arch/arc/include/asm/entry.h
105
GET_CPU_ID \tmp
arch/arc/include/asm/entry.h
106
add2 \tmp, @_current_task, \tmp
arch/arc/include/asm/entry.h
107
st \tsk, [\tmp]
arch/arc/kernel/mcip.c
157
struct mcip_bcr mp;
arch/arc/kernel/mcip.c
159
READ_BCR(ARC_REG_MCIP_BCR, mp);
arch/arc/kernel/mcip.c
163
mp.ver, mp.num_cores,
arch/arc/kernel/mcip.c
164
IS_AVAIL1(mp.ipi, "IPI "),
arch/arc/kernel/mcip.c
165
IS_AVAIL1(mp.idu, "IDU "),
arch/arc/kernel/mcip.c
166
IS_AVAIL1(mp.dbg, "DEBUG "),
arch/arc/kernel/mcip.c
167
IS_AVAIL1(mp.gfrc, "GFRC"));
arch/arc/kernel/mcip.c
381
struct mcip_bcr mp;
arch/arc/kernel/mcip.c
384
READ_BCR(ARC_REG_MCIP_BCR, mp);
arch/arc/kernel/mcip.c
386
if (!mp.idu)
arch/arc/kernel/mcip.c
81
struct mcip_bcr mp;
arch/arc/kernel/mcip.c
83
READ_BCR(ARC_REG_MCIP_BCR, mp);
arch/arc/kernel/mcip.c
89
if (mp.gfrc)
arch/arc/kernel/mcip.c
93
if (mp.dbg)
arch/arc/kernel/setup.c
294
struct mcip_bcr mp;
arch/arc/kernel/setup.c
321
READ_BCR(ARC_REG_MCIP_BCR, mp);
arch/arc/kernel/setup.c
329
IS_AVAIL2(mp.gfrc, "GFRC [SMP 64-bit] ", CONFIG_ARC_TIMERS_64BIT),
arch/arm/include/asm/assembler.h
230
ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count
arch/arm/include/asm/assembler.h
231
add \tmp, \tmp, #1 @ increment it
arch/arm/include/asm/assembler.h
232
str \tmp, [\ti, #TI_PREEMPT]
arch/arm/include/asm/assembler.h
236
ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count
arch/arm/include/asm/assembler.h
237
sub \tmp, \tmp, #1 @ decrement it
arch/arm/include/asm/assembler.h
238
str \tmp, [\ti, #TI_PREEMPT]
arch/arm/include/asm/assembler.h
318
.L0_\@: str_va \rn, __current, \tmp
arch/arm/include/asm/assembler.h
324
str_va \rn, __current, \tmp
arch/arm/include/asm/assembler.h
597
ldr\c \tmp, .La\@
arch/arm/include/asm/assembler.h
606
movw\c \tmp, #:lower16:\sym - .Lpc\@
arch/arm/include/asm/assembler.h
607
movt\c \tmp, #:upper16:\sym - .Lpc\@
arch/arm/include/asm/assembler.h
613
add\c \reg, \tmp, pc
arch/arm/include/asm/assembler.h
615
\op\c \reg, [pc, \tmp]
arch/arm/include/asm/assembler.h
618
.Lb\@: add\c \tmp, \tmp, pc
arch/arm/include/asm/assembler.h
628
\op\c \reg, [\tmp]
arch/arm/include/asm/assembler.h
676
__adldst_l str, \src, \sym, \tmp, \cond
arch/arm/include/asm/assembler.h
683
mov_l \tmp, \sym, \cond
arch/arm/include/asm/assembler.h
695
.L0_\@: sub\cond \tmp, pc, #8 - \offset
arch/arm/include/asm/assembler.h
696
.L1_\@: sub\cond \tmp, \tmp, #4 - \offset
arch/arm/include/asm/assembler.h
699
\op\cond \reg, [\tmp, #\offset]
arch/arm/include/asm/assembler.h
706
.ifnb \tmp
arch/arm/include/asm/assembler.h
707
__ldst_va ldr, \rd, \tmp, \sym, \cond, \offset
arch/arm/include/asm/assembler.h
717
__ldst_va str, \rn, \tmp, \sym, \cond, 0
arch/arm/include/asm/assembler.h
765
eor \tmp, \val, \val, ror #16
arch/arm/include/asm/assembler.h
766
bic \tmp, \tmp, #0x00ff0000
arch/arm/include/asm/assembler.h
768
eor \val, \val, \tmp, lsr #8
arch/arm/include/asm/uaccess-asm.h
21
adds \tmp, \addr, #\size - 1
arch/arm/include/asm/uaccess-asm.h
22
sbcscc \tmp, \tmp, \limit
arch/arm/include/asm/uaccess-asm.h
33
sub \tmp, \limit, #1
arch/arm/include/asm/uaccess-asm.h
34
subs \tmp, \tmp, \addr @ tmp = limit - 1 - addr
arch/arm/include/asm/uaccess-asm.h
35
addhs \tmp, \tmp, #1 @ if (tmp >= 0) {
arch/arm/include/asm/uaccess-asm.h
36
subshs \tmp, \tmp, \size @ tmp = limit - (addr + size) }
arch/arm/include/asm/uaccess-asm.h
49
mov \tmp, #DACR_UACCESS_DISABLE
arch/arm/include/asm/uaccess-asm.h
50
mcr p15, 0, \tmp, c3, c0, 0 @ Set domain register
arch/arm/include/asm/uaccess-asm.h
61
mov \tmp, #DACR_UACCESS_ENABLE
arch/arm/include/asm/uaccess-asm.h
62
mcr p15, 0, \tmp, c3, c0, 0
arch/arm/include/asm/uaccess-asm.h
76
mrc p15, 0, \tmp, c2, c0, 2 @ read TTBCR
arch/arm/include/asm/uaccess-asm.h
77
orr \tmp, \tmp, #TTBCR_EPD0 | TTBCR_T0SZ_MASK
arch/arm/include/asm/uaccess-asm.h
78
orr \tmp, \tmp, #TTBCR_A1
arch/arm/include/asm/uaccess-asm.h
79
mcr p15, 0, \tmp, c2, c0, 2 @ write TTBCR
arch/arm/include/asm/uaccess-asm.h
90
mrc p15, 0, \tmp, c2, c0, 2 @ read TTBCR
arch/arm/include/asm/uaccess-asm.h
91
bic \tmp, \tmp, #TTBCR_EPD0 | TTBCR_T0SZ_MASK
arch/arm/include/asm/uaccess-asm.h
92
bic \tmp, \tmp, #TTBCR_A1
arch/arm/include/asm/uaccess-asm.h
93
mcr p15, 0, \tmp, c2, c0, 2 @ write TTBCR
arch/arm/include/asm/vfpmacros.h
30
ldr \tmp, =elf_hwcap @ may not have MVFR regs
arch/arm/include/asm/vfpmacros.h
31
ldr \tmp, [\tmp, #0]
arch/arm/include/asm/vfpmacros.h
32
tst \tmp, #HWCAP_VFPD32
arch/arm/include/asm/vfpmacros.h
36
VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0
arch/arm/include/asm/vfpmacros.h
37
and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field
arch/arm/include/asm/vfpmacros.h
38
cmp \tmp, #2 @ 32 x 64bit registers?
arch/arm/include/asm/vfpmacros.h
55
ldr \tmp, =elf_hwcap @ may not have MVFR regs
arch/arm/include/asm/vfpmacros.h
56
ldr \tmp, [\tmp, #0]
arch/arm/include/asm/vfpmacros.h
57
tst \tmp, #HWCAP_VFPD32
arch/arm/include/asm/vfpmacros.h
61
VFPFMRX \tmp, MVFR0 @ Media and VFP Feature Register 0
arch/arm/include/asm/vfpmacros.h
62
and \tmp, \tmp, #MVFR0_A_SIMD_MASK @ A_SIMD field
arch/arm/include/asm/vfpmacros.h
63
cmp \tmp, #2 @ 32 x 64bit registers?
arch/arm/lib/bitops.h
16
.arch_extension mp
arch/arm/lib/bitops.h
43
.arch_extension mp
arch/arm/mach-tegra/sleep.h
45
1001: ldr \tmp, [\base]
arch/arm/mach-tegra/sleep.h
46
cmp \tmp, \rn
arch/arm64/include/asm/assembler.h
188
.ifb \tmp
arch/arm64/include/asm/assembler.h
192
adrp \tmp, \sym
arch/arm64/include/asm/assembler.h
193
ldr \dst, [\tmp, :lo12:\sym]
arch/arm64/include/asm/assembler.h
204
adrp \tmp, \sym
arch/arm64/include/asm/assembler.h
205
str \src, [\tmp, :lo12:\sym]
arch/arm64/include/asm/assembler.h
239
adrp \tmp, \sym
arch/arm64/include/asm/assembler.h
240
add \dst, \tmp, #:lo12:\sym
arch/arm64/include/asm/assembler.h
241
get_this_cpu_offset \tmp
arch/arm64/include/asm/assembler.h
242
add \dst, \dst, \tmp
arch/arm64/include/asm/assembler.h
252
get_this_cpu_offset \tmp
arch/arm64/include/asm/assembler.h
253
ldr \dst, [\dst, \tmp]
arch/arm64/include/asm/assembler.h
287
mrs \tmp, ctr_el0 // read CTR
arch/arm64/include/asm/assembler.h
288
ubfm \tmp, \tmp, #16, #19 // cache line size encoding
arch/arm64/include/asm/assembler.h
290
lsl \reg, \reg, \tmp // actual cache line size
arch/arm64/include/asm/assembler.h
297
read_ctr \tmp
arch/arm64/include/asm/assembler.h
298
ubfm \tmp, \tmp, #16, #19 // cache line size encoding
arch/arm64/include/asm/assembler.h
300
lsl \reg, \reg, \tmp // actual cache line size
arch/arm64/include/asm/assembler.h
308
mrs \tmp, ctr_el0 // read CTR
arch/arm64/include/asm/assembler.h
309
and \tmp, \tmp, #0xf // cache line size encoding
arch/arm64/include/asm/assembler.h
311
lsl \reg, \reg, \tmp // actual cache line size
arch/arm64/include/asm/assembler.h
318
read_ctr \tmp
arch/arm64/include/asm/assembler.h
319
and \tmp, \tmp, #0xf // cache line size encoding
arch/arm64/include/asm/assembler.h
321
lsl \reg, \reg, \tmp // actual cache line size
arch/arm64/include/asm/assembler.h
382
sub \tmp, \linesz, #1
arch/arm64/include/asm/assembler.h
383
bic \start, \start, \tmp
arch/arm64/include/asm/assembler.h
385
mov \tmp, \start
arch/arm64/include/asm/assembler.h
410
mov \start, \tmp
arch/arm64/include/asm/assembler.h
411
mov \tmp, xzr
arch/arm64/include/asm/assembler.h
479
phys_to_ttbr \tmp, \zero_page
arch/arm64/include/asm/assembler.h
480
msr ttbr1_el1, \tmp
arch/arm64/include/asm/assembler.h
484
load_ttbr1 \page_table, \tmp, \tmp2
arch/arm64/include/asm/assembler.h
60
mrs \tmp, mdscr_el1
arch/arm64/include/asm/assembler.h
601
mrs \tmp, tcr_el1
arch/arm64/include/asm/assembler.h
602
and \tmp, \tmp, #TCR_EL1_T1SZ_MASK
arch/arm64/include/asm/assembler.h
603
cmp \tmp, #TCR_T1SZ(VA_BITS_MIN)
arch/arm64/include/asm/assembler.h
604
orr \tmp, \ttbr, #TTBR1_BADDR_4852_OFFSET
arch/arm64/include/asm/assembler.h
605
csel \ttbr, \tmp, \ttbr, eq
arch/arm64/include/asm/assembler.h
61
bic \tmp, \tmp, #MDSCR_EL1_SS
arch/arm64/include/asm/assembler.h
62
msr mdscr_el1, \tmp
arch/arm64/include/asm/assembler.h
70
mrs \tmp, mdscr_el1
arch/arm64/include/asm/assembler.h
71
orr \tmp, \tmp, #MDSCR_EL1_SS
arch/arm64/include/asm/assembler.h
72
msr mdscr_el1, \tmp
arch/arm64/include/asm/assembler.h
760
get_current_task \tmp
arch/arm64/include/asm/assembler.h
761
ldr \tmp, [\tmp, #TSK_TI_PREEMPT]
arch/arm64/include/asm/assembler.h
769
cbz \tmp, \lbl
arch/arm64/include/asm/assembler.h
836
mov \tmp, #32 // Patched to correct the immediate
arch/arm64/include/asm/assembler.h
840
subs \tmp, \tmp, #1
arch/arm64/include/asm/assembler.h
851
__mitigate_spectre_bhb_loop \tmp
arch/arm64/include/asm/el2_setup.h
106
mrs \tmp, hcr_el2
arch/arm64/include/asm/el2_setup.h
107
and \tmp, \tmp, #HCR_E2H
arch/arm64/include/asm/el2_setup.h
108
cbz \tmp, \fail
arch/arm64/include/asm/el2_setup.h
132
mrs \tmp, id_aa64dfr0_el1
arch/arm64/include/asm/el2_setup.h
133
ubfx \tmp, \tmp, #ID_AA64DFR0_EL1_PMSVer_SHIFT, #4
arch/arm64/include/asm/el2_setup.h
134
cmp \tmp, \version
arch/arm64/include/asm/el2_setup.h
501
ldr_l \tmp, \idreg\()_el1_sys_val
arch/arm64/include/asm/el2_setup.h
502
ubfx \tmp, \tmp, #\fld, #\width
arch/arm64/include/asm/el2_setup.h
503
cbnz \tmp, \pass
arch/arm64/include/asm/el2_setup.h
508
__check_override \idreg \fld 4 \pass \fail \tmp \ignore
arch/arm64/include/asm/fpsimdmacros.h
39
mrs \tmp, fpcr
arch/arm64/include/asm/fpsimdmacros.h
40
cmp \tmp, \state
arch/arm64/include/asm/kvm_asm.h
309
adr_this_cpu \reg, kvm_host_data, \tmp
arch/arm64/include/asm/kvm_asm.h
324
adr_this_cpu \ctxt, kvm_hyp_ctxt, \tmp
arch/arm64/include/asm/kvm_asm.h
374
mrs \tmp, sp_el0
arch/arm64/include/asm/kvm_asm.h
375
str \tmp, [\ctxt, #CPU_SP_EL0_OFFSET]
arch/arm64/include/asm/kvm_asm.h
379
ldr \tmp, [\ctxt, #CPU_SP_EL0_OFFSET]
arch/arm64/include/asm/kvm_asm.h
380
msr sp_el0, \tmp
arch/arm64/include/asm/kvm_mmu.h
62
ldr_l \tmp, hyp_physvirt_offset
arch/arm64/include/asm/kvm_mmu.h
63
add \reg, \reg, \tmp
arch/arm64/include/asm/kvm_mmu.h
78
hyp_pa \reg, \tmp
arch/arm64/include/asm/kvm_mmu.h
82
movz \tmp, #0
arch/arm64/include/asm/kvm_mmu.h
83
movk \tmp, #0, lsl #16
arch/arm64/include/asm/kvm_mmu.h
84
movk \tmp, #0, lsl #32
arch/arm64/include/asm/kvm_mmu.h
85
movk \tmp, #0, lsl #48
arch/arm64/include/asm/kvm_mmu.h
89
add \reg, \reg, \tmp
arch/loongarch/include/asm/asmmacro.h
192
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
193
PTR_ADD \tmp, \tmp, \thread
arch/loongarch/include/asm/asmmacro.h
194
fst.d $f0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
195
fst.d $f1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
196
fst.d $f2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
197
fst.d $f3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
198
fst.d $f4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
199
fst.d $f5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
200
fst.d $f6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
201
fst.d $f7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
202
fst.d $f8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
203
fst.d $f9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
204
fst.d $f10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
205
fst.d $f11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
206
fst.d $f12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
207
fst.d $f13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
208
fst.d $f14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
209
fst.d $f15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
210
fst.d $f16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
211
fst.d $f17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
212
fst.d $f18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
213
fst.d $f19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
214
fst.d $f20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
215
fst.d $f21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
216
fst.d $f22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
217
fst.d $f23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
218
fst.d $f24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
219
fst.d $f25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
220
fst.d $f26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
221
fst.d $f27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
222
fst.d $f28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
223
fst.d $f29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
224
fst.d $f30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
225
fst.d $f31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
229
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
230
PTR_ADD \tmp, \tmp, \thread
arch/loongarch/include/asm/asmmacro.h
231
fld.d $f0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
232
fld.d $f1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
233
fld.d $f2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
234
fld.d $f3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
235
fld.d $f4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
236
fld.d $f5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
237
fld.d $f6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
238
fld.d $f7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
239
fld.d $f8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
240
fld.d $f9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
241
fld.d $f10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
242
fld.d $f11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
243
fld.d $f12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
244
fld.d $f13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
245
fld.d $f14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
246
fld.d $f15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
247
fld.d $f16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
248
fld.d $f17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
249
fld.d $f18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
250
fld.d $f19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
251
fld.d $f20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
252
fld.d $f21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
253
fld.d $f22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
254
fld.d $f23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
255
fld.d $f24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
256
fld.d $f25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
257
fld.d $f26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
258
fld.d $f27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
259
fld.d $f28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
260
fld.d $f29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
261
fld.d $f30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
262
fld.d $f31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
266
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
267
PTR_ADD \tmp, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
268
vst $vr0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
269
vst $vr1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
270
vst $vr2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
271
vst $vr3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
272
vst $vr4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
273
vst $vr5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
274
vst $vr6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
275
vst $vr7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
276
vst $vr8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
277
vst $vr9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
278
vst $vr10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
279
vst $vr11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
280
vst $vr12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
281
vst $vr13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
282
vst $vr14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
283
vst $vr15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
284
vst $vr16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
285
vst $vr17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
286
vst $vr18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
287
vst $vr19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
288
vst $vr20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
289
vst $vr21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
290
vst $vr22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
291
vst $vr23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
292
vst $vr24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
293
vst $vr25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
294
vst $vr26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
295
vst $vr27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
296
vst $vr28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
297
vst $vr29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
298
vst $vr30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
299
vst $vr31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
303
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
304
PTR_ADD \tmp, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
305
vld $vr0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
306
vld $vr1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
307
vld $vr2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
308
vld $vr3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
309
vld $vr4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
310
vld $vr5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
311
vld $vr6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
312
vld $vr7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
313
vld $vr8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
314
vld $vr9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
315
vld $vr10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
316
vld $vr11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
317
vld $vr12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
318
vld $vr13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
319
vld $vr14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
320
vld $vr15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
321
vld $vr16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
322
vld $vr17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
323
vld $vr18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
324
vld $vr19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
325
vld $vr20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
326
vld $vr21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
327
vld $vr22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
328
vld $vr23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
329
vld $vr24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
330
vld $vr25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
331
vld $vr26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
332
vld $vr27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
333
vld $vr28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
334
vld $vr29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
335
vld $vr30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
336
vld $vr31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
352
vpickve2gr.d \tmp, \vd, 1
arch/loongarch/include/asm/asmmacro.h
353
st.d \tmp, \base, (\off+8)
arch/loongarch/include/asm/asmmacro.h
357
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
358
PTR_ADD \base, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
359
lsx_save_upper $vr0, \base, \tmp, (THREAD_FPR0-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
360
lsx_save_upper $vr1, \base, \tmp, (THREAD_FPR1-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
361
lsx_save_upper $vr2, \base, \tmp, (THREAD_FPR2-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
362
lsx_save_upper $vr3, \base, \tmp, (THREAD_FPR3-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
363
lsx_save_upper $vr4, \base, \tmp, (THREAD_FPR4-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
364
lsx_save_upper $vr5, \base, \tmp, (THREAD_FPR5-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
365
lsx_save_upper $vr6, \base, \tmp, (THREAD_FPR6-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
366
lsx_save_upper $vr7, \base, \tmp, (THREAD_FPR7-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
367
lsx_save_upper $vr8, \base, \tmp, (THREAD_FPR8-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
368
lsx_save_upper $vr9, \base, \tmp, (THREAD_FPR9-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
369
lsx_save_upper $vr10, \base, \tmp, (THREAD_FPR10-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
370
lsx_save_upper $vr11, \base, \tmp, (THREAD_FPR11-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
371
lsx_save_upper $vr12, \base, \tmp, (THREAD_FPR12-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
372
lsx_save_upper $vr13, \base, \tmp, (THREAD_FPR13-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
373
lsx_save_upper $vr14, \base, \tmp, (THREAD_FPR14-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
374
lsx_save_upper $vr15, \base, \tmp, (THREAD_FPR15-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
375
lsx_save_upper $vr16, \base, \tmp, (THREAD_FPR16-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
376
lsx_save_upper $vr17, \base, \tmp, (THREAD_FPR17-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
377
lsx_save_upper $vr18, \base, \tmp, (THREAD_FPR18-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
378
lsx_save_upper $vr19, \base, \tmp, (THREAD_FPR19-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
379
lsx_save_upper $vr20, \base, \tmp, (THREAD_FPR20-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
380
lsx_save_upper $vr21, \base, \tmp, (THREAD_FPR21-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
381
lsx_save_upper $vr22, \base, \tmp, (THREAD_FPR22-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
382
lsx_save_upper $vr23, \base, \tmp, (THREAD_FPR23-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
383
lsx_save_upper $vr24, \base, \tmp, (THREAD_FPR24-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
384
lsx_save_upper $vr25, \base, \tmp, (THREAD_FPR25-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
385
lsx_save_upper $vr26, \base, \tmp, (THREAD_FPR26-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
386
lsx_save_upper $vr27, \base, \tmp, (THREAD_FPR27-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
387
lsx_save_upper $vr28, \base, \tmp, (THREAD_FPR28-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
388
lsx_save_upper $vr29, \base, \tmp, (THREAD_FPR29-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
389
lsx_save_upper $vr30, \base, \tmp, (THREAD_FPR30-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
390
lsx_save_upper $vr31, \base, \tmp, (THREAD_FPR31-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
394
ld.d \tmp, \base, (\off+8)
arch/loongarch/include/asm/asmmacro.h
395
vinsgr2vr.d \vd, \tmp, 1
arch/loongarch/include/asm/asmmacro.h
399
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
400
PTR_ADD \base, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
401
lsx_restore_upper $vr0, \base, \tmp, (THREAD_FPR0-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
402
lsx_restore_upper $vr1, \base, \tmp, (THREAD_FPR1-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
403
lsx_restore_upper $vr2, \base, \tmp, (THREAD_FPR2-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
404
lsx_restore_upper $vr3, \base, \tmp, (THREAD_FPR3-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
405
lsx_restore_upper $vr4, \base, \tmp, (THREAD_FPR4-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
406
lsx_restore_upper $vr5, \base, \tmp, (THREAD_FPR5-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
407
lsx_restore_upper $vr6, \base, \tmp, (THREAD_FPR6-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
408
lsx_restore_upper $vr7, \base, \tmp, (THREAD_FPR7-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
409
lsx_restore_upper $vr8, \base, \tmp, (THREAD_FPR8-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
410
lsx_restore_upper $vr9, \base, \tmp, (THREAD_FPR9-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
411
lsx_restore_upper $vr10, \base, \tmp, (THREAD_FPR10-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
412
lsx_restore_upper $vr11, \base, \tmp, (THREAD_FPR11-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
413
lsx_restore_upper $vr12, \base, \tmp, (THREAD_FPR12-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
414
lsx_restore_upper $vr13, \base, \tmp, (THREAD_FPR13-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
415
lsx_restore_upper $vr14, \base, \tmp, (THREAD_FPR14-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
416
lsx_restore_upper $vr15, \base, \tmp, (THREAD_FPR15-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
417
lsx_restore_upper $vr16, \base, \tmp, (THREAD_FPR16-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
418
lsx_restore_upper $vr17, \base, \tmp, (THREAD_FPR17-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
419
lsx_restore_upper $vr18, \base, \tmp, (THREAD_FPR18-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
420
lsx_restore_upper $vr19, \base, \tmp, (THREAD_FPR19-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
421
lsx_restore_upper $vr20, \base, \tmp, (THREAD_FPR20-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
422
lsx_restore_upper $vr21, \base, \tmp, (THREAD_FPR21-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
423
lsx_restore_upper $vr22, \base, \tmp, (THREAD_FPR22-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
424
lsx_restore_upper $vr23, \base, \tmp, (THREAD_FPR23-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
425
lsx_restore_upper $vr24, \base, \tmp, (THREAD_FPR24-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
426
lsx_restore_upper $vr25, \base, \tmp, (THREAD_FPR25-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
427
lsx_restore_upper $vr26, \base, \tmp, (THREAD_FPR26-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
428
lsx_restore_upper $vr27, \base, \tmp, (THREAD_FPR27-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
429
lsx_restore_upper $vr28, \base, \tmp, (THREAD_FPR28-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
430
lsx_restore_upper $vr29, \base, \tmp, (THREAD_FPR29-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
431
lsx_restore_upper $vr30, \base, \tmp, (THREAD_FPR30-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
432
lsx_restore_upper $vr31, \base, \tmp, (THREAD_FPR31-THREAD_FPR0)
arch/loongarch/include/asm/asmmacro.h
436
vinsgr2vr.d \vd, \tmp, 1
arch/loongarch/include/asm/asmmacro.h
440
not \tmp, zero
arch/loongarch/include/asm/asmmacro.h
441
lsx_init_upper $vr0 \tmp
arch/loongarch/include/asm/asmmacro.h
442
lsx_init_upper $vr1 \tmp
arch/loongarch/include/asm/asmmacro.h
443
lsx_init_upper $vr2 \tmp
arch/loongarch/include/asm/asmmacro.h
444
lsx_init_upper $vr3 \tmp
arch/loongarch/include/asm/asmmacro.h
445
lsx_init_upper $vr4 \tmp
arch/loongarch/include/asm/asmmacro.h
446
lsx_init_upper $vr5 \tmp
arch/loongarch/include/asm/asmmacro.h
447
lsx_init_upper $vr6 \tmp
arch/loongarch/include/asm/asmmacro.h
448
lsx_init_upper $vr7 \tmp
arch/loongarch/include/asm/asmmacro.h
449
lsx_init_upper $vr8 \tmp
arch/loongarch/include/asm/asmmacro.h
450
lsx_init_upper $vr9 \tmp
arch/loongarch/include/asm/asmmacro.h
451
lsx_init_upper $vr10 \tmp
arch/loongarch/include/asm/asmmacro.h
452
lsx_init_upper $vr11 \tmp
arch/loongarch/include/asm/asmmacro.h
453
lsx_init_upper $vr12 \tmp
arch/loongarch/include/asm/asmmacro.h
454
lsx_init_upper $vr13 \tmp
arch/loongarch/include/asm/asmmacro.h
455
lsx_init_upper $vr14 \tmp
arch/loongarch/include/asm/asmmacro.h
456
lsx_init_upper $vr15 \tmp
arch/loongarch/include/asm/asmmacro.h
457
lsx_init_upper $vr16 \tmp
arch/loongarch/include/asm/asmmacro.h
458
lsx_init_upper $vr17 \tmp
arch/loongarch/include/asm/asmmacro.h
459
lsx_init_upper $vr18 \tmp
arch/loongarch/include/asm/asmmacro.h
460
lsx_init_upper $vr19 \tmp
arch/loongarch/include/asm/asmmacro.h
461
lsx_init_upper $vr20 \tmp
arch/loongarch/include/asm/asmmacro.h
462
lsx_init_upper $vr21 \tmp
arch/loongarch/include/asm/asmmacro.h
463
lsx_init_upper $vr22 \tmp
arch/loongarch/include/asm/asmmacro.h
464
lsx_init_upper $vr23 \tmp
arch/loongarch/include/asm/asmmacro.h
465
lsx_init_upper $vr24 \tmp
arch/loongarch/include/asm/asmmacro.h
466
lsx_init_upper $vr25 \tmp
arch/loongarch/include/asm/asmmacro.h
467
lsx_init_upper $vr26 \tmp
arch/loongarch/include/asm/asmmacro.h
468
lsx_init_upper $vr27 \tmp
arch/loongarch/include/asm/asmmacro.h
469
lsx_init_upper $vr28 \tmp
arch/loongarch/include/asm/asmmacro.h
470
lsx_init_upper $vr29 \tmp
arch/loongarch/include/asm/asmmacro.h
471
lsx_init_upper $vr30 \tmp
arch/loongarch/include/asm/asmmacro.h
472
lsx_init_upper $vr31 \tmp
arch/loongarch/include/asm/asmmacro.h
476
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
477
PTR_ADD \tmp, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
478
xvst $xr0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
479
xvst $xr1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
480
xvst $xr2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
481
xvst $xr3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
482
xvst $xr4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
483
xvst $xr5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
484
xvst $xr6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
485
xvst $xr7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
486
xvst $xr8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
487
xvst $xr9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
488
xvst $xr10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
489
xvst $xr11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
490
xvst $xr12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
491
xvst $xr13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
492
xvst $xr14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
493
xvst $xr15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
494
xvst $xr16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
495
xvst $xr17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
496
xvst $xr18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
497
xvst $xr19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
498
xvst $xr20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
499
xvst $xr21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
500
xvst $xr22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
501
xvst $xr23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
502
xvst $xr24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
503
xvst $xr25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
504
xvst $xr26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
505
xvst $xr27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
506
xvst $xr28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
507
xvst $xr29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
508
xvst $xr30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
509
xvst $xr31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
51
movfcsr2gr \tmp, fcsr0
arch/loongarch/include/asm/asmmacro.h
513
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
514
PTR_ADD \tmp, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
515
xvld $xr0, \tmp, THREAD_FPR0 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
516
xvld $xr1, \tmp, THREAD_FPR1 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
517
xvld $xr2, \tmp, THREAD_FPR2 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
518
xvld $xr3, \tmp, THREAD_FPR3 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
519
xvld $xr4, \tmp, THREAD_FPR4 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
520
xvld $xr5, \tmp, THREAD_FPR5 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
521
xvld $xr6, \tmp, THREAD_FPR6 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
522
xvld $xr7, \tmp, THREAD_FPR7 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
523
xvld $xr8, \tmp, THREAD_FPR8 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
524
xvld $xr9, \tmp, THREAD_FPR9 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
525
xvld $xr10, \tmp, THREAD_FPR10 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
526
xvld $xr11, \tmp, THREAD_FPR11 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
527
xvld $xr12, \tmp, THREAD_FPR12 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
528
xvld $xr13, \tmp, THREAD_FPR13 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
529
xvld $xr14, \tmp, THREAD_FPR14 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
53
st.w \tmp, \thread, THREAD_FCSR
arch/loongarch/include/asm/asmmacro.h
530
xvld $xr15, \tmp, THREAD_FPR15 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
531
xvld $xr16, \tmp, THREAD_FPR16 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
532
xvld $xr17, \tmp, THREAD_FPR17 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
533
xvld $xr18, \tmp, THREAD_FPR18 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
534
xvld $xr19, \tmp, THREAD_FPR19 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
535
xvld $xr20, \tmp, THREAD_FPR20 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
536
xvld $xr21, \tmp, THREAD_FPR21 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
537
xvld $xr22, \tmp, THREAD_FPR22 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
538
xvld $xr23, \tmp, THREAD_FPR23 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
539
xvld $xr24, \tmp, THREAD_FPR24 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
540
xvld $xr25, \tmp, THREAD_FPR25 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
541
xvld $xr26, \tmp, THREAD_FPR26 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
542
xvld $xr27, \tmp, THREAD_FPR27 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
543
xvld $xr28, \tmp, THREAD_FPR28 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
544
xvld $xr29, \tmp, THREAD_FPR29 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
545
xvld $xr30, \tmp, THREAD_FPR30 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
546
xvld $xr31, \tmp, THREAD_FPR31 - THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
55
stptr.w \tmp, \thread, THREAD_FCSR
arch/loongarch/include/asm/asmmacro.h
575
li.w \tmp, THREAD_FPR0
arch/loongarch/include/asm/asmmacro.h
576
PTR_ADD \base, \thread, \tmp
arch/loongarch/include/asm/asmmacro.h
59
andi \tmp, \tmp, FPU_CSR_TM
arch/loongarch/include/asm/asmmacro.h
60
beqz \tmp, 1f
arch/loongarch/include/asm/asmmacro.h
618
xvinsgr2vr.d \xd, \tmp, 2
arch/loongarch/include/asm/asmmacro.h
619
xvinsgr2vr.d \xd, \tmp, 3
arch/loongarch/include/asm/asmmacro.h
62
x86mftop \tmp
arch/loongarch/include/asm/asmmacro.h
623
not \tmp, zero
arch/loongarch/include/asm/asmmacro.h
624
lasx_init_upper $xr0 \tmp
arch/loongarch/include/asm/asmmacro.h
625
lasx_init_upper $xr1 \tmp
arch/loongarch/include/asm/asmmacro.h
626
lasx_init_upper $xr2 \tmp
arch/loongarch/include/asm/asmmacro.h
627
lasx_init_upper $xr3 \tmp
arch/loongarch/include/asm/asmmacro.h
628
lasx_init_upper $xr4 \tmp
arch/loongarch/include/asm/asmmacro.h
629
lasx_init_upper $xr5 \tmp
arch/loongarch/include/asm/asmmacro.h
63
stptr.w \tmp, \thread, THREAD_FTOP
arch/loongarch/include/asm/asmmacro.h
630
lasx_init_upper $xr6 \tmp
arch/loongarch/include/asm/asmmacro.h
631
lasx_init_upper $xr7 \tmp
arch/loongarch/include/asm/asmmacro.h
632
lasx_init_upper $xr8 \tmp
arch/loongarch/include/asm/asmmacro.h
633
lasx_init_upper $xr9 \tmp
arch/loongarch/include/asm/asmmacro.h
634
lasx_init_upper $xr10 \tmp
arch/loongarch/include/asm/asmmacro.h
635
lasx_init_upper $xr11 \tmp
arch/loongarch/include/asm/asmmacro.h
636
lasx_init_upper $xr12 \tmp
arch/loongarch/include/asm/asmmacro.h
637
lasx_init_upper $xr13 \tmp
arch/loongarch/include/asm/asmmacro.h
638
lasx_init_upper $xr14 \tmp
arch/loongarch/include/asm/asmmacro.h
639
lasx_init_upper $xr15 \tmp
arch/loongarch/include/asm/asmmacro.h
640
lasx_init_upper $xr16 \tmp
arch/loongarch/include/asm/asmmacro.h
641
lasx_init_upper $xr17 \tmp
arch/loongarch/include/asm/asmmacro.h
642
lasx_init_upper $xr18 \tmp
arch/loongarch/include/asm/asmmacro.h
643
lasx_init_upper $xr19 \tmp
arch/loongarch/include/asm/asmmacro.h
644
lasx_init_upper $xr20 \tmp
arch/loongarch/include/asm/asmmacro.h
645
lasx_init_upper $xr21 \tmp
arch/loongarch/include/asm/asmmacro.h
646
lasx_init_upper $xr22 \tmp
arch/loongarch/include/asm/asmmacro.h
647
lasx_init_upper $xr23 \tmp
arch/loongarch/include/asm/asmmacro.h
648
lasx_init_upper $xr24 \tmp
arch/loongarch/include/asm/asmmacro.h
649
lasx_init_upper $xr25 \tmp
arch/loongarch/include/asm/asmmacro.h
650
lasx_init_upper $xr26 \tmp
arch/loongarch/include/asm/asmmacro.h
651
lasx_init_upper $xr27 \tmp
arch/loongarch/include/asm/asmmacro.h
652
lasx_init_upper $xr28 \tmp
arch/loongarch/include/asm/asmmacro.h
653
lasx_init_upper $xr29 \tmp
arch/loongarch/include/asm/asmmacro.h
654
lasx_init_upper $xr30 \tmp
arch/loongarch/include/asm/asmmacro.h
655
lasx_init_upper $xr31 \tmp
arch/mips/include/asm/asmmacro-32.h
19
cfc1 \tmp, fcr31
arch/mips/include/asm/asmmacro-32.h
36
sw \tmp, THREAD_FCR31(\thread)
arch/mips/include/asm/asmmacro-32.h
43
lw \tmp, THREAD_FCR31(\thread)
arch/mips/include/asm/asmmacro-32.h
60
ctc1 \tmp, fcr31
arch/mips/include/asm/asmmacro.h
104
sw \tmp, THREAD_FCR31(\thread)
arch/mips/include/asm/asmmacro.h
135
sll \tmp, \status, 5
arch/mips/include/asm/asmmacro.h
136
bgez \tmp, 10f
arch/mips/include/asm/asmmacro.h
140
fpu_save_16even \thread \tmp
arch/mips/include/asm/asmmacro.h
146
lw \tmp, THREAD_FCR31(\thread)
arch/mips/include/asm/asmmacro.h
163
ctc1 \tmp, fcr31
arch/mips/include/asm/asmmacro.h
194
sll \tmp, \status, 5
arch/mips/include/asm/asmmacro.h
195
bgez \tmp, 10f # 16 register mode?
arch/mips/include/asm/asmmacro.h
200
fpu_restore_16even \thread \tmp
arch/mips/include/asm/asmmacro.h
87
cfc1 \tmp, fcr31
arch/powerpc/mm/book3s32/tlb.c
83
struct vm_area_struct *mp;
arch/powerpc/mm/book3s32/tlb.c
92
for_each_vma(vmi, mp)
arch/powerpc/mm/book3s32/tlb.c
93
hash__flush_range(mp->vm_mm, mp->vm_start, mp->vm_end);
arch/riscv/include/asm/asm.h
100
add \dst, \dst, \tmp
arch/riscv/include/asm/asm.h
109
asm_per_cpu \dst \ptr \tmp
arch/riscv/include/asm/asm.h
94
lw \tmp, TASK_TI_CPU_NUM(tp)
arch/riscv/include/asm/asm.h
95
slli \tmp, \tmp, RISCV_LGPTR
arch/riscv/include/asm/asm.h
97
add \dst, \dst, \tmp
arch/riscv/include/asm/asm.h
98
REG_L \tmp, 0(\dst)
arch/riscv/include/asm/scs.h
18
load_per_cpu gp, irq_shadow_call_stack_ptr, \tmp
arch/s390/kernel/perf_pai.c
128
static void pai_free(struct pai_mapptr *mp)
arch/s390/kernel/perf_pai.c
130
if (mp->mapptr->fullpage)
arch/s390/kernel/perf_pai.c
131
free_page((unsigned long)mp->mapptr->area);
arch/s390/kernel/perf_pai.c
133
kfree(mp->mapptr->area);
arch/s390/kernel/perf_pai.c
134
kfree(mp->mapptr->paiext_cb);
arch/s390/kernel/perf_pai.c
135
kvfree(mp->mapptr->save);
arch/s390/kernel/perf_pai.c
136
kfree(mp->mapptr);
arch/s390/kernel/perf_pai.c
137
mp->mapptr = NULL;
arch/s390/kernel/perf_pai.c
146
struct pai_mapptr *mp = per_cpu_ptr(pai_root[idx].mapptr, cpu);
arch/s390/kernel/perf_pai.c
147
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
155
pai_free(mp);
arch/s390/kernel/perf_pai.c
195
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
197
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
242
struct pai_mapptr *mp;
arch/s390/kernel/perf_pai.c
251
mp = per_cpu_ptr(pai_root[idx].mapptr, cpu);
arch/s390/kernel/perf_pai.c
252
cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
261
mp->mapptr = cpump;
arch/s390/kernel/perf_pai.c
288
pai_free(mp);
arch/s390/kernel/perf_pai.c
443
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
444
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
470
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
471
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
503
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
504
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
528
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
529
struct pai_map *cpump = mp->mapptr;
arch/s390/kernel/perf_pai.c
662
struct pai_mapptr *mp = this_cpu_ptr(pai_root[idx].mapptr);
arch/s390/kernel/perf_pai.c
663
struct pai_map *cpump = mp->mapptr;
arch/sparc/include/asm/vio.h
333
u64 mp;
arch/sparc/kernel/chmc.c
254
struct chmc_obp_map *mp;
arch/sparc/kernel/chmc.c
259
mp = &p->map[0];
arch/sparc/kernel/chmc.c
261
mp = &p->map[1];
arch/sparc/kernel/chmc.c
267
map_val = mp->dimm_map[dimm_map_index];
arch/sparc/kernel/chmc.c
270
*pin_p = mp->pin_map[cache_line_offset];
arch/sparc/kernel/ds.c
1184
val = mdesc_get_property(hp, vdev->mp, "id", NULL);
arch/sparc/kernel/ldc.c
2392
u64 mp;
arch/sparc/kernel/ldc.c
2398
mp = mdesc_node_by_name(hp, MDESC_NODE_NULL, "platform");
arch/sparc/kernel/ldc.c
2400
if (mp == MDESC_NODE_NULL)
arch/sparc/kernel/ldc.c
2403
v = mdesc_get_property(hp, mp, "domaining-enabled", NULL);
arch/sparc/kernel/mdesc.c
1008
mp = mdesc_node_by_name(hp, MDESC_NODE_NULL, "sockets");
arch/sparc/kernel/mdesc.c
1009
if (mp != MDESC_NODE_NULL)
arch/sparc/kernel/mdesc.c
1010
set_sock_ids_by_socket(hp, mp);
arch/sparc/kernel/mdesc.c
1013
static void mark_proc_ids(struct mdesc_handle *hp, u64 mp, int proc_id)
arch/sparc/kernel/mdesc.c
1017
mdesc_for_each_arc(a, hp, mp, MDESC_ARC_TYPE_BACK) {
arch/sparc/kernel/mdesc.c
1035
u64 mp;
arch/sparc/kernel/mdesc.c
1038
mdesc_for_each_node_by_name(hp, mp, exec_unit_name) {
arch/sparc/kernel/mdesc.c
1042
type = mdesc_get_property(hp, mp, "type", &len);
arch/sparc/kernel/mdesc.c
1047
mark_proc_ids(hp, mp, idx);
arch/sparc/kernel/mdesc.c
1080
static void get_mondo_data(struct mdesc_handle *hp, u64 mp,
arch/sparc/kernel/mdesc.c
1086
val = mdesc_get_property(hp, mp, "q-cpu-mondo-#bits", NULL);
arch/sparc/kernel/mdesc.c
1089
val = mdesc_get_property(hp, mp, "q-dev-mondo-#bits", NULL);
arch/sparc/kernel/mdesc.c
1092
val = mdesc_get_property(hp, mp, "q-resumable-#bits", NULL);
arch/sparc/kernel/mdesc.c
1095
val = mdesc_get_property(hp, mp, "q-nonresumable-#bits", NULL);
arch/sparc/kernel/mdesc.c
1111
u64 mp;
arch/sparc/kernel/mdesc.c
1113
mdesc_for_each_node_by_name(hp, mp, "cpu") {
arch/sparc/kernel/mdesc.c
1114
const u64 *id = mdesc_get_property(hp, mp, "id", NULL);
arch/sparc/kernel/mdesc.c
1128
ret = func(hp, mp, cpuid, arg);
arch/sparc/kernel/mdesc.c
1137
static void *record_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid,
arch/sparc/kernel/mdesc.c
1156
static void * __init check_one_pgsz(struct mdesc_handle *hp, u64 mp, int cpuid, void *arg)
arch/sparc/kernel/mdesc.c
1158
const u64 *pgsz_prop = mdesc_get_property(hp, mp, "mmu-page-size-list", NULL);
arch/sparc/kernel/mdesc.c
1180
static void *fill_in_one_cpu(struct mdesc_handle *hp, u64 mp, int cpuid,
arch/sparc/kernel/mdesc.c
1183
const u64 *cfreq = mdesc_get_property(hp, mp, "clock-frequency", NULL);
arch/sparc/kernel/mdesc.c
1202
get_mondo_data(hp, mp, tb);
arch/sparc/kernel/mdesc.c
1204
mdesc_for_each_arc(a, hp, mp, MDESC_ARC_TYPE_FWD) {
arch/sparc/kernel/mdesc.c
818
static void fill_in_one_cache(cpuinfo_sparc *c, struct mdesc_handle *hp, u64 mp)
arch/sparc/kernel/mdesc.c
820
const u64 *level = mdesc_get_property(hp, mp, "level", NULL);
arch/sparc/kernel/mdesc.c
821
const u64 *size = mdesc_get_property(hp, mp, "size", NULL);
arch/sparc/kernel/mdesc.c
822
const u64 *line_size = mdesc_get_property(hp, mp, "line-size", NULL);
arch/sparc/kernel/mdesc.c
826
type = mdesc_get_property(hp, mp, "type", &type_len);
arch/sparc/kernel/mdesc.c
851
mdesc_for_each_arc(a, hp, mp, MDESC_ARC_TYPE_FWD) {
arch/sparc/kernel/mdesc.c
908
static void mark_core_ids(struct mdesc_handle *hp, u64 mp,
arch/sparc/kernel/mdesc.c
911
find_back_node_value(hp, mp, "cpu", __mark_core_id, core_id, 10);
arch/sparc/kernel/mdesc.c
914
static void mark_max_cache_ids(struct mdesc_handle *hp, u64 mp,
arch/sparc/kernel/mdesc.c
917
find_back_node_value(hp, mp, "cpu", __mark_max_cache_id,
arch/sparc/kernel/mdesc.c
924
u64 mp;
arch/sparc/kernel/mdesc.c
931
mdesc_for_each_node_by_name(hp, mp, "cache") {
arch/sparc/kernel/mdesc.c
936
level = mdesc_get_property(hp, mp, "level", NULL);
arch/sparc/kernel/mdesc.c
940
type = mdesc_get_property(hp, mp, "type", &len);
arch/sparc/kernel/mdesc.c
944
mark_core_ids(hp, mp, idx);
arch/sparc/kernel/mdesc.c
951
u64 mp;
arch/sparc/kernel/mdesc.c
959
mdesc_for_each_node_by_name(hp, mp, "cache") {
arch/sparc/kernel/mdesc.c
962
cur_lvl = mdesc_get_property(hp, mp, "level", NULL);
arch/sparc/kernel/mdesc.c
965
mark_max_cache_ids(hp, mp, idx);
arch/sparc/kernel/mdesc.c
972
static void set_sock_ids_by_socket(struct mdesc_handle *hp, u64 mp)
arch/sparc/kernel/mdesc.c
976
mdesc_for_each_node_by_name(hp, mp, "socket") {
arch/sparc/kernel/mdesc.c
979
mdesc_for_each_arc(a, hp, mp, MDESC_ARC_TYPE_FWD) {
arch/sparc/kernel/mdesc.c
998
u64 mp;
arch/sparc/kernel/vio.c
251
static void vio_fill_channel_info(struct mdesc_handle *hp, u64 mp,
arch/sparc/kernel/vio.c
259
mdesc_for_each_arc(a, hp, mp, MDESC_ARC_TYPE_FWD) {
arch/sparc/kernel/vio.c
291
static struct vio_dev *vio_create_one(struct mdesc_handle *hp, u64 mp,
arch/sparc/kernel/vio.c
301
type = mdesc_get_property(hp, mp, "device-type", &tlen);
arch/sparc/kernel/vio.c
303
type = mdesc_get_property(hp, mp, "name", &tlen);
arch/sparc/kernel/vio.c
305
type = mdesc_node_name(hp, mp);
arch/sparc/kernel/vio.c
315
id = mdesc_get_property(hp, mp, "id", NULL);
arch/sparc/kernel/vio.c
317
cfg_handle = vio_cfg_handle(hp, mp);
arch/sparc/kernel/vio.c
319
compat = mdesc_get_property(hp, mp, "device-type", &clen);
arch/sparc/kernel/vio.c
334
vdev->mp = mp;
arch/sparc/kernel/vio.c
346
vio_fill_channel_info(hp, mp, vdev);
arch/sparc/kernel/vio.c
384
err = mdesc_get_node_info(hp, mp, node_name,
arch/x86/kernel/mpparse.c
41
static int __init mpf_checksum(unsigned char *mp, int len)
arch/x86/kernel/mpparse.c
46
sum += *mp++;
arch/xtensa/include/asm/asmmacro.h
201
extui \tmp, \a, 16, 16
arch/xtensa/include/asm/asmmacro.h
202
bnez \tmp, 0f
arch/xtensa/include/asm/asmmacro.h
206
extui \tmp, \a, 24, 8
arch/xtensa/include/asm/asmmacro.h
207
bnez \tmp, 1f
arch/xtensa/include/asm/asmmacro.h
211
movi \tmp, __nsau_data
arch/xtensa/include/asm/asmmacro.h
213
add \tmp, \tmp, \a
arch/xtensa/include/asm/asmmacro.h
214
l8ui \tmp, \tmp, 0
arch/xtensa/include/asm/asmmacro.h
215
add \cnt, \cnt, \tmp
arch/xtensa/include/asm/asmmacro.h
223
neg \tmp, \src
arch/xtensa/include/asm/asmmacro.h
224
movgez \tmp, \src, \src
arch/xtensa/include/asm/asmmacro.h
225
mov \dst, \tmp
drivers/block/aoe/aoeblk.c
334
mempool_t *mp;
drivers/block/aoe/aoeblk.c
357
mp = mempool_create(MIN_BUFS, mempool_alloc_slab, mempool_free_slab,
drivers/block/aoe/aoeblk.c
359
if (mp == NULL) {
drivers/block/aoe/aoeblk.c
391
d->bufpool = mp;
drivers/block/aoe/aoeblk.c
426
mempool_destroy(mp);
drivers/block/aoe/aoechr.c
139
char *mp;
drivers/block/aoe/aoechr.c
152
mp = kmemdup(msg, n, GFP_ATOMIC);
drivers/block/aoe/aoechr.c
153
if (!mp)
drivers/block/aoe/aoechr.c
156
em->msg = mp;
drivers/block/aoe/aoechr.c
224
char *mp;
drivers/block/aoe/aoechr.c
262
mp = em->msg;
drivers/block/aoe/aoechr.c
272
n = copy_to_user(buf, mp, len);
drivers/block/aoe/aoechr.c
273
kfree(mp);
drivers/block/sunvdc.c
1015
ldc_timeout = mdesc_get_property(hp, vdev->mp, "vdc-timeout", NULL);
drivers/clocksource/arc_timer.c
105
struct mcip_bcr mp;
drivers/clocksource/arc_timer.c
108
READ_BCR(ARC_REG_MCIP_BCR, mp);
drivers/clocksource/arc_timer.c
109
if (!mp.gfrc) {
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10160
dml_print("DML::%s: PrefetchAndImmediateFlipSupported = %0d\n", __func__, mode_lib->mp.PrefetchAndImmediateFlipSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10161
return mode_lib->mp.PrefetchAndImmediateFlipSupported;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10211
dml_uint_t plane_idx = mode_lib->mp.pipe_plane[pipe_idx];
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10220
plane_idx = mode_lib->mp.pipe_plane[surface_idx]; \
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10229
dml_get_var_func(wm_urgent, dml_float_t, mode_lib->mp.Watermark.UrgentWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10230
dml_get_var_func(wm_stutter_exit, dml_float_t, mode_lib->mp.Watermark.StutterExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10231
dml_get_var_func(wm_stutter_enter_exit, dml_float_t, mode_lib->mp.Watermark.StutterEnterPlusExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10232
dml_get_var_func(wm_memory_trip, dml_float_t, mode_lib->mp.UrgentLatency);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10233
dml_get_var_func(wm_fclk_change, dml_float_t, mode_lib->mp.Watermark.FCLKChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10234
dml_get_var_func(wm_usr_retraining, dml_float_t, mode_lib->mp.Watermark.USRRetrainingWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10235
dml_get_var_func(wm_dram_clock_change, dml_float_t, mode_lib->mp.Watermark.DRAMClockChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10236
dml_get_var_func(wm_z8_stutter_enter_exit, dml_float_t, mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10237
dml_get_var_func(wm_z8_stutter, dml_float_t, mode_lib->mp.Watermark.Z8StutterExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10238
dml_get_var_func(fraction_of_urgent_bandwidth, dml_float_t, mode_lib->mp.FractionOfUrgentBandwidth);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10239
dml_get_var_func(fraction_of_urgent_bandwidth_imm_flip, dml_float_t, mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10240
dml_get_var_func(urgent_latency, dml_float_t, mode_lib->mp.UrgentLatency);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10241
dml_get_var_func(clk_dcf_deepsleep, dml_float_t, mode_lib->mp.DCFCLKDeepSleep);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10242
dml_get_var_func(wm_writeback_dram_clock_change, dml_float_t, mode_lib->mp.Watermark.WritebackDRAMClockChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10243
dml_get_var_func(wm_writeback_urgent, dml_float_t, mode_lib->mp.Watermark.WritebackUrgentWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10244
dml_get_var_func(stutter_efficiency, dml_float_t, mode_lib->mp.StutterEfficiency);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10245
dml_get_var_func(stutter_efficiency_no_vblank, dml_float_t, mode_lib->mp.StutterEfficiencyNotIncludingVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10246
dml_get_var_func(stutter_efficiency_z8, dml_float_t, mode_lib->mp.Z8StutterEfficiency);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10247
dml_get_var_func(stutter_num_bursts_z8, dml_float_t, mode_lib->mp.Z8NumberOfStutterBurstsPerFrame);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10248
dml_get_var_func(stutter_period, dml_float_t, mode_lib->mp.StutterPeriod);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10249
dml_get_var_func(stutter_efficiency_z8_bestcase, dml_float_t, mode_lib->mp.Z8StutterEfficiencyBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10250
dml_get_var_func(stutter_num_bursts_z8_bestcase, dml_float_t, mode_lib->mp.Z8NumberOfStutterBurstsPerFrameBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10251
dml_get_var_func(stutter_period_bestcase, dml_float_t, mode_lib->mp.StutterPeriodBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10252
dml_get_var_func(urgent_extra_latency, dml_float_t, mode_lib->mp.UrgentExtraLatency);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10253
dml_get_var_func(fclk_change_latency, dml_float_t, mode_lib->mp.MaxActiveFCLKChangeLatencySupported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10254
dml_get_var_func(dispclk_calculated, dml_float_t, mode_lib->mp.Dispclk_calculated);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10255
dml_get_var_func(total_data_read_bw, dml_float_t, mode_lib->mp.TotalDataReadBandwidth);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10258
dml_get_var_func(tcalc, dml_float_t, mode_lib->mp.TCalc);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10259
dml_get_var_func(comp_buffer_size_kbytes, dml_uint_t, mode_lib->mp.CompressedBufferSizeInkByte);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10265
dml_get_var_func(total_immediate_flip_bytes, dml_uint_t, mode_lib->mp.TotImmediateFlipBytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10267
dml_get_per_surface_var_func(dsc_delay, dml_uint_t, mode_lib->mp.DSCDelay); // this is the dsc latency
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10268
dml_get_per_surface_var_func(dppclk_calculated, dml_float_t, mode_lib->mp.Dppclk_calculated);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10269
dml_get_per_surface_var_func(dscclk_calculated, dml_float_t, mode_lib->mp.DSCCLK_calculated);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10270
dml_get_per_surface_var_func(min_ttu_vblank_in_us, dml_float_t, mode_lib->mp.MinTTUVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10271
dml_get_per_surface_var_func(vratio_prefetch_l, dml_float_t, mode_lib->mp.VRatioPrefetchY);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10272
dml_get_per_surface_var_func(vratio_prefetch_c, dml_float_t, mode_lib->mp.VRatioPrefetchC);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10273
dml_get_per_surface_var_func(dst_x_after_scaler, dml_uint_t, mode_lib->mp.DSTXAfterScaler);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10274
dml_get_per_surface_var_func(dst_y_after_scaler, dml_uint_t, mode_lib->mp.DSTYAfterScaler);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10275
dml_get_per_surface_var_func(dst_y_per_vm_vblank, dml_float_t, mode_lib->mp.DestinationLinesToRequestVMInVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10276
dml_get_per_surface_var_func(dst_y_per_row_vblank, dml_float_t, mode_lib->mp.DestinationLinesToRequestRowInVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10277
dml_get_per_surface_var_func(dst_y_prefetch, dml_float_t, mode_lib->mp.DestinationLinesForPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10278
dml_get_per_surface_var_func(dst_y_per_vm_flip, dml_float_t, mode_lib->mp.DestinationLinesToRequestVMInImmediateFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10279
dml_get_per_surface_var_func(dst_y_per_row_flip, dml_float_t, mode_lib->mp.DestinationLinesToRequestRowInImmediateFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10280
dml_get_per_surface_var_func(dst_y_per_pte_row_nom_l, dml_float_t, mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_L);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10281
dml_get_per_surface_var_func(dst_y_per_pte_row_nom_c, dml_float_t, mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_C);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10282
dml_get_per_surface_var_func(dst_y_per_meta_row_nom_l, dml_float_t, mode_lib->mp.DST_Y_PER_META_ROW_NOM_L);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10283
dml_get_per_surface_var_func(dst_y_per_meta_row_nom_c, dml_float_t, mode_lib->mp.DST_Y_PER_META_ROW_NOM_C);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10284
dml_get_per_surface_var_func(refcyc_per_vm_group_vblank_in_us, dml_float_t, mode_lib->mp.TimePerVMGroupVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10285
dml_get_per_surface_var_func(refcyc_per_vm_group_flip_in_us, dml_float_t, mode_lib->mp.TimePerVMGroupFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10286
dml_get_per_surface_var_func(refcyc_per_vm_req_vblank_in_us, dml_float_t, mode_lib->mp.TimePerVMRequestVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10287
dml_get_per_surface_var_func(refcyc_per_vm_req_flip_in_us, dml_float_t, mode_lib->mp.TimePerVMRequestFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10288
dml_get_per_surface_var_func(refcyc_per_vm_dmdata_in_us, dml_float_t, mode_lib->mp.Tdmdl_vm);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10289
dml_get_per_surface_var_func(dmdata_dl_delta_in_us, dml_float_t, mode_lib->mp.Tdmdl);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10290
dml_get_per_surface_var_func(refcyc_per_line_delivery_l_in_us, dml_float_t, mode_lib->mp.DisplayPipeLineDeliveryTimeLuma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10291
dml_get_per_surface_var_func(refcyc_per_line_delivery_c_in_us, dml_float_t, mode_lib->mp.DisplayPipeLineDeliveryTimeChroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10292
dml_get_per_surface_var_func(refcyc_per_line_delivery_pre_l_in_us, dml_float_t, mode_lib->mp.DisplayPipeLineDeliveryTimeLumaPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10293
dml_get_per_surface_var_func(refcyc_per_line_delivery_pre_c_in_us, dml_float_t, mode_lib->mp.DisplayPipeLineDeliveryTimeChromaPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10294
dml_get_per_surface_var_func(refcyc_per_req_delivery_l_in_us, dml_float_t, mode_lib->mp.DisplayPipeRequestDeliveryTimeLuma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10295
dml_get_per_surface_var_func(refcyc_per_req_delivery_c_in_us, dml_float_t, mode_lib->mp.DisplayPipeRequestDeliveryTimeChroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10296
dml_get_per_surface_var_func(refcyc_per_req_delivery_pre_l_in_us, dml_float_t, mode_lib->mp.DisplayPipeRequestDeliveryTimeLumaPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10297
dml_get_per_surface_var_func(refcyc_per_req_delivery_pre_c_in_us, dml_float_t, mode_lib->mp.DisplayPipeRequestDeliveryTimeChromaPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10298
dml_get_per_surface_var_func(refcyc_per_cursor_req_delivery_in_us, dml_float_t, mode_lib->mp.CursorRequestDeliveryTime);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10299
dml_get_per_surface_var_func(refcyc_per_cursor_req_delivery_pre_in_us, dml_float_t, mode_lib->mp.CursorRequestDeliveryTimePrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10300
dml_get_per_surface_var_func(refcyc_per_meta_chunk_nom_l_in_us, dml_float_t, mode_lib->mp.TimePerMetaChunkNominal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10301
dml_get_per_surface_var_func(refcyc_per_meta_chunk_nom_c_in_us, dml_float_t, mode_lib->mp.TimePerChromaMetaChunkNominal);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10302
dml_get_per_surface_var_func(refcyc_per_meta_chunk_vblank_l_in_us, dml_float_t, mode_lib->mp.TimePerMetaChunkVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10303
dml_get_per_surface_var_func(refcyc_per_meta_chunk_vblank_c_in_us, dml_float_t, mode_lib->mp.TimePerChromaMetaChunkVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10304
dml_get_per_surface_var_func(refcyc_per_meta_chunk_flip_l_in_us, dml_float_t, mode_lib->mp.TimePerMetaChunkFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10305
dml_get_per_surface_var_func(refcyc_per_meta_chunk_flip_c_in_us, dml_float_t, mode_lib->mp.TimePerChromaMetaChunkFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10306
dml_get_per_surface_var_func(refcyc_per_pte_group_nom_l_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_nom_luma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10307
dml_get_per_surface_var_func(refcyc_per_pte_group_nom_c_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_nom_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10308
dml_get_per_surface_var_func(refcyc_per_pte_group_vblank_l_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_vblank_luma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10309
dml_get_per_surface_var_func(refcyc_per_pte_group_vblank_c_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_vblank_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10310
dml_get_per_surface_var_func(refcyc_per_pte_group_flip_l_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_flip_luma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10311
dml_get_per_surface_var_func(refcyc_per_pte_group_flip_c_in_us, dml_float_t, mode_lib->mp.time_per_pte_group_flip_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10312
dml_get_per_surface_var_func(dpte_group_size_in_bytes, dml_uint_t, mode_lib->mp.dpte_group_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10313
dml_get_per_surface_var_func(vm_group_size_in_bytes, dml_uint_t, mode_lib->mp.vm_group_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10316
dml_get_per_surface_var_func(dpte_row_height_l, dml_uint_t, mode_lib->mp.dpte_row_height);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10317
dml_get_per_surface_var_func(dpte_row_height_c, dml_uint_t, mode_lib->mp.dpte_row_height_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10318
dml_get_per_surface_var_func(dpte_row_height_linear_l, dml_uint_t, mode_lib->mp.dpte_row_height_linear);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10319
dml_get_per_surface_var_func(dpte_row_height_linear_c, dml_uint_t, mode_lib->mp.dpte_row_height_linear_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10320
dml_get_per_surface_var_func(meta_row_height_l, dml_uint_t, mode_lib->mp.meta_row_height);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10321
dml_get_per_surface_var_func(meta_row_height_c, dml_uint_t, mode_lib->mp.meta_row_height_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10323
dml_get_per_surface_var_func(vstartup_calculated, dml_uint_t, mode_lib->mp.VStartup);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10324
dml_get_per_surface_var_func(vupdate_offset, dml_uint_t, mode_lib->mp.VUpdateOffsetPix);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10325
dml_get_per_surface_var_func(vupdate_width, dml_uint_t, mode_lib->mp.VUpdateWidthPix);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10326
dml_get_per_surface_var_func(vready_offset, dml_uint_t, mode_lib->mp.VReadyOffsetPix);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10327
dml_get_per_surface_var_func(vready_at_or_after_vsync, dml_uint_t, mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10328
dml_get_per_surface_var_func(min_dst_y_next_start, dml_uint_t, mode_lib->mp.MIN_DST_Y_NEXT_START);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10331
dml_get_per_surface_var_func(use_mall_for_static_screen, dml_uint_t, mode_lib->mp.UsesMALLForStaticScreen);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10332
dml_get_per_surface_var_func(surface_size_for_mall, dml_uint_t, mode_lib->mp.SurfaceSizeInTheMALL);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10333
dml_get_per_surface_var_func(dcc_max_uncompressed_block_l, dml_uint_t, mode_lib->mp.DCCYMaxUncompressedBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10334
dml_get_per_surface_var_func(dcc_max_compressed_block_l, dml_uint_t, mode_lib->mp.DCCYMaxCompressedBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10335
dml_get_per_surface_var_func(dcc_independent_block_l, dml_uint_t, mode_lib->mp.DCCYIndependentBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10336
dml_get_per_surface_var_func(dcc_max_uncompressed_block_c, dml_uint_t, mode_lib->mp.DCCCMaxUncompressedBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10337
dml_get_per_surface_var_func(dcc_max_compressed_block_c, dml_uint_t, mode_lib->mp.DCCCMaxCompressedBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10338
dml_get_per_surface_var_func(dcc_independent_block_c, dml_uint_t, mode_lib->mp.DCCCIndependentBlock);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10339
dml_get_per_surface_var_func(max_active_dram_clock_change_latency_supported, dml_uint_t, mode_lib->mp.MaxActiveDRAMClockChangeLatencySupported);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10340
dml_get_per_surface_var_func(pte_buffer_mode, dml_uint_t, mode_lib->mp.PTE_BUFFER_MODE);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10341
dml_get_per_surface_var_func(bigk_fragment_size, dml_uint_t, mode_lib->mp.BIGK_FRAGMENT_SIZE);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10342
dml_get_per_surface_var_func(dpte_bytes_per_row, dml_uint_t, mode_lib->mp.PixelPTEBytesPerRow);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
10343
dml_get_per_surface_var_func(meta_bytes_per_row, dml_uint_t, mode_lib->mp.MetaRowByte);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
6473
&mode_lib->mp.FractionOfUrgentBandwidth, // dml_float_t *FractionOfUrgentBandwidth
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
7801
mode_lib->mp.UrgentLatency = mode_lib->ms.UrgLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8308
struct mode_program_st *locals = &mode_lib->mp;
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8315
mode_lib->mp.num_active_pipes = dml_get_num_active_pipes(&mode_lib->ms.cache_display_cfg);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8316
dml_calc_pipe_plane_mapping(&mode_lib->ms.cache_display_cfg.hw, mode_lib->mp.pipe_plane);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
8321
dml_print("DML::%s: num_active_pipes = %u\n", __func__, mode_lib->mp.num_active_pipes);
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core.c
9489
memmove(&mode_lib->mp.Watermark, CalculateWatermarks_params->Watermark, sizeof(struct Watermarks));
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_core_structs.h
1911
struct mode_program_st mp; // struct for mode programming
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
759
dml_uint_t plane_idx = mode_lib->mp.pipe_plane[pipe_idx];
drivers/gpu/drm/amd/display/dc/dml2_0/display_mode_util.c
771
if (plane_idx == mode_lib->mp.pipe_plane[i]) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
318
programming->plane_programming[plane_index].num_dpps_required = core->clean_me_up.mode_lib.mp.NoOfDPP[plane_index];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
580
in_out->programming->plane_programming[plane_index].num_dpps_required = core->clean_me_up.mode_lib.mp.NoOfDPP[plane_index];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
589
if (core->clean_me_up.mode_lib.mp.MaxActiveDRAMClockChangeLatencySupported[plane_index] >= core->clean_me_up.mode_lib.soc.power_management_parameters.dram_clk_change_blackout_us)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4.c
591
else if (core->clean_me_up.mode_lib.mp.TWait[plane_index] >= core->clean_me_up.mode_lib.soc.power_management_parameters.dram_clk_change_blackout_us)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10413
memset(&mode_lib->mp, 0, sizeof(struct dml2_core_internal_mode_program));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10418
mode_lib->mp.num_active_pipes = dml_get_num_active_pipes(display_cfg->num_planes, cfg_support_info);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10419
dml_calc_pipe_plane_mapping(cfg_support_info, mode_lib->mp.pipe_plane);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10421
mode_lib->mp.Dcfclk = programming->min_clocks.dcn4x.active.dcfclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10422
mode_lib->mp.FabricClock = programming->min_clocks.dcn4x.active.fclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10423
mode_lib->mp.dram_bw_mbps = uclk_khz_to_dram_bw_mbps(programming->min_clocks.dcn4x.active.uclk_khz, &mode_lib->soc.clk_table.dram_config, &min_clk_table->dram_bw_table);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10424
mode_lib->mp.uclk_freq_mhz = programming->min_clocks.dcn4x.active.uclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10425
mode_lib->mp.GlobalDPPCLK = programming->min_clocks.dcn4x.dpprefclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10427
mode_lib->mp.qos_param_index = get_qos_param_index(programming->min_clocks.dcn4x.active.uclk_khz, mode_lib->soc.qos_parameters.qos_params.dcn4x.per_uclk_dpm_params);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10428
mode_lib->mp.active_min_uclk_dpm_index = get_active_min_uclk_dpm_index(programming->min_clocks.dcn4x.active.uclk_khz, &mode_lib->soc.clk_table);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10442
mode_lib->mp.ODMMode[k] = dml2_odm_mode_combine_4to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10445
mode_lib->mp.ODMMode[k] = dml2_odm_mode_combine_3to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10448
mode_lib->mp.ODMMode[k] = dml2_odm_mode_combine_2to1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10452
mode_lib->mp.ODMMode[k] = dml2_odm_mode_mso_1to4;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10454
mode_lib->mp.ODMMode[k] = dml2_odm_mode_mso_1to2;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10456
mode_lib->mp.ODMMode[k] = dml2_odm_mode_bypass;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10462
mode_lib->mp.NoOfDPP[k] = cfg_support_info->plane_support_info[k].dpps_used;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10463
mode_lib->mp.Dppclk[k] = programming->plane_programming[k].min_clocks.dcn4x.dppclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10464
DML_ASSERT(mode_lib->mp.Dppclk[k] > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10469
mode_lib->mp.DSCCLK[k] = programming->stream_programming[stream_index].min_clocks.dcn4x.dscclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10470
DML_LOG_VERBOSE("DML::%s: k=%d stream_index=%d, mode_lib->mp.DSCCLK = %f\n", __func__, k, stream_index, mode_lib->mp.DSCCLK[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10473
mode_lib->mp.Dispclk = programming->min_clocks.dcn4x.dispclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10474
mode_lib->mp.DCFCLKDeepSleep = programming->min_clocks.dcn4x.deepsleep_dcfclk_khz / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10476
DML_ASSERT(mode_lib->mp.Dcfclk > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10477
DML_ASSERT(mode_lib->mp.FabricClock > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10478
DML_ASSERT(mode_lib->mp.dram_bw_mbps > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10479
DML_ASSERT(mode_lib->mp.uclk_freq_mhz > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10480
DML_ASSERT(mode_lib->mp.GlobalDPPCLK > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10481
DML_ASSERT(mode_lib->mp.Dispclk > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10482
DML_ASSERT(mode_lib->mp.DCFCLKDeepSleep > 0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10487
DML_LOG_VERBOSE("DML::%s: num_active_pipes = %u\n", __func__, mode_lib->mp.num_active_pipes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10488
DML_LOG_VERBOSE("DML::%s: Dcfclk = %f\n", __func__, mode_lib->mp.Dcfclk);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10489
DML_LOG_VERBOSE("DML::%s: FabricClock = %f\n", __func__, mode_lib->mp.FabricClock);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10490
DML_LOG_VERBOSE("DML::%s: dram_bw_mbps = %f\n", __func__, mode_lib->mp.dram_bw_mbps);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10491
DML_LOG_VERBOSE("DML::%s: uclk_freq_mhz = %f\n", __func__, mode_lib->mp.uclk_freq_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10492
DML_LOG_VERBOSE("DML::%s: Dispclk = %f\n", __func__, mode_lib->mp.Dispclk);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10494
DML_LOG_VERBOSE("DML::%s: Dppclk[%0d] = %f\n", __func__, k, mode_lib->mp.Dppclk[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10496
DML_LOG_VERBOSE("DML::%s: GlobalDPPCLK = %f\n", __func__, mode_lib->mp.GlobalDPPCLK);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10497
DML_LOG_VERBOSE("DML::%s: DCFCLKDeepSleep = %f\n", __func__, mode_lib->mp.DCFCLKDeepSleep);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10505
for (k = 0; k < mode_lib->mp.num_active_pipes; ++k) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10506
DML_LOG_VERBOSE("DML::%s: pipe=%d is in plane=%d\n", __func__, k, mode_lib->mp.pipe_plane[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10507
DML_LOG_VERBOSE("DML::%s: Per-plane DPPPerSurface[%0d] = %d\n", __func__, k, mode_lib->mp.NoOfDPP[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10547
&mode_lib->mp.PSCL_THROUGHPUT[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10548
&mode_lib->mp.PSCL_THROUGHPUT_CHROMA[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10549
&mode_lib->mp.DPPCLKUsingSingleDPP[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10560
&mode_lib->mp.BytePerPixelY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10561
&mode_lib->mp.BytePerPixelC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10562
&mode_lib->mp.BytePerPixelInDETY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10563
&mode_lib->mp.BytePerPixelInDETC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10564
&mode_lib->mp.Read256BlockHeightY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10565
&mode_lib->mp.Read256BlockHeightC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10566
&mode_lib->mp.Read256BlockWidthY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10567
&mode_lib->mp.Read256BlockWidthC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10568
&mode_lib->mp.MacroTileHeightY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10569
&mode_lib->mp.MacroTileHeightC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10570
&mode_lib->mp.MacroTileWidthY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10571
&mode_lib->mp.MacroTileWidthC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10572
&mode_lib->mp.surf_linear128_l[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10573
&mode_lib->mp.surf_linear128_c[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10580
mode_lib->mp.ODMMode,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10581
mode_lib->mp.BytePerPixelY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10582
mode_lib->mp.BytePerPixelC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10583
mode_lib->mp.Read256BlockHeightY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10584
mode_lib->mp.Read256BlockHeightC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10585
mode_lib->mp.Read256BlockWidthY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10586
mode_lib->mp.Read256BlockWidthC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10587
mode_lib->mp.surf_linear128_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10588
mode_lib->mp.surf_linear128_c,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10589
mode_lib->mp.NoOfDPP,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10592
mode_lib->mp.req_per_swath_ub_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10593
mode_lib->mp.req_per_swath_ub_c,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10594
mode_lib->mp.SwathWidthSingleDPPY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10595
mode_lib->mp.SwathWidthSingleDPPC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10596
mode_lib->mp.SwathWidthY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10597
mode_lib->mp.SwathWidthC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10600
mode_lib->mp.swath_width_luma_ub,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10601
mode_lib->mp.swath_width_chroma_ub);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10604
mode_lib->mp.cursor_bw[k] = display_cfg->plane_descriptors[k].cursor.num_cursors * display_cfg->plane_descriptors[k].cursor.cursor_width * display_cfg->plane_descriptors[k].cursor.cursor_bpp / 8.0 /
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10606
mode_lib->mp.vactive_sw_bw_l[k] = mode_lib->mp.SwathWidthSingleDPPY[k] * mode_lib->mp.BytePerPixelY[k] / (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total / ((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000)) * display_cfg->plane_descriptors[k].composition.scaler_info.plane0.v_ratio;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10607
mode_lib->mp.vactive_sw_bw_c[k] = mode_lib->mp.SwathWidthSingleDPPC[k] * mode_lib->mp.BytePerPixelC[k] / (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total / ((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000)) * display_cfg->plane_descriptors[k].composition.scaler_info.plane1.v_ratio;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10608
DML_LOG_VERBOSE("DML::%s: vactive_sw_bw_l[%i] = %fBps\n", __func__, k, mode_lib->mp.vactive_sw_bw_l[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10609
DML_LOG_VERBOSE("DML::%s: vactive_sw_bw_c[%i] = %fBps\n", __func__, k, mode_lib->mp.vactive_sw_bw_c[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10625
CalculateSwathAndDETConfiguration_params->ReadBandwidthLuma = mode_lib->mp.vactive_sw_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10626
CalculateSwathAndDETConfiguration_params->ReadBandwidthChroma = mode_lib->mp.vactive_sw_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10629
CalculateSwathAndDETConfiguration_params->Read256BytesBlockHeightY = mode_lib->mp.Read256BlockHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10630
CalculateSwathAndDETConfiguration_params->Read256BytesBlockHeightC = mode_lib->mp.Read256BlockHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10631
CalculateSwathAndDETConfiguration_params->Read256BytesBlockWidthY = mode_lib->mp.Read256BlockWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10632
CalculateSwathAndDETConfiguration_params->Read256BytesBlockWidthC = mode_lib->mp.Read256BlockWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10633
CalculateSwathAndDETConfiguration_params->surf_linear128_l = mode_lib->mp.surf_linear128_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10634
CalculateSwathAndDETConfiguration_params->surf_linear128_c = mode_lib->mp.surf_linear128_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10635
CalculateSwathAndDETConfiguration_params->ODMMode = mode_lib->mp.ODMMode;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10636
CalculateSwathAndDETConfiguration_params->DPPPerSurface = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10637
CalculateSwathAndDETConfiguration_params->BytePerPixY = mode_lib->mp.BytePerPixelY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10638
CalculateSwathAndDETConfiguration_params->BytePerPixC = mode_lib->mp.BytePerPixelC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10639
CalculateSwathAndDETConfiguration_params->BytePerPixDETY = mode_lib->mp.BytePerPixelInDETY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10640
CalculateSwathAndDETConfiguration_params->BytePerPixDETC = mode_lib->mp.BytePerPixelInDETC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10644
CalculateSwathAndDETConfiguration_params->req_per_swath_ub_l = mode_lib->mp.req_per_swath_ub_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10645
CalculateSwathAndDETConfiguration_params->req_per_swath_ub_c = mode_lib->mp.req_per_swath_ub_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10650
CalculateSwathAndDETConfiguration_params->SwathHeightY = mode_lib->mp.SwathHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10651
CalculateSwathAndDETConfiguration_params->SwathHeightC = mode_lib->mp.SwathHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10652
CalculateSwathAndDETConfiguration_params->request_size_bytes_luma = mode_lib->mp.request_size_bytes_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10653
CalculateSwathAndDETConfiguration_params->request_size_bytes_chroma = mode_lib->mp.request_size_bytes_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10654
CalculateSwathAndDETConfiguration_params->DETBufferSizeInKByte = mode_lib->mp.DETBufferSizeInKByte;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10655
CalculateSwathAndDETConfiguration_params->DETBufferSizeY = mode_lib->mp.DETBufferSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10656
CalculateSwathAndDETConfiguration_params->DETBufferSizeC = mode_lib->mp.DETBufferSizeC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10659
CalculateSwathAndDETConfiguration_params->UnboundedRequestEnabled = &mode_lib->mp.UnboundedRequestEnabled;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10660
CalculateSwathAndDETConfiguration_params->compbuf_reserved_space_64b = &mode_lib->mp.compbuf_reserved_space_64b;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10661
CalculateSwathAndDETConfiguration_params->hw_debug5 = &mode_lib->mp.hw_debug5;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10662
CalculateSwathAndDETConfiguration_params->CompressedBufferSizeInkByte = &mode_lib->mp.CompressedBufferSizeInkByte;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10671
mode_lib->mp.DSCDelay[k] = DSCDelayRequirement(cfg_support_info->stream_support_info[display_cfg->plane_descriptors[k].stream_index].dsc_enable,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10672
mode_lib->mp.ODMMode[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10687
mode_lib->mp.SurfaceSizeInTheMALL[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10693
mode_lib->mp.BytePerPixelY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10694
mode_lib->mp.BytePerPixelC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10695
mode_lib->mp.Read256BlockWidthY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10696
mode_lib->mp.Read256BlockWidthC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10697
mode_lib->mp.Read256BlockHeightY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10698
mode_lib->mp.Read256BlockHeightC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10699
mode_lib->mp.MacroTileWidthY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10700
mode_lib->mp.MacroTileWidthC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10701
mode_lib->mp.MacroTileHeightY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10702
mode_lib->mp.MacroTileHeightC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10705
mode_lib->mp.SurfaceSizeInTheMALL,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10711
s->SurfaceParameters[k].DPPPerSurface = mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10715
s->SurfaceParameters[k].BlockWidth256BytesY = mode_lib->mp.Read256BlockWidthY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10716
s->SurfaceParameters[k].BlockHeight256BytesY = mode_lib->mp.Read256BlockHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10717
s->SurfaceParameters[k].BlockWidth256BytesC = mode_lib->mp.Read256BlockWidthC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10718
s->SurfaceParameters[k].BlockHeight256BytesC = mode_lib->mp.Read256BlockHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10719
s->SurfaceParameters[k].BlockWidthY = mode_lib->mp.MacroTileWidthY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10720
s->SurfaceParameters[k].BlockHeightY = mode_lib->mp.MacroTileHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10721
s->SurfaceParameters[k].BlockWidthC = mode_lib->mp.MacroTileWidthC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10722
s->SurfaceParameters[k].BlockHeightC = mode_lib->mp.MacroTileHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10728
s->SurfaceParameters[k].BytePerPixelY = mode_lib->mp.BytePerPixelY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10729
s->SurfaceParameters[k].BytePerPixelC = mode_lib->mp.BytePerPixelC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10743
s->SurfaceParameters[k].SwathHeightY = mode_lib->mp.SwathHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10744
s->SurfaceParameters[k].SwathHeightC = mode_lib->mp.SwathHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10752
CalculateVMRowAndSwath_params->SurfaceSizeInMALL = mode_lib->mp.SurfaceSizeInTheMALL;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10756
CalculateVMRowAndSwath_params->SwathWidthY = mode_lib->mp.SwathWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10757
CalculateVMRowAndSwath_params->SwathWidthC = mode_lib->mp.SwathWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10764
CalculateVMRowAndSwath_params->dpte_row_width_luma_ub = mode_lib->mp.dpte_row_width_luma_ub;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10765
CalculateVMRowAndSwath_params->dpte_row_width_chroma_ub = mode_lib->mp.dpte_row_width_chroma_ub;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10766
CalculateVMRowAndSwath_params->dpte_row_height_luma = mode_lib->mp.dpte_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10767
CalculateVMRowAndSwath_params->dpte_row_height_chroma = mode_lib->mp.dpte_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10768
CalculateVMRowAndSwath_params->dpte_row_height_linear_luma = mode_lib->mp.dpte_row_height_linear;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10769
CalculateVMRowAndSwath_params->dpte_row_height_linear_chroma = mode_lib->mp.dpte_row_height_linear_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10770
CalculateVMRowAndSwath_params->vm_group_bytes = mode_lib->mp.vm_group_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10771
CalculateVMRowAndSwath_params->dpte_group_bytes = mode_lib->mp.dpte_group_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10772
CalculateVMRowAndSwath_params->PixelPTEReqWidthY = mode_lib->mp.PixelPTEReqWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10773
CalculateVMRowAndSwath_params->PixelPTEReqHeightY = mode_lib->mp.PixelPTEReqHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10774
CalculateVMRowAndSwath_params->PTERequestSizeY = mode_lib->mp.PTERequestSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10775
CalculateVMRowAndSwath_params->PixelPTEReqWidthC = mode_lib->mp.PixelPTEReqWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10776
CalculateVMRowAndSwath_params->PixelPTEReqHeightC = mode_lib->mp.PixelPTEReqHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10777
CalculateVMRowAndSwath_params->PTERequestSizeC = mode_lib->mp.PTERequestSizeC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10782
CalculateVMRowAndSwath_params->dpde0_bytes_per_frame_ub_l = mode_lib->mp.dpde0_bytes_per_frame_ub_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10783
CalculateVMRowAndSwath_params->dpde0_bytes_per_frame_ub_c = mode_lib->mp.dpde0_bytes_per_frame_ub_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10784
CalculateVMRowAndSwath_params->PrefetchSourceLinesY = mode_lib->mp.PrefetchSourceLinesY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10785
CalculateVMRowAndSwath_params->PrefetchSourceLinesC = mode_lib->mp.PrefetchSourceLinesC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10786
CalculateVMRowAndSwath_params->VInitPreFillY = mode_lib->mp.VInitPreFillY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10787
CalculateVMRowAndSwath_params->VInitPreFillC = mode_lib->mp.VInitPreFillC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10788
CalculateVMRowAndSwath_params->MaxNumSwathY = mode_lib->mp.MaxNumSwathY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10789
CalculateVMRowAndSwath_params->MaxNumSwathC = mode_lib->mp.MaxNumSwathC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10790
CalculateVMRowAndSwath_params->dpte_row_bw = mode_lib->mp.dpte_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10791
CalculateVMRowAndSwath_params->PixelPTEBytesPerRow = mode_lib->mp.PixelPTEBytesPerRow;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10794
CalculateVMRowAndSwath_params->vm_bytes = mode_lib->mp.vm_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10795
CalculateVMRowAndSwath_params->use_one_row_for_frame = mode_lib->mp.use_one_row_for_frame;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10796
CalculateVMRowAndSwath_params->use_one_row_for_frame_flip = mode_lib->mp.use_one_row_for_frame_flip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10797
CalculateVMRowAndSwath_params->is_using_mall_for_ss = mode_lib->mp.is_using_mall_for_ss;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10798
CalculateVMRowAndSwath_params->PTE_BUFFER_MODE = mode_lib->mp.PTE_BUFFER_MODE;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10799
CalculateVMRowAndSwath_params->BIGK_FRAGMENT_SIZE = mode_lib->mp.BIGK_FRAGMENT_SIZE;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10801
CalculateVMRowAndSwath_params->meta_row_bw = mode_lib->mp.meta_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10802
CalculateVMRowAndSwath_params->meta_row_bytes = mode_lib->mp.meta_row_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10805
CalculateVMRowAndSwath_params->meta_req_width_luma = mode_lib->mp.meta_req_width;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10806
CalculateVMRowAndSwath_params->meta_req_height_luma = mode_lib->mp.meta_req_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10807
CalculateVMRowAndSwath_params->meta_row_width_luma = mode_lib->mp.meta_row_width;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10808
CalculateVMRowAndSwath_params->meta_row_height_luma = mode_lib->mp.meta_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10809
CalculateVMRowAndSwath_params->meta_pte_bytes_per_frame_ub_l = mode_lib->mp.meta_pte_bytes_per_frame_ub_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10810
CalculateVMRowAndSwath_params->meta_req_width_chroma = mode_lib->mp.meta_req_width_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10811
CalculateVMRowAndSwath_params->meta_row_height_chroma = mode_lib->mp.meta_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10812
CalculateVMRowAndSwath_params->meta_row_width_chroma = mode_lib->mp.meta_row_width_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10813
CalculateVMRowAndSwath_params->meta_req_height_chroma = mode_lib->mp.meta_req_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10814
CalculateVMRowAndSwath_params->meta_pte_bytes_per_frame_ub_c = mode_lib->mp.meta_pte_bytes_per_frame_ub_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10821
mode_lib->mp.mall_prefetch_sdp_overhead_factor[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10822
mode_lib->mp.mall_prefetch_dram_overhead_factor[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10823
mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p0[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10824
mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p0[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10825
mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p1[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10826
mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p1[k] = 1.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10848
calculate_mcache_setting_params->blk_width_l = mode_lib->mp.MacroTileWidthY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10849
calculate_mcache_setting_params->blk_height_l = mode_lib->mp.MacroTileHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10853
calculate_mcache_setting_params->bytes_per_pixel_l = mode_lib->mp.BytePerPixelY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10859
calculate_mcache_setting_params->blk_width_c = mode_lib->mp.MacroTileWidthC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10860
calculate_mcache_setting_params->blk_height_c = mode_lib->mp.MacroTileHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10864
calculate_mcache_setting_params->bytes_per_pixel_c = mode_lib->mp.BytePerPixelC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10867
calculate_mcache_setting_params->dcc_dram_bw_nom_overhead_factor_l = &mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p0[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10868
calculate_mcache_setting_params->dcc_dram_bw_pref_overhead_factor_l = &mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p0[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10869
calculate_mcache_setting_params->dcc_dram_bw_nom_overhead_factor_c = &mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p1[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10870
calculate_mcache_setting_params->dcc_dram_bw_pref_overhead_factor_c = &mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p1[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10872
calculate_mcache_setting_params->num_mcaches_l = &mode_lib->mp.num_mcaches_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10873
calculate_mcache_setting_params->mcache_row_bytes_l = &mode_lib->mp.mcache_row_bytes_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10874
calculate_mcache_setting_params->mcache_row_bytes_per_channel_l = &mode_lib->mp.mcache_row_bytes_per_channel_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10875
calculate_mcache_setting_params->mcache_offsets_l = mode_lib->mp.mcache_offsets_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10876
calculate_mcache_setting_params->mcache_shift_granularity_l = &mode_lib->mp.mcache_shift_granularity_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10878
calculate_mcache_setting_params->num_mcaches_c = &mode_lib->mp.num_mcaches_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10879
calculate_mcache_setting_params->mcache_row_bytes_c = &mode_lib->mp.mcache_row_bytes_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10880
calculate_mcache_setting_params->mcache_row_bytes_per_channel_c = &mode_lib->mp.mcache_row_bytes_per_channel_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10881
calculate_mcache_setting_params->mcache_offsets_c = mode_lib->mp.mcache_offsets_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10882
calculate_mcache_setting_params->mcache_shift_granularity_c = &mode_lib->mp.mcache_shift_granularity_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10884
calculate_mcache_setting_params->mall_comb_mcache_l = &mode_lib->mp.mall_comb_mcache_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10885
calculate_mcache_setting_params->mall_comb_mcache_c = &mode_lib->mp.mall_comb_mcache_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10886
calculate_mcache_setting_params->lc_comb_mcache = &mode_lib->mp.lc_comb_mcache[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10891
mode_lib->mp.mall_prefetch_sdp_overhead_factor,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10892
mode_lib->mp.mall_prefetch_dram_overhead_factor,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10901
mode_lib->mp.avg_bandwidth_available_min,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10902
mode_lib->mp.avg_bandwidth_available,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10903
mode_lib->mp.urg_bandwidth_available_min,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10904
mode_lib->mp.urg_bandwidth_available,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10905
mode_lib->mp.urg_bandwidth_available_vm_only,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10906
mode_lib->mp.urg_bandwidth_available_pixel_and_vm,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10910
mode_lib->mp.Dcfclk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10911
mode_lib->mp.FabricClock,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10912
mode_lib->mp.dram_bw_mbps);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10923
mode_lib->mp.urg_bandwidth_available_pixel_and_vm[dml2_core_internal_soc_state_sys_active],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10924
mode_lib->mp.urg_bandwidth_available_vm_only[dml2_core_internal_soc_state_sys_active]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10929
s->TotalActiveDPP = s->TotalActiveDPP + mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10931
s->TotalDCCActiveDPP = s->TotalDCCActiveDPP + mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10935
calculate_tdlut_setting_params->dispclk_mhz = mode_lib->mp.Dispclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10964
mode_lib->mp.Dcfclk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10965
mode_lib->mp.FabricClock,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10967
mode_lib->mp.urg_bandwidth_available_min[dml2_core_internal_soc_state_sys_active],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10969
mode_lib->mp.NoOfDPP,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10970
mode_lib->mp.dpte_group_bytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10978
mode_lib->mp.request_size_bytes_luma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10979
mode_lib->mp.request_size_bytes_chroma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10982
mode_lib->mp.TripToMemory,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10986
&mode_lib->mp.ExtraLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10987
&mode_lib->mp.ExtraLatency_sr,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10988
&mode_lib->mp.ExtraLatencyPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10990
mode_lib->mp.TCalc = 24.0 / mode_lib->mp.DCFCLKDeepSleep;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
10994
mode_lib->mp.WritebackDelay[k] =
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11004
display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total) / mode_lib->mp.Dispclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11006
mode_lib->mp.WritebackDelay[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11014
calculate_bytes_to_fetch_required_to_hide_latency_params->num_of_dpp = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11015
calculate_bytes_to_fetch_required_to_hide_latency_params->meta_row_height_l = mode_lib->mp.meta_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11016
calculate_bytes_to_fetch_required_to_hide_latency_params->meta_row_height_c = mode_lib->mp.meta_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11019
calculate_bytes_to_fetch_required_to_hide_latency_params->dpte_row_height_l = mode_lib->mp.dpte_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11020
calculate_bytes_to_fetch_required_to_hide_latency_params->dpte_row_height_c = mode_lib->mp.dpte_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11023
calculate_bytes_to_fetch_required_to_hide_latency_params->byte_per_pix_l = mode_lib->mp.BytePerPixelY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11024
calculate_bytes_to_fetch_required_to_hide_latency_params->byte_per_pix_c = mode_lib->mp.BytePerPixelC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11025
calculate_bytes_to_fetch_required_to_hide_latency_params->swath_width_l = mode_lib->mp.SwathWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11026
calculate_bytes_to_fetch_required_to_hide_latency_params->swath_width_c = mode_lib->mp.SwathWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11027
calculate_bytes_to_fetch_required_to_hide_latency_params->swath_height_l = mode_lib->mp.SwathHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11028
calculate_bytes_to_fetch_required_to_hide_latency_params->swath_height_c = mode_lib->mp.SwathHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11044
mode_lib->mp.excess_vactive_fill_bw_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11045
mode_lib->mp.excess_vactive_fill_bw_c);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11047
mode_lib->mp.UrgentLatency = CalculateUrgentLatency(
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11054
mode_lib->mp.FabricClock,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11055
mode_lib->mp.uclk_freq_mhz,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11057
mode_lib->soc.qos_parameters.qos_params.dcn4x.per_uclk_dpm_params[mode_lib->mp.qos_param_index].urgent_ramp_uclk_cycles,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11064
mode_lib->mp.TripToMemory = CalculateTripToMemory(
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11065
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11066
mode_lib->mp.FabricClock,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11067
mode_lib->mp.uclk_freq_mhz,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11069
mode_lib->soc.qos_parameters.qos_params.dcn4x.per_uclk_dpm_params[mode_lib->mp.qos_param_index].trip_to_memory_uclk_cycles,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11075
mode_lib->mp.TripToMemory = math_max2(mode_lib->mp.UrgentLatency, mode_lib->mp.TripToMemory);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11077
mode_lib->mp.MetaTripToMemory = CalculateMetaTripToMemory(
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11078
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11079
mode_lib->mp.FabricClock,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11080
mode_lib->mp.uclk_freq_mhz,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11082
mode_lib->soc.qos_parameters.qos_params.dcn4x.per_uclk_dpm_params[mode_lib->mp.qos_param_index].meta_trip_to_memory_uclk_cycles,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11095
mode_lib->mp.NoOfDPP[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11102
mode_lib->mp.NoOfDPP[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11125
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11128
&mode_lib->mp.UrgentBurstFactorCursor[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11131
mode_lib->mp.UrgentBurstFactorCursorPre[k] = mode_lib->mp.UrgentBurstFactorCursor[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11135
mode_lib->mp.swath_width_luma_ub[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11136
mode_lib->mp.swath_width_chroma_ub[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11137
mode_lib->mp.SwathHeightY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11138
mode_lib->mp.SwathHeightC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11140
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11143
mode_lib->mp.BytePerPixelInDETY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11144
mode_lib->mp.BytePerPixelInDETC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11145
mode_lib->mp.DETBufferSizeY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11146
mode_lib->mp.DETBufferSizeC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11149
&mode_lib->mp.UrgentBurstFactorLuma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11150
&mode_lib->mp.UrgentBurstFactorChroma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11151
&mode_lib->mp.NotEnoughUrgentLatencyHiding[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11153
mode_lib->mp.NotEnoughUrgentLatencyHiding[k] = mode_lib->mp.NotEnoughUrgentLatencyHiding[k] || cursor_not_enough_urgent_latency_hiding;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11161
mode_lib->mp.WritebackDelay[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11165
DML_LOG_VERBOSE("DML::%s: k=%u WritebackDelay = %f\n", __func__, k, mode_lib->mp.WritebackDelay[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11184
CheckGlobalPrefetchAdmissibility_params->swath_height_l = mode_lib->mp.SwathHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11185
CheckGlobalPrefetchAdmissibility_params->swath_height_c = mode_lib->mp.SwathHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11187
CheckGlobalPrefetchAdmissibility_params->compressed_buffer_size_kbytes = mode_lib->mp.CompressedBufferSizeInkByte;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11188
CheckGlobalPrefetchAdmissibility_params->detile_buffer_size_bytes_l = mode_lib->mp.DETBufferSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11189
CheckGlobalPrefetchAdmissibility_params->detile_buffer_size_bytes_c = mode_lib->mp.DETBufferSizeC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11195
CheckGlobalPrefetchAdmissibility_params->estimated_urg_bandwidth_required_mbps = mode_lib->mp.urg_bandwidth_available[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11196
CheckGlobalPrefetchAdmissibility_params->estimated_dcfclk_mhz = mode_lib->mp.Dcfclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11198
CheckGlobalPrefetchAdmissibility_params->dst_y_prefetch = mode_lib->mp.dst_y_prefetch;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11216
mode_lib->mp.TWait[k] = CalculateTWait(
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11218
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11219
mode_lib->mp.TripToMemory,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11221
get_g6_temp_read_blackout_us(&mode_lib->soc, (unsigned int)(mode_lib->mp.uclk_freq_mhz * 1000), in_out_params->min_clk_index) : 0.0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11223
myPipe->Dppclk = mode_lib->mp.Dppclk[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11224
myPipe->Dispclk = mode_lib->mp.Dispclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11226
myPipe->DCFClkDeepSleep = mode_lib->mp.DCFCLKDeepSleep;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11227
myPipe->DPPPerSurface = mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11235
myPipe->BlockWidth256BytesY = mode_lib->mp.Read256BlockWidthY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11236
myPipe->BlockHeight256BytesY = mode_lib->mp.Read256BlockHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11237
myPipe->BlockWidth256BytesC = mode_lib->mp.Read256BlockWidthC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11238
myPipe->BlockHeight256BytesC = mode_lib->mp.Read256BlockHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11245
myPipe->ODMMode = mode_lib->mp.ODMMode[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11247
myPipe->BytePerPixelY = mode_lib->mp.BytePerPixelY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11248
myPipe->BytePerPixelC = mode_lib->mp.BytePerPixelC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11257
CalculatePrefetchSchedule_params->DSCDelay = mode_lib->mp.DSCDelay[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11263
CalculatePrefetchSchedule_params->DPP_RECOUT_WIDTH = (unsigned int)(mode_lib->mp.SwathWidthY[k] / display_cfg->plane_descriptors[k].composition.scaler_info.plane0.h_ratio);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11272
CalculatePrefetchSchedule_params->UrgentLatency = mode_lib->mp.UrgentLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11273
CalculatePrefetchSchedule_params->ExtraLatencyPrefetch = mode_lib->mp.ExtraLatencyPrefetch;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11274
CalculatePrefetchSchedule_params->TCalc = mode_lib->mp.TCalc;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11275
CalculatePrefetchSchedule_params->vm_bytes = mode_lib->mp.vm_bytes[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11276
CalculatePrefetchSchedule_params->PixelPTEBytesPerRow = mode_lib->mp.PixelPTEBytesPerRow[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11277
CalculatePrefetchSchedule_params->PrefetchSourceLinesY = mode_lib->mp.PrefetchSourceLinesY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11278
CalculatePrefetchSchedule_params->VInitPreFillY = mode_lib->mp.VInitPreFillY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11279
CalculatePrefetchSchedule_params->MaxNumSwathY = mode_lib->mp.MaxNumSwathY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11280
CalculatePrefetchSchedule_params->PrefetchSourceLinesC = mode_lib->mp.PrefetchSourceLinesC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11281
CalculatePrefetchSchedule_params->VInitPreFillC = mode_lib->mp.VInitPreFillC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11282
CalculatePrefetchSchedule_params->MaxNumSwathC = mode_lib->mp.MaxNumSwathC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11283
CalculatePrefetchSchedule_params->swath_width_luma_ub = mode_lib->mp.swath_width_luma_ub[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11284
CalculatePrefetchSchedule_params->swath_width_chroma_ub = mode_lib->mp.swath_width_chroma_ub[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11285
CalculatePrefetchSchedule_params->SwathHeightY = mode_lib->mp.SwathHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11286
CalculatePrefetchSchedule_params->SwathHeightC = mode_lib->mp.SwathHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11287
CalculatePrefetchSchedule_params->TWait = mode_lib->mp.TWait[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11288
CalculatePrefetchSchedule_params->Ttrip = mode_lib->mp.TripToMemory;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11289
CalculatePrefetchSchedule_params->Turg = mode_lib->mp.UrgentLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11300
CalculatePrefetchSchedule_params->meta_row_bytes = mode_lib->mp.meta_row_bytes[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11301
CalculatePrefetchSchedule_params->mall_prefetch_sdp_overhead_factor = mode_lib->mp.mall_prefetch_sdp_overhead_factor[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11303
CalculatePrefetchSchedule_params->vactive_sw_bw_l = mode_lib->mp.vactive_sw_bw_l[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11304
CalculatePrefetchSchedule_params->vactive_sw_bw_c = mode_lib->mp.vactive_sw_bw_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11307
CalculatePrefetchSchedule_params->DSTXAfterScaler = &mode_lib->mp.DSTXAfterScaler[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11308
CalculatePrefetchSchedule_params->DSTYAfterScaler = &mode_lib->mp.DSTYAfterScaler[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11309
CalculatePrefetchSchedule_params->dst_y_prefetch = &mode_lib->mp.dst_y_prefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11310
CalculatePrefetchSchedule_params->dst_y_per_vm_vblank = &mode_lib->mp.dst_y_per_vm_vblank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11311
CalculatePrefetchSchedule_params->dst_y_per_row_vblank = &mode_lib->mp.dst_y_per_row_vblank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11312
CalculatePrefetchSchedule_params->VRatioPrefetchY = &mode_lib->mp.VRatioPrefetchY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11313
CalculatePrefetchSchedule_params->VRatioPrefetchC = &mode_lib->mp.VRatioPrefetchC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11314
CalculatePrefetchSchedule_params->RequiredPrefetchPixelDataBWLuma = &mode_lib->mp.RequiredPrefetchPixelDataBWLuma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11315
CalculatePrefetchSchedule_params->RequiredPrefetchPixelDataBWChroma = &mode_lib->mp.RequiredPrefetchPixelDataBWChroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11317
CalculatePrefetchSchedule_params->NotEnoughTimeForDynamicMetadata = &mode_lib->mp.NotEnoughTimeForDynamicMetadata[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11318
CalculatePrefetchSchedule_params->Tno_bw = &mode_lib->mp.Tno_bw[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11319
CalculatePrefetchSchedule_params->Tno_bw_flip = &mode_lib->mp.Tno_bw_flip[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11320
CalculatePrefetchSchedule_params->prefetch_vmrow_bw = &mode_lib->mp.prefetch_vmrow_bw[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11321
CalculatePrefetchSchedule_params->Tdmdl_vm = &mode_lib->mp.Tdmdl_vm[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11322
CalculatePrefetchSchedule_params->Tdmdl = &mode_lib->mp.Tdmdl[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11323
CalculatePrefetchSchedule_params->TSetup = &mode_lib->mp.TSetup[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11330
CalculatePrefetchSchedule_params->VUpdateOffsetPix = &mode_lib->mp.VUpdateOffsetPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11331
CalculatePrefetchSchedule_params->VUpdateWidthPix = &mode_lib->mp.VUpdateWidthPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11332
CalculatePrefetchSchedule_params->VReadyOffsetPix = &mode_lib->mp.VReadyOffsetPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11333
CalculatePrefetchSchedule_params->prefetch_cursor_bw = &mode_lib->mp.prefetch_cursor_bw[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11339
mode_lib->mp.NoTimeToPrefetch[k] = CalculatePrefetchSchedule(&mode_lib->scratch, CalculatePrefetchSchedule_params);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11342
mode_lib->mp.impacted_prefetch_margin_us[k] = (mode_lib->mp.dst_y_prefetch[k] - s->impacted_dst_y_pre[k]) * s->line_times[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11344
mode_lib->mp.impacted_prefetch_margin_us[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11347
DML_LOG_VERBOSE("DML::%s: k=%0u NoTimeToPrefetch=%0d\n", __func__, k, mode_lib->mp.NoTimeToPrefetch[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11349
mode_lib->mp.VStartupMin[k] = s->MaxVStartupLines[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11352
mode_lib->mp.PrefetchModeSupported = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11354
if (mode_lib->mp.NoTimeToPrefetch[k] == true ||
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11355
mode_lib->mp.NotEnoughTimeForDynamicMetadata[k] ||
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11356
mode_lib->mp.DSTYAfterScaler[k] > 8) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11357
DML_LOG_VERBOSE("DML::%s: k=%u, NoTimeToPrefetch = %0d\n", __func__, k, mode_lib->mp.NoTimeToPrefetch[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11358
DML_LOG_VERBOSE("DML::%s: k=%u, NotEnoughTimeForDynamicMetadata=%u\n", __func__, k, mode_lib->mp.NotEnoughTimeForDynamicMetadata[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11359
DML_LOG_VERBOSE("DML::%s: k=%u, DSTYAfterScaler=%u (should be <= 0)\n", __func__, k, mode_lib->mp.DSTYAfterScaler[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11360
mode_lib->mp.PrefetchModeSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11362
if (mode_lib->mp.dst_y_prefetch[k] < 2)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11365
if (mode_lib->mp.VRatioPrefetchY[k] > __DML2_CALCS_MAX_VRATIO_PRE__ ||
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11366
mode_lib->mp.VRatioPrefetchC[k] > __DML2_CALCS_MAX_VRATIO_PRE__) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11368
DML_LOG_VERBOSE("DML::%s: k=%d, VRatioPrefetchY=%f (should not be < %f)\n", __func__, k, mode_lib->mp.VRatioPrefetchY[k], __DML2_CALCS_MAX_VRATIO_PRE__);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11369
DML_LOG_VERBOSE("DML::%s: k=%d, VRatioPrefetchC=%f (should not be < %f)\n", __func__, k, mode_lib->mp.VRatioPrefetchC[k], __DML2_CALCS_MAX_VRATIO_PRE__);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11373
if (mode_lib->mp.NotEnoughUrgentLatencyHiding[k]) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11374
DML_LOG_VERBOSE("DML::%s: k=%u, NotEnoughUrgentLatencyHiding = %u\n", __func__, k, mode_lib->mp.NotEnoughUrgentLatencyHiding[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11375
mode_lib->mp.PrefetchModeSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11382
mode_lib->mp.PrefetchModeSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11386
mode_lib->mp.PrefetchModeSupported ? "" : "NOT ", CalculatePrefetchSchedule_params->VStartup);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11389
if (mode_lib->mp.PrefetchModeSupported == true) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11395
mode_lib->mp.swath_width_luma_ub[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11396
mode_lib->mp.swath_width_chroma_ub[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11397
mode_lib->mp.SwathHeightY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11398
mode_lib->mp.SwathHeightC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11400
mode_lib->mp.UrgentLatency,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11401
mode_lib->mp.VRatioPrefetchY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11402
mode_lib->mp.VRatioPrefetchC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11403
mode_lib->mp.BytePerPixelInDETY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11404
mode_lib->mp.BytePerPixelInDETC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11405
mode_lib->mp.DETBufferSizeY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11406
mode_lib->mp.DETBufferSizeC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11408
&mode_lib->mp.UrgentBurstFactorLumaPre[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11409
&mode_lib->mp.UrgentBurstFactorChromaPre[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11410
&mode_lib->mp.NotEnoughUrgentLatencyHidingPre[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11413
DML_LOG_VERBOSE("DML::%s: k=%0u DPPPerSurface=%u\n", __func__, k, mode_lib->mp.NoOfDPP[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11414
DML_LOG_VERBOSE("DML::%s: k=%0u UrgentBurstFactorLuma=%f\n", __func__, k, mode_lib->mp.UrgentBurstFactorLuma[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11415
DML_LOG_VERBOSE("DML::%s: k=%0u UrgentBurstFactorChroma=%f\n", __func__, k, mode_lib->mp.UrgentBurstFactorChroma[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11416
DML_LOG_VERBOSE("DML::%s: k=%0u UrgentBurstFactorLumaPre=%f\n", __func__, k, mode_lib->mp.UrgentBurstFactorLumaPre[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11417
DML_LOG_VERBOSE("DML::%s: k=%0u UrgentBurstFactorChromaPre=%f\n", __func__, k, mode_lib->mp.UrgentBurstFactorChromaPre[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11419
DML_LOG_VERBOSE("DML::%s: k=%0u VRatioPrefetchY=%f\n", __func__, k, mode_lib->mp.VRatioPrefetchY[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11422
DML_LOG_VERBOSE("DML::%s: k=%0u prefetch_vmrow_bw=%f\n", __func__, k, mode_lib->mp.prefetch_vmrow_bw[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11423
DML_LOG_VERBOSE("DML::%s: k=%0u vactive_sw_bw_l=%f\n", __func__, k, mode_lib->mp.vactive_sw_bw_l[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11424
DML_LOG_VERBOSE("DML::%s: k=%0u vactive_sw_bw_c=%f\n", __func__, k, mode_lib->mp.vactive_sw_bw_c[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11425
DML_LOG_VERBOSE("DML::%s: k=%0u cursor_bw=%f\n", __func__, k, mode_lib->mp.cursor_bw[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11426
DML_LOG_VERBOSE("DML::%s: k=%0u dpte_row_bw=%f\n", __func__, k, mode_lib->mp.dpte_row_bw[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11427
DML_LOG_VERBOSE("DML::%s: k=%0u meta_row_bw=%f\n", __func__, k, mode_lib->mp.meta_row_bw[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11428
DML_LOG_VERBOSE("DML::%s: k=%0u RequiredPrefetchPixelDataBWLuma=%f\n", __func__, k, mode_lib->mp.RequiredPrefetchPixelDataBWLuma[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11429
DML_LOG_VERBOSE("DML::%s: k=%0u RequiredPrefetchPixelDataBWChroma=%f\n", __func__, k, mode_lib->mp.RequiredPrefetchPixelDataBWChroma[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11430
DML_LOG_VERBOSE("DML::%s: k=%0u prefetch_cursor_bw=%f\n", __func__, k, mode_lib->mp.prefetch_cursor_bw[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11435
mode_lib->mp.final_flip_bw[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11437
calculate_peak_bandwidth_params->urg_vactive_bandwidth_required = mode_lib->mp.urg_vactive_bandwidth_required;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11438
calculate_peak_bandwidth_params->urg_bandwidth_required = mode_lib->mp.urg_bandwidth_required;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11439
calculate_peak_bandwidth_params->urg_bandwidth_required_qual = mode_lib->mp.urg_bandwidth_required_qual;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11440
calculate_peak_bandwidth_params->non_urg_bandwidth_required = mode_lib->mp.non_urg_bandwidth_required;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11447
calculate_peak_bandwidth_params->num_of_dpp = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11448
calculate_peak_bandwidth_params->dcc_dram_bw_nom_overhead_factor_p0 = mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11449
calculate_peak_bandwidth_params->dcc_dram_bw_nom_overhead_factor_p1 = mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11450
calculate_peak_bandwidth_params->dcc_dram_bw_pref_overhead_factor_p0 = mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11451
calculate_peak_bandwidth_params->dcc_dram_bw_pref_overhead_factor_p1 = mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11452
calculate_peak_bandwidth_params->mall_prefetch_sdp_overhead_factor = mode_lib->mp.mall_prefetch_sdp_overhead_factor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11453
calculate_peak_bandwidth_params->mall_prefetch_dram_overhead_factor = mode_lib->mp.mall_prefetch_dram_overhead_factor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11455
calculate_peak_bandwidth_params->surface_read_bandwidth_l = mode_lib->mp.vactive_sw_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11456
calculate_peak_bandwidth_params->surface_read_bandwidth_c = mode_lib->mp.vactive_sw_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11457
calculate_peak_bandwidth_params->prefetch_bandwidth_l = mode_lib->mp.RequiredPrefetchPixelDataBWLuma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11458
calculate_peak_bandwidth_params->prefetch_bandwidth_c = mode_lib->mp.RequiredPrefetchPixelDataBWChroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11460
calculate_peak_bandwidth_params->excess_vactive_fill_bw_l = mode_lib->mp.excess_vactive_fill_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11461
calculate_peak_bandwidth_params->excess_vactive_fill_bw_c = mode_lib->mp.excess_vactive_fill_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11462
calculate_peak_bandwidth_params->cursor_bw = mode_lib->mp.cursor_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11463
calculate_peak_bandwidth_params->dpte_row_bw = mode_lib->mp.dpte_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11464
calculate_peak_bandwidth_params->meta_row_bw = mode_lib->mp.meta_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11465
calculate_peak_bandwidth_params->prefetch_cursor_bw = mode_lib->mp.prefetch_cursor_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11466
calculate_peak_bandwidth_params->prefetch_vmrow_bw = mode_lib->mp.prefetch_vmrow_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11467
calculate_peak_bandwidth_params->flip_bw = mode_lib->mp.final_flip_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11468
calculate_peak_bandwidth_params->urgent_burst_factor_l = mode_lib->mp.UrgentBurstFactorLuma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11469
calculate_peak_bandwidth_params->urgent_burst_factor_c = mode_lib->mp.UrgentBurstFactorChroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11470
calculate_peak_bandwidth_params->urgent_burst_factor_cursor = mode_lib->mp.UrgentBurstFactorCursor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11471
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_l = mode_lib->mp.UrgentBurstFactorLumaPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11472
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_c = mode_lib->mp.UrgentBurstFactorChromaPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11473
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_cursor = mode_lib->mp.UrgentBurstFactorCursorPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11482
&mode_lib->mp.FractionOfUrgentBandwidth, // double* frac_urg_bandwidth
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11483
&mode_lib->mp.FractionOfUrgentBandwidthMALL, // double* frac_urg_bandwidth_mall
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11485
&mode_lib->mp.PrefetchModeSupported, // prefetch bw ok
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11488
mode_lib->mp.non_urg_bandwidth_required,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11489
mode_lib->mp.urg_vactive_bandwidth_required,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11490
mode_lib->mp.urg_bandwidth_required,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11491
mode_lib->mp.urg_bandwidth_available);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11493
if (!mode_lib->mp.PrefetchModeSupported)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11497
if (mode_lib->mp.NotEnoughUrgentLatencyHidingPre[k]) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11498
DML_LOG_VERBOSE("DML::%s: k=%u, NotEnoughUrgentLatencyHidingPre = %u\n", __func__, k, mode_lib->mp.NotEnoughUrgentLatencyHidingPre[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11499
mode_lib->mp.PrefetchModeSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11505
if (mode_lib->mp.PrefetchModeSupported == true) { // prefetch schedule and prefetch bw ok, now check flip bw
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11507
mode_lib->mp.BandwidthAvailableForImmediateFlip =
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11510
mode_lib->mp.urg_bandwidth_required_qual, // no flip
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11511
mode_lib->mp.urg_bandwidth_available);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11512
mode_lib->mp.TotImmediateFlipBytes = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11516
mode_lib->mp.vm_bytes[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11517
mode_lib->mp.PixelPTEBytesPerRow[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11518
mode_lib->mp.meta_row_bytes[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11522
mode_lib->mp.TotImmediateFlipBytes += s->per_pipe_flip_bytes[k] * mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11525
DML_LOG_VERBOSE("DML::%s: DPPPerSurface = %u\n", __func__, mode_lib->mp.NoOfDPP[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11526
DML_LOG_VERBOSE("DML::%s: vm_bytes = %u\n", __func__, mode_lib->mp.vm_bytes[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11527
DML_LOG_VERBOSE("DML::%s: PixelPTEBytesPerRow = %u\n", __func__, mode_lib->mp.PixelPTEBytesPerRow[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11528
DML_LOG_VERBOSE("DML::%s: meta_row_bytes = %u\n", __func__, mode_lib->mp.meta_row_bytes[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11529
DML_LOG_VERBOSE("DML::%s: TotImmediateFlipBytes = %u\n", __func__, mode_lib->mp.TotImmediateFlipBytes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11543
mode_lib->mp.vm_bytes[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11544
mode_lib->mp.PixelPTEBytesPerRow[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11545
mode_lib->mp.BandwidthAvailableForImmediateFlip,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11546
mode_lib->mp.TotImmediateFlipBytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11551
mode_lib->mp.Tno_bw[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11552
mode_lib->mp.dpte_row_height[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11553
mode_lib->mp.dpte_row_height_chroma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11554
mode_lib->mp.use_one_row_for_frame_flip[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11558
mode_lib->mp.meta_row_bytes[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11559
mode_lib->mp.meta_row_height[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11560
mode_lib->mp.meta_row_height_chroma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11564
&mode_lib->mp.dst_y_per_vm_flip[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11565
&mode_lib->mp.dst_y_per_row_flip[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11566
&mode_lib->mp.final_flip_bw[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11567
&mode_lib->mp.ImmediateFlipSupportedForPipe[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11571
calculate_peak_bandwidth_params->urg_bandwidth_required = mode_lib->mp.urg_bandwidth_required_flip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11573
calculate_peak_bandwidth_params->non_urg_bandwidth_required = mode_lib->mp.non_urg_bandwidth_required_flip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11580
calculate_peak_bandwidth_params->num_of_dpp = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11581
calculate_peak_bandwidth_params->dcc_dram_bw_nom_overhead_factor_p0 = mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11582
calculate_peak_bandwidth_params->dcc_dram_bw_nom_overhead_factor_p1 = mode_lib->mp.dcc_dram_bw_nom_overhead_factor_p1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11583
calculate_peak_bandwidth_params->dcc_dram_bw_pref_overhead_factor_p0 = mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11584
calculate_peak_bandwidth_params->dcc_dram_bw_pref_overhead_factor_p1 = mode_lib->mp.dcc_dram_bw_pref_overhead_factor_p1;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11585
calculate_peak_bandwidth_params->mall_prefetch_sdp_overhead_factor = mode_lib->mp.mall_prefetch_sdp_overhead_factor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11586
calculate_peak_bandwidth_params->mall_prefetch_dram_overhead_factor = mode_lib->mp.mall_prefetch_dram_overhead_factor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11588
calculate_peak_bandwidth_params->surface_read_bandwidth_l = mode_lib->mp.vactive_sw_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11589
calculate_peak_bandwidth_params->surface_read_bandwidth_c = mode_lib->mp.vactive_sw_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11590
calculate_peak_bandwidth_params->prefetch_bandwidth_l = mode_lib->mp.RequiredPrefetchPixelDataBWLuma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11591
calculate_peak_bandwidth_params->prefetch_bandwidth_c = mode_lib->mp.RequiredPrefetchPixelDataBWChroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11592
calculate_peak_bandwidth_params->excess_vactive_fill_bw_l = mode_lib->mp.excess_vactive_fill_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11593
calculate_peak_bandwidth_params->excess_vactive_fill_bw_c = mode_lib->mp.excess_vactive_fill_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11594
calculate_peak_bandwidth_params->cursor_bw = mode_lib->mp.cursor_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11595
calculate_peak_bandwidth_params->dpte_row_bw = mode_lib->mp.dpte_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11596
calculate_peak_bandwidth_params->meta_row_bw = mode_lib->mp.meta_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11597
calculate_peak_bandwidth_params->prefetch_cursor_bw = mode_lib->mp.prefetch_cursor_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11598
calculate_peak_bandwidth_params->prefetch_vmrow_bw = mode_lib->mp.prefetch_vmrow_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11600
calculate_peak_bandwidth_params->flip_bw = mode_lib->mp.final_flip_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11601
calculate_peak_bandwidth_params->urgent_burst_factor_l = mode_lib->mp.UrgentBurstFactorLuma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11602
calculate_peak_bandwidth_params->urgent_burst_factor_c = mode_lib->mp.UrgentBurstFactorChroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11603
calculate_peak_bandwidth_params->urgent_burst_factor_cursor = mode_lib->mp.UrgentBurstFactorCursor;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11604
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_l = mode_lib->mp.UrgentBurstFactorLumaPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11605
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_c = mode_lib->mp.UrgentBurstFactorChromaPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11606
calculate_peak_bandwidth_params->urgent_burst_factor_prefetch_cursor = mode_lib->mp.UrgentBurstFactorCursorPre;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11613
&mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip, // double* frac_urg_bandwidth_flip
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11614
&mode_lib->mp.ImmediateFlipSupported, // bool* flip_bandwidth_support_ok
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11617
mode_lib->mp.urg_bandwidth_required_flip,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11618
mode_lib->mp.non_urg_bandwidth_required_flip,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11619
mode_lib->mp.urg_bandwidth_available);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11621
if (!mode_lib->mp.ImmediateFlipSupported)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11625
if (display_cfg->plane_descriptors[k].immediate_flip && mode_lib->mp.ImmediateFlipSupportedForPipe[k] == false) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11626
mode_lib->mp.ImmediateFlipSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11633
mode_lib->mp.ImmediateFlipSupported = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11638
mode_lib->mp.PrefetchAndImmediateFlipSupported = (mode_lib->mp.PrefetchModeSupported == true && (!must_support_iflip || mode_lib->mp.ImmediateFlipSupported));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11641
DML_LOG_VERBOSE("DML::%s: PrefetchModeSupported = %u\n", __func__, mode_lib->mp.PrefetchModeSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11645
DML_LOG_VERBOSE("DML::%s: ImmediateFlipSupported = %u\n", __func__, mode_lib->mp.ImmediateFlipSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11646
DML_LOG_VERBOSE("DML::%s: PrefetchAndImmediateFlipSupported = %u\n", __func__, mode_lib->mp.PrefetchAndImmediateFlipSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11654
if (!mode_lib->mp.PrefetchAndImmediateFlipSupported) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11673
mode_lib->mp.Read256BlockHeightY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11674
mode_lib->mp.Read256BlockHeightC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11676
mode_lib->mp.BytePerPixelY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11677
mode_lib->mp.BytePerPixelC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11678
mode_lib->mp.BytePerPixelInDETY[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11679
mode_lib->mp.BytePerPixelInDETC[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11683
&mode_lib->mp.RequestLuma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11684
&mode_lib->mp.RequestChroma[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11685
&mode_lib->mp.DCCYMaxUncompressedBlock[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11686
&mode_lib->mp.DCCCMaxUncompressedBlock[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11687
&mode_lib->mp.DCCYMaxCompressedBlock[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11688
&mode_lib->mp.DCCCMaxCompressedBlock[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11689
&mode_lib->mp.DCCYIndependentBlock[k],
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11690
&mode_lib->mp.DCCCIndependentBlock[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11694
s->mmSOCParameters.UrgentLatency = mode_lib->mp.UrgentLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11695
s->mmSOCParameters.ExtraLatency = mode_lib->mp.ExtraLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11696
s->mmSOCParameters.ExtraLatency_sr = mode_lib->mp.ExtraLatency_sr;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11706
s->mmSOCParameters.g6_temp_read_blackout_us = get_g6_temp_read_blackout_us(&mode_lib->soc, (unsigned int)(mode_lib->mp.uclk_freq_mhz * 1000), in_out_params->min_clk_index);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11707
s->mmSOCParameters.max_urgent_latency_us = get_max_urgent_latency_us(&mode_lib->soc.qos_parameters.qos_params.dcn4x, mode_lib->mp.uclk_freq_mhz, mode_lib->mp.FabricClock, in_out_params->min_clk_index);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11708
s->mmSOCParameters.df_response_time_us = mode_lib->soc.qos_parameters.qos_params.dcn4x.df_qos_response_time_fclk_cycles / mode_lib->mp.FabricClock;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11717
CalculateWatermarks_params->DCFCLK = mode_lib->mp.Dcfclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11720
CalculateWatermarks_params->dpte_group_bytes = mode_lib->mp.dpte_group_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11724
CalculateWatermarks_params->DCFClkDeepSleep = mode_lib->mp.DCFCLKDeepSleep;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11725
CalculateWatermarks_params->DETBufferSizeY = mode_lib->mp.DETBufferSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11726
CalculateWatermarks_params->DETBufferSizeC = mode_lib->mp.DETBufferSizeC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11727
CalculateWatermarks_params->SwathHeightY = mode_lib->mp.SwathHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11728
CalculateWatermarks_params->SwathHeightC = mode_lib->mp.SwathHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11729
CalculateWatermarks_params->SwathWidthY = mode_lib->mp.SwathWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11730
CalculateWatermarks_params->SwathWidthC = mode_lib->mp.SwathWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11731
CalculateWatermarks_params->BytePerPixelDETY = mode_lib->mp.BytePerPixelInDETY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11732
CalculateWatermarks_params->BytePerPixelDETC = mode_lib->mp.BytePerPixelInDETC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11733
CalculateWatermarks_params->DSTXAfterScaler = mode_lib->mp.DSTXAfterScaler;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11734
CalculateWatermarks_params->DSTYAfterScaler = mode_lib->mp.DSTYAfterScaler;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11735
CalculateWatermarks_params->UnboundedRequestEnabled = mode_lib->mp.UnboundedRequestEnabled;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11736
CalculateWatermarks_params->CompressedBufferSizeInkByte = mode_lib->mp.CompressedBufferSizeInkByte;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11737
CalculateWatermarks_params->meta_row_height_l = mode_lib->mp.meta_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11738
CalculateWatermarks_params->meta_row_height_c = mode_lib->mp.meta_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11739
CalculateWatermarks_params->DPPPerSurface = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11742
CalculateWatermarks_params->Watermark = &mode_lib->mp.Watermark;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11743
CalculateWatermarks_params->DRAMClockChangeSupport = mode_lib->mp.DRAMClockChangeSupport;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11744
CalculateWatermarks_params->global_dram_clock_change_supported = &mode_lib->mp.global_dram_clock_change_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11745
CalculateWatermarks_params->MaxActiveDRAMClockChangeLatencySupported = mode_lib->mp.MaxActiveDRAMClockChangeLatencySupported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11746
CalculateWatermarks_params->SubViewportLinesNeededInMALL = mode_lib->mp.SubViewportLinesNeededInMALL;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11747
CalculateWatermarks_params->FCLKChangeSupport = mode_lib->mp.FCLKChangeSupport;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11748
CalculateWatermarks_params->global_fclk_change_supported = &mode_lib->mp.global_fclk_change_supported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11749
CalculateWatermarks_params->MaxActiveFCLKChangeLatencySupported = &mode_lib->mp.MaxActiveFCLKChangeLatencySupported;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11750
CalculateWatermarks_params->USRRetrainingSupport = &mode_lib->mp.USRRetrainingSupport;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11751
CalculateWatermarks_params->g6_temp_read_support = &mode_lib->mp.g6_temp_read_support;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11759
mode_lib->mp.WritebackAllowDRAMClockChangeEndPosition[k] = math_max2(0, mode_lib->mp.VStartupMin[k] * display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total /
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11760
((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000) - mode_lib->mp.Watermark.WritebackDRAMClockChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11761
mode_lib->mp.WritebackAllowFCLKChangeEndPosition[k] = math_max2(0, mode_lib->mp.VStartupMin[k] * display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total /
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11762
((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000) - mode_lib->mp.Watermark.WritebackFCLKChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11764
mode_lib->mp.WritebackAllowDRAMClockChangeEndPosition[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11765
mode_lib->mp.WritebackAllowFCLKChangeEndPosition[k] = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11769
calculate_pstate_keepout_dst_lines(display_cfg, &mode_lib->mp.Watermark, mode_lib->mp.pstate_keepout_dst_lines);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11779
mode_lib->mp.VRatioPrefetchY,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11780
mode_lib->mp.VRatioPrefetchC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11781
mode_lib->mp.swath_width_luma_ub,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11782
mode_lib->mp.swath_width_chroma_ub,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11783
mode_lib->mp.PSCL_THROUGHPUT,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11784
mode_lib->mp.PSCL_THROUGHPUT_CHROMA,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11785
mode_lib->mp.Dppclk,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11786
mode_lib->mp.BytePerPixelC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11787
mode_lib->mp.req_per_swath_ub_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11788
mode_lib->mp.req_per_swath_ub_c,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11791
mode_lib->mp.DisplayPipeLineDeliveryTimeLuma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11792
mode_lib->mp.DisplayPipeLineDeliveryTimeChroma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11793
mode_lib->mp.DisplayPipeLineDeliveryTimeLumaPrefetch,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11794
mode_lib->mp.DisplayPipeLineDeliveryTimeChromaPrefetch,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11795
mode_lib->mp.DisplayPipeRequestDeliveryTimeLuma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11796
mode_lib->mp.DisplayPipeRequestDeliveryTimeChroma,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11797
mode_lib->mp.DisplayPipeRequestDeliveryTimeLumaPrefetch,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11798
mode_lib->mp.DisplayPipeRequestDeliveryTimeChromaPrefetch);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11803
CalculateMetaAndPTETimes_params->use_one_row_for_frame = mode_lib->mp.use_one_row_for_frame;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11804
CalculateMetaAndPTETimes_params->dst_y_per_row_vblank = mode_lib->mp.dst_y_per_row_vblank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11805
CalculateMetaAndPTETimes_params->dst_y_per_row_flip = mode_lib->mp.dst_y_per_row_flip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11806
CalculateMetaAndPTETimes_params->BytePerPixelY = mode_lib->mp.BytePerPixelY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11807
CalculateMetaAndPTETimes_params->BytePerPixelC = mode_lib->mp.BytePerPixelC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11808
CalculateMetaAndPTETimes_params->dpte_row_height = mode_lib->mp.dpte_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11809
CalculateMetaAndPTETimes_params->dpte_row_height_chroma = mode_lib->mp.dpte_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11810
CalculateMetaAndPTETimes_params->dpte_group_bytes = mode_lib->mp.dpte_group_bytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11811
CalculateMetaAndPTETimes_params->PTERequestSizeY = mode_lib->mp.PTERequestSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11812
CalculateMetaAndPTETimes_params->PTERequestSizeC = mode_lib->mp.PTERequestSizeC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11813
CalculateMetaAndPTETimes_params->PixelPTEReqWidthY = mode_lib->mp.PixelPTEReqWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11814
CalculateMetaAndPTETimes_params->PixelPTEReqHeightY = mode_lib->mp.PixelPTEReqHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11815
CalculateMetaAndPTETimes_params->PixelPTEReqWidthC = mode_lib->mp.PixelPTEReqWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11816
CalculateMetaAndPTETimes_params->PixelPTEReqHeightC = mode_lib->mp.PixelPTEReqHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11817
CalculateMetaAndPTETimes_params->dpte_row_width_luma_ub = mode_lib->mp.dpte_row_width_luma_ub;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11818
CalculateMetaAndPTETimes_params->dpte_row_width_chroma_ub = mode_lib->mp.dpte_row_width_chroma_ub;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11824
CalculateMetaAndPTETimes_params->meta_row_width = mode_lib->mp.meta_row_width;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11825
CalculateMetaAndPTETimes_params->meta_row_width_chroma = mode_lib->mp.meta_row_width_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11826
CalculateMetaAndPTETimes_params->meta_row_height = mode_lib->mp.meta_row_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11827
CalculateMetaAndPTETimes_params->meta_row_height_chroma = mode_lib->mp.meta_row_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11828
CalculateMetaAndPTETimes_params->meta_req_width = mode_lib->mp.meta_req_width;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11829
CalculateMetaAndPTETimes_params->meta_req_width_chroma = mode_lib->mp.meta_req_width_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11830
CalculateMetaAndPTETimes_params->meta_req_height = mode_lib->mp.meta_req_height;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11831
CalculateMetaAndPTETimes_params->meta_req_height_chroma = mode_lib->mp.meta_req_height_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11833
CalculateMetaAndPTETimes_params->time_per_tdlut_group = mode_lib->mp.time_per_tdlut_group;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11834
CalculateMetaAndPTETimes_params->DST_Y_PER_PTE_ROW_NOM_L = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_L;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11835
CalculateMetaAndPTETimes_params->DST_Y_PER_PTE_ROW_NOM_C = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_C;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11836
CalculateMetaAndPTETimes_params->time_per_pte_group_nom_luma = mode_lib->mp.time_per_pte_group_nom_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11837
CalculateMetaAndPTETimes_params->time_per_pte_group_vblank_luma = mode_lib->mp.time_per_pte_group_vblank_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11838
CalculateMetaAndPTETimes_params->time_per_pte_group_flip_luma = mode_lib->mp.time_per_pte_group_flip_luma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11839
CalculateMetaAndPTETimes_params->time_per_pte_group_nom_chroma = mode_lib->mp.time_per_pte_group_nom_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11840
CalculateMetaAndPTETimes_params->time_per_pte_group_vblank_chroma = mode_lib->mp.time_per_pte_group_vblank_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11841
CalculateMetaAndPTETimes_params->time_per_pte_group_flip_chroma = mode_lib->mp.time_per_pte_group_flip_chroma;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11842
CalculateMetaAndPTETimes_params->DST_Y_PER_META_ROW_NOM_L = mode_lib->mp.DST_Y_PER_META_ROW_NOM_L;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11843
CalculateMetaAndPTETimes_params->DST_Y_PER_META_ROW_NOM_C = mode_lib->mp.DST_Y_PER_META_ROW_NOM_C;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11844
CalculateMetaAndPTETimes_params->TimePerMetaChunkNominal = mode_lib->mp.TimePerMetaChunkNominal;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11845
CalculateMetaAndPTETimes_params->TimePerChromaMetaChunkNominal = mode_lib->mp.TimePerChromaMetaChunkNominal;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11846
CalculateMetaAndPTETimes_params->TimePerMetaChunkVBlank = mode_lib->mp.TimePerMetaChunkVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11847
CalculateMetaAndPTETimes_params->TimePerChromaMetaChunkVBlank = mode_lib->mp.TimePerChromaMetaChunkVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11848
CalculateMetaAndPTETimes_params->TimePerMetaChunkFlip = mode_lib->mp.TimePerMetaChunkFlip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11849
CalculateMetaAndPTETimes_params->TimePerChromaMetaChunkFlip = mode_lib->mp.TimePerChromaMetaChunkFlip;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11856
mode_lib->mp.BytePerPixelC,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11857
mode_lib->mp.dst_y_per_vm_vblank,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11858
mode_lib->mp.dst_y_per_vm_flip,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11859
mode_lib->mp.dpte_row_width_luma_ub,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11860
mode_lib->mp.dpte_row_width_chroma_ub,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11861
mode_lib->mp.vm_group_bytes,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11862
mode_lib->mp.dpde0_bytes_per_frame_ub_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11863
mode_lib->mp.dpde0_bytes_per_frame_ub_c,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11865
mode_lib->mp.meta_pte_bytes_per_frame_ub_l,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11866
mode_lib->mp.meta_pte_bytes_per_frame_ub_c,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11870
mode_lib->mp.TimePerVMGroupVBlank,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11871
mode_lib->mp.TimePerVMGroupFlip,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11872
mode_lib->mp.TimePerVMRequestVBlank,
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11873
mode_lib->mp.TimePerVMRequestFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11879
mode_lib->mp.MinTTUVBlank[k] = mode_lib->mp.TWait[k] + mode_lib->mp.ExtraLatency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11881
mode_lib->mp.MinTTUVBlank[k] = mode_lib->mp.TCalc + mode_lib->mp.MinTTUVBlank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11884
DML_LOG_VERBOSE("DML::%s: k=%u, MinTTUVBlank = %f (before vstartup margin)\n", __func__, k, mode_lib->mp.MinTTUVBlank[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11886
s->Tvstartup_margin = (s->MaxVStartupLines[k] - mode_lib->mp.VStartupMin[k]) * display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total / ((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11887
mode_lib->mp.MinTTUVBlank[k] = mode_lib->mp.MinTTUVBlank[k] + s->Tvstartup_margin;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11892
DML_LOG_VERBOSE("DML::%s: k=%u, MinTTUVBlank = %f\n", __func__, k, mode_lib->mp.MinTTUVBlank[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11895
mode_lib->mp.Tdmdl[k] = mode_lib->mp.Tdmdl[k] + s->Tvstartup_margin;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11897
mode_lib->mp.Tdmdl_vm[k] = mode_lib->mp.Tdmdl_vm[k] + s->Tvstartup_margin;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11903
mode_lib->mp.VStartup[k] = (isInterlaceTiming ? (2 * s->MaxVStartupLines[k]) : s->MaxVStartupLines[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11907
s->LSetup = math_floor2(4.0 * mode_lib->mp.TSetup[k] / ((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total / ((double)display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.pixel_clock_khz / 1000)), 1.0) / 4.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11908
s->blank_lines_remaining = (display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_total - display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_active) - mode_lib->mp.VStartup[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11915
mode_lib->mp.MIN_DST_Y_NEXT_START[k] = s->dlg_vblank_start + s->blank_lines_remaining + s->LSetup;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11918
if (((mode_lib->mp.VUpdateOffsetPix[k] + mode_lib->mp.VUpdateWidthPix[k] + mode_lib->mp.VReadyOffsetPix[k]) / (double) display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.h_total) <=
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11920
math_floor2((display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_total - display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_active - display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_front_porch - mode_lib->mp.VStartup[k]) / 2.0, 1.0) :
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11921
(int)(display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_total - display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_active - display_cfg->stream_descriptors[display_cfg->plane_descriptors[k].stream_index].timing.v_front_porch - mode_lib->mp.VStartup[k]))) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11922
mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC[k] = true;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11924
mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC[k] = false;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11927
DML_LOG_VERBOSE("DML::%s: k=%u, VStartup = %u (max)\n", __func__, k, mode_lib->mp.VStartup[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11928
DML_LOG_VERBOSE("DML::%s: k=%u, VStartupMin = %u (max)\n", __func__, k, mode_lib->mp.VStartupMin[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11929
DML_LOG_VERBOSE("DML::%s: k=%u, VUpdateOffsetPix = %u\n", __func__, k, mode_lib->mp.VUpdateOffsetPix[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11930
DML_LOG_VERBOSE("DML::%s: k=%u, VUpdateWidthPix = %u\n", __func__, k, mode_lib->mp.VUpdateWidthPix[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11931
DML_LOG_VERBOSE("DML::%s: k=%u, VReadyOffsetPix = %u\n", __func__, k, mode_lib->mp.VReadyOffsetPix[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11936
DML_LOG_VERBOSE("DML::%s: k=%u, TSetup = %f\n", __func__, k, mode_lib->mp.TSetup[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11937
DML_LOG_VERBOSE("DML::%s: k=%u, MIN_DST_Y_NEXT_START = %f\n", __func__, k, mode_lib->mp.MIN_DST_Y_NEXT_START[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11938
DML_LOG_VERBOSE("DML::%s: k=%u, VREADY_AT_OR_AFTER_VSYNC = %u\n", __func__, k, mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11943
mode_lib->mp.TotalWRBandwidth = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11952
mode_lib->mp.TotalWRBandwidth = mode_lib->mp.TotalWRBandwidth + s->WRBandwidth;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11956
mode_lib->mp.TotalDataReadBandwidth = 0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11958
mode_lib->mp.TotalDataReadBandwidth = mode_lib->mp.TotalDataReadBandwidth + mode_lib->mp.vactive_sw_bw_l[k] + mode_lib->mp.vactive_sw_bw_c[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11960
DML_LOG_VERBOSE("DML::%s: k=%u, TotalDataReadBandwidth = %f\n", __func__, k, mode_lib->mp.TotalDataReadBandwidth);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11961
DML_LOG_VERBOSE("DML::%s: k=%u, vactive_sw_bw_l = %f\n", __func__, k, mode_lib->mp.vactive_sw_bw_l[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11962
DML_LOG_VERBOSE("DML::%s: k=%u, vactive_sw_bw_c = %f\n", __func__, k, mode_lib->mp.vactive_sw_bw_c[k]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11967
CalculateStutterEfficiency_params->CompressedBufferSizeInkByte = mode_lib->mp.CompressedBufferSizeInkByte;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11968
CalculateStutterEfficiency_params->UnboundedRequestEnabled = mode_lib->mp.UnboundedRequestEnabled;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11974
CalculateStutterEfficiency_params->TotalDataReadBandwidth = mode_lib->mp.TotalDataReadBandwidth;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11975
CalculateStutterEfficiency_params->DCFCLK = mode_lib->mp.Dcfclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11976
CalculateStutterEfficiency_params->ReturnBW = mode_lib->mp.urg_bandwidth_available_min[dml2_core_internal_soc_state_sys_active];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11977
CalculateStutterEfficiency_params->CompbufReservedSpace64B = mode_lib->mp.compbuf_reserved_space_64b;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11982
CalculateStutterEfficiency_params->StutterEnterPlusExitWatermark = mode_lib->mp.Watermark.StutterEnterPlusExitWatermark;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11983
CalculateStutterEfficiency_params->Z8StutterEnterPlusExitWatermark = mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11985
CalculateStutterEfficiency_params->MinTTUVBlank = mode_lib->mp.MinTTUVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11986
CalculateStutterEfficiency_params->DPPPerSurface = mode_lib->mp.NoOfDPP;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11987
CalculateStutterEfficiency_params->DETBufferSizeY = mode_lib->mp.DETBufferSizeY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11988
CalculateStutterEfficiency_params->BytePerPixelY = mode_lib->mp.BytePerPixelY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11989
CalculateStutterEfficiency_params->BytePerPixelDETY = mode_lib->mp.BytePerPixelInDETY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11990
CalculateStutterEfficiency_params->SwathWidthY = mode_lib->mp.SwathWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11991
CalculateStutterEfficiency_params->SwathHeightY = mode_lib->mp.SwathHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11992
CalculateStutterEfficiency_params->SwathHeightC = mode_lib->mp.SwathHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11993
CalculateStutterEfficiency_params->BlockHeight256BytesY = mode_lib->mp.Read256BlockHeightY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11994
CalculateStutterEfficiency_params->BlockWidth256BytesY = mode_lib->mp.Read256BlockWidthY;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11995
CalculateStutterEfficiency_params->BlockHeight256BytesC = mode_lib->mp.Read256BlockHeightC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11996
CalculateStutterEfficiency_params->BlockWidth256BytesC = mode_lib->mp.Read256BlockWidthC;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11997
CalculateStutterEfficiency_params->DCCYMaxUncompressedBlock = mode_lib->mp.DCCYMaxUncompressedBlock;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11998
CalculateStutterEfficiency_params->DCCCMaxUncompressedBlock = mode_lib->mp.DCCCMaxUncompressedBlock;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
11999
CalculateStutterEfficiency_params->ReadBandwidthSurfaceLuma = mode_lib->mp.vactive_sw_bw_l;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12000
CalculateStutterEfficiency_params->ReadBandwidthSurfaceChroma = mode_lib->mp.vactive_sw_bw_c;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12001
CalculateStutterEfficiency_params->dpte_row_bw = mode_lib->mp.dpte_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12002
CalculateStutterEfficiency_params->meta_row_bw = mode_lib->mp.meta_row_bw;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12006
CalculateStutterEfficiency_params->StutterEfficiencyNotIncludingVBlank = &mode_lib->mp.StutterEfficiencyNotIncludingVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12007
CalculateStutterEfficiency_params->StutterEfficiency = &mode_lib->mp.StutterEfficiency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12008
CalculateStutterEfficiency_params->NumberOfStutterBurstsPerFrame = &mode_lib->mp.NumberOfStutterBurstsPerFrame;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12009
CalculateStutterEfficiency_params->Z8StutterEfficiencyNotIncludingVBlank = &mode_lib->mp.Z8StutterEfficiencyNotIncludingVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12010
CalculateStutterEfficiency_params->Z8StutterEfficiency = &mode_lib->mp.Z8StutterEfficiency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12011
CalculateStutterEfficiency_params->Z8NumberOfStutterBurstsPerFrame = &mode_lib->mp.Z8NumberOfStutterBurstsPerFrame;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12012
CalculateStutterEfficiency_params->StutterPeriod = &mode_lib->mp.StutterPeriod;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12013
CalculateStutterEfficiency_params->DCHUBBUB_ARB_CSTATE_MAX_CAP_MODE = &mode_lib->mp.DCHUBBUB_ARB_CSTATE_MAX_CAP_MODE;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12023
CalculateStutterEfficiency_params->Z8StutterEfficiencyNotIncludingVBlank = &mode_lib->mp.Z8StutterEfficiencyNotIncludingVBlankBestCase;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12024
CalculateStutterEfficiency_params->Z8StutterEfficiency = &mode_lib->mp.Z8StutterEfficiencyBestCase;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12025
CalculateStutterEfficiency_params->Z8NumberOfStutterBurstsPerFrame = &mode_lib->mp.Z8NumberOfStutterBurstsPerFrameBestCase;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12026
CalculateStutterEfficiency_params->StutterPeriod = &mode_lib->mp.StutterPeriodBestCase;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12031
mode_lib->mp.Z8StutterEfficiencyNotIncludingVBlankBestCase = mode_lib->mp.Z8StutterEfficiencyNotIncludingVBlank;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12032
mode_lib->mp.Z8StutterEfficiencyBestCase = mode_lib->mp.Z8StutterEfficiency;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12033
mode_lib->mp.Z8NumberOfStutterBurstsPerFrameBestCase = mode_lib->mp.Z8NumberOfStutterBurstsPerFrame;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12034
mode_lib->mp.StutterPeriodBestCase = mode_lib->mp.StutterPeriod;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12040
mode_lib->mp.min_return_latency_in_dcfclk = (unsigned int)min_return_latency_in_DCFCLK_cycles;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12041
mode_lib->mp.dcfclk_deep_sleep_hysteresis = (unsigned int)math_max2(32, (double)mode_lib->ip.pixel_chunk_size_kbytes * 1024 * 3 / 4 / 64 - min_return_latency_in_DCFCLK_cycles);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12042
DML_ASSERT(mode_lib->mp.dcfclk_deep_sleep_hysteresis < 256);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12051
DML_LOG_VERBOSE("DML::%s: dcfclk_deep_sleep_hysteresis = %d \n", __func__, mode_lib->mp.dcfclk_deep_sleep_hysteresis);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12054
return (in_out_params->mode_lib->mp.PrefetchAndImmediateFlipSupported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12200
unsigned int plane_idx = mode_lib->mp.pipe_plane[pipe_idx];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12208
wm_regs->fclk_pstate = (int unsigned)(mode_lib->mp.Watermark.FCLKChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12209
wm_regs->sr_enter = (int unsigned)(mode_lib->mp.Watermark.StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12210
wm_regs->sr_exit = (int unsigned)(mode_lib->mp.Watermark.StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12211
wm_regs->sr_enter_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12212
wm_regs->sr_exit_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12213
wm_regs->temp_read_or_ppt = (int unsigned)(mode_lib->mp.Watermark.temp_read_or_ppt_watermark_us * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12214
wm_regs->uclk_pstate = (int unsigned)(mode_lib->mp.Watermark.DRAMClockChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12215
wm_regs->urgent = (int unsigned)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12216
wm_regs->usr = (int unsigned)(mode_lib->mp.Watermark.USRRetrainingWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12217
wm_regs->refcyc_per_trip_to_mem = (unsigned int)(mode_lib->mp.UrgentLatency * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12218
wm_regs->refcyc_per_meta_trip_to_mem = (unsigned int)(mode_lib->mp.MetaTripToMemory * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12219
wm_regs->frac_urg_bw_flip = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12220
wm_regs->frac_urg_bw_nom = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidth * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12221
wm_regs->frac_urg_bw_mall = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthMALL * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12425
l->odm_mode = mode_lib->mp.ODMMode[l->plane_idx];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12457
l->num_active_pipes = mode_lib->mp.num_active_pipes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12490
l->min_ttu_vblank = mode_lib->mp.MinTTUVBlank[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12491
l->min_dst_y_next_start = (unsigned int)(mode_lib->mp.MIN_DST_Y_NEXT_START[mode_lib->mp.pipe_plane[pipe_idx]]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12497
l->vready_after_vcount0 = (unsigned int)(mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC[mode_lib->mp.pipe_plane[pipe_idx]]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12502
l->dst_x_after_scaler = (unsigned int)(mode_lib->mp.DSTXAfterScaler[mode_lib->mp.pipe_plane[pipe_idx]]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12503
l->dst_y_after_scaler = (unsigned int)(mode_lib->mp.DSTYAfterScaler[mode_lib->mp.pipe_plane[pipe_idx]]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12508
l->dst_y_prefetch = mode_lib->mp.dst_y_prefetch[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12509
l->dst_y_per_vm_vblank = mode_lib->mp.dst_y_per_vm_vblank[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12510
l->dst_y_per_row_vblank = mode_lib->mp.dst_y_per_row_vblank[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12511
l->dst_y_per_vm_flip = mode_lib->mp.dst_y_per_vm_flip[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12512
l->dst_y_per_row_flip = mode_lib->mp.dst_y_per_row_flip[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12524
l->vratio_pre_l = mode_lib->mp.VRatioPrefetchY[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12525
l->vratio_pre_c = mode_lib->mp.VRatioPrefetchC[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12531
l->refcyc_per_line_delivery_pre_l = mode_lib->mp.DisplayPipeLineDeliveryTimeLumaPrefetch[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12532
l->refcyc_per_line_delivery_l = mode_lib->mp.DisplayPipeLineDeliveryTimeLuma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12541
l->refcyc_per_line_delivery_pre_c = mode_lib->mp.DisplayPipeLineDeliveryTimeChromaPrefetch[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12542
l->refcyc_per_line_delivery_c = mode_lib->mp.DisplayPipeLineDeliveryTimeChroma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12548
disp_dlg_regs->refcyc_per_vm_dmdata = (unsigned int)(mode_lib->mp.Tdmdl_vm[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12549
disp_dlg_regs->dmdata_dl_delta = (unsigned int)(mode_lib->mp.Tdmdl[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12551
l->refcyc_per_req_delivery_pre_l = mode_lib->mp.DisplayPipeRequestDeliveryTimeLumaPrefetch[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12552
l->refcyc_per_req_delivery_l = mode_lib->mp.DisplayPipeRequestDeliveryTimeLuma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12560
l->refcyc_per_req_delivery_pre_c = mode_lib->mp.DisplayPipeRequestDeliveryTimeChromaPrefetch[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12561
l->refcyc_per_req_delivery_c = mode_lib->mp.DisplayPipeRequestDeliveryTimeChroma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12590
disp_dlg_regs->refcyc_per_vm_group_vblank = (unsigned int)(mode_lib->mp.TimePerVMGroupVBlank[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12591
disp_dlg_regs->refcyc_per_vm_group_flip = (unsigned int)(mode_lib->mp.TimePerVMGroupFlip[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12592
disp_dlg_regs->refcyc_per_vm_req_vblank = (unsigned int)(mode_lib->mp.TimePerVMRequestVBlank[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz * math_pow(2, 10));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12593
disp_dlg_regs->refcyc_per_vm_req_flip = (unsigned int)(mode_lib->mp.TimePerVMRequestFlip[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz * math_pow(2, 10));
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12595
l->dst_y_per_pte_row_nom_l = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_L[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12596
l->dst_y_per_pte_row_nom_c = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_C[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12597
l->refcyc_per_pte_group_nom_l = mode_lib->mp.time_per_pte_group_nom_luma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12598
l->refcyc_per_pte_group_nom_c = mode_lib->mp.time_per_pte_group_nom_chroma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12599
l->refcyc_per_pte_group_vblank_l = mode_lib->mp.time_per_pte_group_vblank_luma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12600
l->refcyc_per_pte_group_vblank_c = mode_lib->mp.time_per_pte_group_vblank_chroma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12601
l->refcyc_per_pte_group_flip_l = mode_lib->mp.time_per_pte_group_flip_luma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12602
l->refcyc_per_pte_group_flip_c = mode_lib->mp.time_per_pte_group_flip_chroma[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12603
l->refcyc_per_tdlut_group = mode_lib->mp.time_per_tdlut_group[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12619
l->dst_y_per_meta_row_nom_l = mode_lib->mp.DST_Y_PER_META_ROW_NOM_L[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12620
l->dst_y_per_meta_row_nom_c = mode_lib->mp.DST_Y_PER_META_ROW_NOM_C[mode_lib->mp.pipe_plane[pipe_idx]];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12621
l->refcyc_per_meta_chunk_nom_l = mode_lib->mp.TimePerMetaChunkNominal[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12622
l->refcyc_per_meta_chunk_nom_c = mode_lib->mp.TimePerChromaMetaChunkNominal[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12623
l->refcyc_per_meta_chunk_vblank_l = mode_lib->mp.TimePerMetaChunkVBlank[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12624
l->refcyc_per_meta_chunk_vblank_c = mode_lib->mp.TimePerChromaMetaChunkVBlank[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12625
l->refcyc_per_meta_chunk_flip_l = mode_lib->mp.TimePerMetaChunkFlip[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12626
l->refcyc_per_meta_chunk_flip_c = mode_lib->mp.TimePerChromaMetaChunkFlip[mode_lib->mp.pipe_plane[pipe_idx]] * l->refclk_freq_in_mhz;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12723
arb_param->compbuf_size = mode_lib->mp.CompressedBufferSizeInkByte / mode_lib->ip.compressed_buffer_segment_size_in_kbytes;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
12838
mode_lib->mp.MinTTUVBlank[i] <= mode_lib->mp.Watermark.DRAMClockChangeWatermark) {
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13152
out->informative.mall.total_surface_size_in_mall_bytes += mode_lib->mp.SurfaceSizeInTheMALL[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13153
out->informative.dpp.total_num_dpps_required += mode_lib->mp.NoOfDPP[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13156
out->informative.qos.min_return_latency_in_dcfclk = mode_lib->mp.min_return_latency_in_dcfclk;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13249
out->informative.misc.min_ttu_vblank_us[k] = mode_lib->mp.MinTTUVBlank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13250
out->informative.mall.subviewport_lines_needed_in_mall[k] = mode_lib->mp.SubViewportLinesNeededInMALL[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13251
out->informative.crb.det_size_in_kbytes[k] = mode_lib->mp.DETBufferSizeInKByte[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13252
out->informative.crb.DETBufferSizeY[k] = mode_lib->mp.DETBufferSizeY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13253
out->informative.misc.ImmediateFlipSupportedForPipe[k] = mode_lib->mp.ImmediateFlipSupportedForPipe[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13254
out->informative.misc.UsesMALLForStaticScreen[k] = mode_lib->mp.is_using_mall_for_ss[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13255
out->informative.plane_info[k].dpte_row_height_plane0 = mode_lib->mp.dpte_row_height[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13256
out->informative.plane_info[k].dpte_row_height_plane1 = mode_lib->mp.dpte_row_height_chroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13257
out->informative.plane_info[k].meta_row_height_plane0 = mode_lib->mp.meta_row_height[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13258
out->informative.plane_info[k].meta_row_height_plane1 = mode_lib->mp.meta_row_height_chroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13259
out->informative.dcc_control[k].max_uncompressed_block_plane0 = mode_lib->mp.DCCYMaxUncompressedBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13260
out->informative.dcc_control[k].max_compressed_block_plane0 = mode_lib->mp.DCCYMaxCompressedBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13261
out->informative.dcc_control[k].independent_block_plane0 = mode_lib->mp.DCCYIndependentBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13262
out->informative.dcc_control[k].max_uncompressed_block_plane1 = mode_lib->mp.DCCCMaxUncompressedBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13263
out->informative.dcc_control[k].max_compressed_block_plane1 = mode_lib->mp.DCCCMaxCompressedBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13264
out->informative.dcc_control[k].independent_block_plane1 = mode_lib->mp.DCCCIndependentBlock[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13265
out->informative.misc.dst_x_after_scaler[k] = mode_lib->mp.DSTXAfterScaler[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13266
out->informative.misc.dst_y_after_scaler[k] = mode_lib->mp.DSTYAfterScaler[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13267
out->informative.misc.prefetch_source_lines_plane0[k] = mode_lib->mp.PrefetchSourceLinesY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13268
out->informative.misc.prefetch_source_lines_plane1[k] = mode_lib->mp.PrefetchSourceLinesC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13269
out->informative.misc.vready_at_or_after_vsync[k] = mode_lib->mp.VREADY_AT_OR_AFTER_VSYNC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13270
out->informative.misc.min_dst_y_next_start[k] = mode_lib->mp.MIN_DST_Y_NEXT_START[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13271
out->informative.plane_info[k].swath_width_plane0 = mode_lib->mp.SwathWidthY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13272
out->informative.plane_info[k].swath_height_plane0 = mode_lib->mp.SwathHeightY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13273
out->informative.plane_info[k].swath_height_plane1 = mode_lib->mp.SwathHeightC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13274
out->informative.misc.CursorDstXOffset[k] = mode_lib->mp.CursorDstXOffset[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13275
out->informative.misc.CursorDstYOffset[k] = mode_lib->mp.CursorDstYOffset[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13276
out->informative.misc.CursorChunkHDLAdjust[k] = mode_lib->mp.CursorChunkHDLAdjust[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13277
out->informative.misc.dpte_group_bytes[k] = mode_lib->mp.dpte_group_bytes[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13278
out->informative.misc.vm_group_bytes[k] = mode_lib->mp.vm_group_bytes[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13279
out->informative.misc.DisplayPipeRequestDeliveryTimeLuma[k] = mode_lib->mp.DisplayPipeRequestDeliveryTimeLuma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13280
out->informative.misc.DisplayPipeRequestDeliveryTimeChroma[k] = mode_lib->mp.DisplayPipeRequestDeliveryTimeChroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13281
out->informative.misc.DisplayPipeRequestDeliveryTimeLumaPrefetch[k] = mode_lib->mp.DisplayPipeRequestDeliveryTimeLumaPrefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13282
out->informative.misc.DisplayPipeRequestDeliveryTimeChromaPrefetch[k] = mode_lib->mp.DisplayPipeRequestDeliveryTimeChromaPrefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13283
out->informative.misc.TimePerVMGroupVBlank[k] = mode_lib->mp.TimePerVMGroupVBlank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13284
out->informative.misc.TimePerVMGroupFlip[k] = mode_lib->mp.TimePerVMGroupFlip[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13285
out->informative.misc.TimePerVMRequestVBlank[k] = mode_lib->mp.TimePerVMRequestVBlank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13286
out->informative.misc.TimePerVMRequestFlip[k] = mode_lib->mp.TimePerVMRequestFlip[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13287
out->informative.misc.Tdmdl_vm[k] = mode_lib->mp.Tdmdl_vm[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13288
out->informative.misc.Tdmdl[k] = mode_lib->mp.Tdmdl[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13289
out->informative.misc.VStartup[k] = mode_lib->mp.VStartup[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13290
out->informative.misc.VUpdateOffsetPix[k] = mode_lib->mp.VUpdateOffsetPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13291
out->informative.misc.VUpdateWidthPix[k] = mode_lib->mp.VUpdateWidthPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13292
out->informative.misc.VReadyOffsetPix[k] = mode_lib->mp.VReadyOffsetPix[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13294
out->informative.misc.DST_Y_PER_PTE_ROW_NOM_L[k] = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_L[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13295
out->informative.misc.DST_Y_PER_PTE_ROW_NOM_C[k] = mode_lib->mp.DST_Y_PER_PTE_ROW_NOM_C[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13296
out->informative.misc.time_per_pte_group_nom_luma[k] = mode_lib->mp.time_per_pte_group_nom_luma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13297
out->informative.misc.time_per_pte_group_nom_chroma[k] = mode_lib->mp.time_per_pte_group_nom_chroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13298
out->informative.misc.time_per_pte_group_vblank_luma[k] = mode_lib->mp.time_per_pte_group_vblank_luma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13299
out->informative.misc.time_per_pte_group_vblank_chroma[k] = mode_lib->mp.time_per_pte_group_vblank_chroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13300
out->informative.misc.time_per_pte_group_flip_luma[k] = mode_lib->mp.time_per_pte_group_flip_luma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13301
out->informative.misc.time_per_pte_group_flip_chroma[k] = mode_lib->mp.time_per_pte_group_flip_chroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13302
out->informative.misc.VRatioPrefetchY[k] = mode_lib->mp.VRatioPrefetchY[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13303
out->informative.misc.VRatioPrefetchC[k] = mode_lib->mp.VRatioPrefetchC[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13304
out->informative.misc.DestinationLinesForPrefetch[k] = mode_lib->mp.dst_y_prefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13305
out->informative.misc.DestinationLinesToRequestVMInVBlank[k] = mode_lib->mp.dst_y_per_vm_vblank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13306
out->informative.misc.DestinationLinesToRequestRowInVBlank[k] = mode_lib->mp.dst_y_per_row_vblank[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13307
out->informative.misc.DestinationLinesToRequestVMInImmediateFlip[k] = mode_lib->mp.dst_y_per_vm_flip[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13308
out->informative.misc.DestinationLinesToRequestRowInImmediateFlip[k] = mode_lib->mp.dst_y_per_row_flip[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13309
out->informative.misc.DisplayPipeLineDeliveryTimeLuma[k] = mode_lib->mp.DisplayPipeLineDeliveryTimeLuma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13310
out->informative.misc.DisplayPipeLineDeliveryTimeChroma[k] = mode_lib->mp.DisplayPipeLineDeliveryTimeChroma[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13311
out->informative.misc.DisplayPipeLineDeliveryTimeLumaPrefetch[k] = mode_lib->mp.DisplayPipeLineDeliveryTimeLumaPrefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13312
out->informative.misc.DisplayPipeLineDeliveryTimeChromaPrefetch[k] = mode_lib->mp.DisplayPipeLineDeliveryTimeChromaPrefetch[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13314
out->informative.misc.WritebackRequiredBandwidth = mode_lib->mp.TotalWRBandwidth / 1000.0;
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13315
out->informative.misc.WritebackAllowDRAMClockChangeEndPosition[k] = mode_lib->mp.WritebackAllowDRAMClockChangeEndPosition[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13316
out->informative.misc.WritebackAllowFCLKChangeEndPosition[k] = mode_lib->mp.WritebackAllowFCLKChangeEndPosition[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13317
out->informative.misc.DSCCLK_calculated[k] = mode_lib->mp.DSCCLK[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13318
out->informative.misc.BIGK_FRAGMENT_SIZE[k] = mode_lib->mp.BIGK_FRAGMENT_SIZE[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13319
out->informative.misc.PTE_BUFFER_MODE[k] = mode_lib->mp.PTE_BUFFER_MODE[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13320
out->informative.misc.DSCDelay[k] = mode_lib->mp.DSCDelay[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13321
out->informative.misc.MaxActiveDRAMClockChangeLatencySupported[k] = mode_lib->mp.MaxActiveDRAMClockChangeLatencySupported[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13323
if (mode_lib->mp.impacted_prefetch_margin_us[k] < out->informative.misc.LowestPrefetchMargin)
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
13324
out->informative.misc.LowestPrefetchMargin = mode_lib->mp.impacted_prefetch_margin_us[k];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
247
unsigned int plane_idx = mode_lib->mp.pipe_plane[pipe_idx];
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
257
plane_idx = mode_lib->mp.pipe_plane[pipe_idx]; \
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
261
dml_get_per_pipe_var_func(dpte_group_size_in_bytes, unsigned int, mode_lib->mp.dpte_group_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
262
dml_get_per_pipe_var_func(vm_group_size_in_bytes, unsigned int, mode_lib->mp.vm_group_bytes);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
263
dml_get_per_pipe_var_func(swath_height_l, unsigned int, mode_lib->mp.SwathHeightY);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
264
dml_get_per_pipe_var_func(swath_height_c, unsigned int, mode_lib->mp.SwathHeightC);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
265
dml_get_per_pipe_var_func(dpte_row_height_linear_l, unsigned int, mode_lib->mp.dpte_row_height_linear);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
266
dml_get_per_pipe_var_func(dpte_row_height_linear_c, unsigned int, mode_lib->mp.dpte_row_height_linear_chroma);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
268
dml_get_per_pipe_var_func(vstartup_calculated, unsigned int, mode_lib->mp.VStartup);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
269
dml_get_per_pipe_var_func(vupdate_offset, unsigned int, mode_lib->mp.VUpdateOffsetPix);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
270
dml_get_per_pipe_var_func(vupdate_width, unsigned int, mode_lib->mp.VUpdateWidthPix);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
271
dml_get_per_pipe_var_func(vready_offset, unsigned int, mode_lib->mp.VReadyOffsetPix);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
272
dml_get_per_pipe_var_func(pstate_keepout_dst_lines, unsigned int, mode_lib->mp.pstate_keepout_dst_lines);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
273
dml_get_per_pipe_var_func(det_stored_buffer_size_l_bytes, unsigned int, mode_lib->mp.DETBufferSizeY);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
274
dml_get_per_pipe_var_func(det_stored_buffer_size_c_bytes, unsigned int, mode_lib->mp.DETBufferSizeC);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
275
dml_get_per_pipe_var_func(det_buffer_size_kbytes, unsigned int, mode_lib->mp.DETBufferSizeInKByte);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
276
dml_get_per_pipe_var_func(surface_size_in_mall_bytes, unsigned int, mode_lib->mp.SurfaceSizeInTheMALL);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
308
dml_get_var_func(wm_urgent, double, mode_lib->mp.Watermark.UrgentWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
309
dml_get_var_func(wm_stutter_exit, double, mode_lib->mp.Watermark.StutterExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
310
dml_get_var_func(wm_stutter_enter_exit, double, mode_lib->mp.Watermark.StutterEnterPlusExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
311
dml_get_var_func(wm_z8_stutter_exit, double, mode_lib->mp.Watermark.Z8StutterExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
312
dml_get_var_func(wm_z8_stutter_enter_exit, double, mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
313
dml_get_var_func(wm_memory_trip, double, mode_lib->mp.UrgentLatency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
314
dml_get_var_func(meta_trip_memory_us, double, mode_lib->mp.MetaTripToMemory);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
316
dml_get_var_func(wm_fclk_change, double, mode_lib->mp.Watermark.FCLKChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
317
dml_get_var_func(wm_usr_retraining, double, mode_lib->mp.Watermark.USRRetrainingWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
318
dml_get_var_func(wm_temp_read_or_ppt, double, mode_lib->mp.Watermark.temp_read_or_ppt_watermark_us);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
319
dml_get_var_func(wm_dram_clock_change, double, mode_lib->mp.Watermark.DRAMClockChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
320
dml_get_var_func(fraction_of_urgent_bandwidth, double, mode_lib->mp.FractionOfUrgentBandwidth);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
321
dml_get_var_func(fraction_of_urgent_bandwidth_imm_flip, double, mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
322
dml_get_var_func(fraction_of_urgent_bandwidth_mall, double, mode_lib->mp.FractionOfUrgentBandwidthMALL);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
323
dml_get_var_func(wm_writeback_dram_clock_change, double, mode_lib->mp.Watermark.WritebackDRAMClockChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
324
dml_get_var_func(wm_writeback_fclk_change, double, mode_lib->mp.Watermark.WritebackFCLKChangeWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
325
dml_get_var_func(stutter_efficiency, double, mode_lib->mp.StutterEfficiency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
326
dml_get_var_func(stutter_efficiency_no_vblank, double, mode_lib->mp.StutterEfficiencyNotIncludingVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
327
dml_get_var_func(stutter_num_bursts, double, mode_lib->mp.NumberOfStutterBurstsPerFrame);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
328
dml_get_var_func(stutter_efficiency_z8, double, mode_lib->mp.Z8StutterEfficiency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
329
dml_get_var_func(stutter_efficiency_no_vblank_z8, double, mode_lib->mp.Z8StutterEfficiencyNotIncludingVBlank);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
330
dml_get_var_func(stutter_num_bursts_z8, double, mode_lib->mp.Z8NumberOfStutterBurstsPerFrame);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
331
dml_get_var_func(stutter_period, double, mode_lib->mp.StutterPeriod);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
332
dml_get_var_func(stutter_efficiency_z8_bestcase, double, mode_lib->mp.Z8StutterEfficiencyBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
333
dml_get_var_func(stutter_num_bursts_z8_bestcase, double, mode_lib->mp.Z8NumberOfStutterBurstsPerFrameBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
334
dml_get_var_func(stutter_period_bestcase, double, mode_lib->mp.StutterPeriodBestCase);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
335
dml_get_var_func(fclk_change_latency, double, mode_lib->mp.MaxActiveFCLKChangeLatencySupported);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
336
dml_get_var_func(global_dppclk_khz, double, mode_lib->mp.GlobalDPPCLK * 1000.0);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
344
dml_get_var_func(sys_active_avg_bw_available_sdp, double, mode_lib->mp.avg_bandwidth_available[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
345
dml_get_var_func(sys_active_avg_bw_available_dram, double, mode_lib->mp.avg_bandwidth_available[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
347
dml_get_var_func(svp_prefetch_avg_bw_available_sdp, double, mode_lib->mp.avg_bandwidth_available[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
348
dml_get_var_func(svp_prefetch_avg_bw_available_dram, double, mode_lib->mp.avg_bandwidth_available[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
350
dml_get_var_func(sys_active_urg_bw_available_sdp, double, mode_lib->mp.urg_bandwidth_available[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
351
dml_get_var_func(sys_active_urg_bw_available_dram, double, mode_lib->mp.urg_bandwidth_available[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
352
dml_get_var_func(sys_active_urg_bw_available_dram_vm_only, double, mode_lib->mp.urg_bandwidth_available_vm_only[dml2_core_internal_soc_state_sys_active]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
354
dml_get_var_func(svp_prefetch_urg_bw_available_sdp, double, mode_lib->mp.urg_bandwidth_available[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
355
dml_get_var_func(svp_prefetch_urg_bw_available_dram, double, mode_lib->mp.urg_bandwidth_available[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
356
dml_get_var_func(svp_prefetch_urg_bw_available_dram_vm_only, double, mode_lib->mp.urg_bandwidth_available_vm_only[dml2_core_internal_soc_state_svp_prefetch]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
358
dml_get_var_func(urgent_latency, double, mode_lib->mp.UrgentLatency);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
364
dml_get_var_func(sys_active_urg_bw_required_sdp, double, mode_lib->mp.urg_bandwidth_required[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
365
dml_get_var_func(sys_active_urg_bw_required_dram, double, mode_lib->mp.urg_bandwidth_required[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
366
dml_get_var_func(svp_prefetch_urg_bw_required_sdp, double, mode_lib->mp.urg_bandwidth_required[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
367
dml_get_var_func(svp_prefetch_urg_bw_required_dram, double, mode_lib->mp.urg_bandwidth_required[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
369
dml_get_var_func(sys_active_non_urg_required_sdp, double, mode_lib->mp.non_urg_bandwidth_required[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
370
dml_get_var_func(sys_active_non_urg_required_dram, double, mode_lib->mp.non_urg_bandwidth_required[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
371
dml_get_var_func(svp_prefetch_non_urg_bw_required_sdp, double, mode_lib->mp.non_urg_bandwidth_required[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
372
dml_get_var_func(svp_prefetch_non_urg_bw_required_dram, double, mode_lib->mp.non_urg_bandwidth_required[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
374
dml_get_var_func(sys_active_urg_bw_required_sdp_flip, double, mode_lib->mp.urg_bandwidth_required_flip[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
375
dml_get_var_func(sys_active_urg_bw_required_dram_flip, double, mode_lib->mp.urg_bandwidth_required_flip[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
376
dml_get_var_func(svp_prefetch_urg_bw_required_sdp_flip, double, mode_lib->mp.urg_bandwidth_required_flip[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
377
dml_get_var_func(svp_prefetch_urg_bw_required_dram_flip, double, mode_lib->mp.urg_bandwidth_required_flip[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
379
dml_get_var_func(sys_active_non_urg_required_sdp_flip, double, mode_lib->mp.non_urg_bandwidth_required_flip[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
380
dml_get_var_func(sys_active_non_urg_required_dram_flip, double, mode_lib->mp.non_urg_bandwidth_required_flip[dml2_core_internal_soc_state_sys_active][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
381
dml_get_var_func(svp_prefetch_non_urg_bw_required_sdp_flip, double, mode_lib->mp.non_urg_bandwidth_required_flip[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_sdp]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
382
dml_get_var_func(svp_prefetch_non_urg_bw_required_dram_flip, double, mode_lib->mp.non_urg_bandwidth_required_flip[dml2_core_internal_soc_state_svp_prefetch][dml2_core_internal_bw_dram]);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
384
dml_get_var_func(comp_buffer_size_kbytes, unsigned int, mode_lib->mp.CompressedBufferSizeInkByte);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
386
dml_get_var_func(unbounded_request_enabled, bool, mode_lib->mp.UnboundedRequestEnabled);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
387
dml_get_var_func(wm_writeback_urgent, double, mode_lib->mp.Watermark.WritebackUrgentWatermark);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
388
dml_get_var_func(cstate_max_cap_mode, bool, mode_lib->mp.DCHUBBUB_ARB_CSTATE_MAX_CAP_MODE);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
389
dml_get_var_func(compbuf_reserved_space_64b, unsigned int, mode_lib->mp.compbuf_reserved_space_64b);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
390
dml_get_var_func(hw_debug5, bool, mode_lib->mp.hw_debug5);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_dcn4_calcs.c
391
dml_get_var_func(dcfclk_deep_sleep_hysteresis, unsigned int, mode_lib->mp.dcfclk_deep_sleep_hysteresis);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_core/dml2_core_shared_types.h
2312
struct dml2_core_internal_mode_program mp; // struct for mode programming
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
770
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].fclk_pstate = (int unsigned)(mode_lib->mp.Watermark.FCLKChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
771
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].sr_enter = (int unsigned)(mode_lib->mp.Watermark.StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
772
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].sr_exit = (int unsigned)(mode_lib->mp.Watermark.StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
773
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].sr_enter_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
774
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].sr_exit_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
775
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].temp_read_or_ppt = (int unsigned)(mode_lib->mp.Watermark.temp_read_or_ppt_watermark_us * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
776
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].uclk_pstate = (int unsigned)(mode_lib->mp.Watermark.DRAMClockChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
777
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].urgent = (int unsigned)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
778
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].usr = (int unsigned)(mode_lib->mp.Watermark.USRRetrainingWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
779
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].refcyc_per_trip_to_mem = (unsigned int)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
780
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].refcyc_per_meta_trip_to_mem = (unsigned int)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
781
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].frac_urg_bw_flip = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
782
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].frac_urg_bw_nom = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidth * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
783
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_A].frac_urg_bw_mall = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthMALL * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
786
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].fclk_pstate = (int unsigned)(mode_lib->mp.Watermark.FCLKChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
787
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].sr_enter = (int unsigned)(mode_lib->mp.Watermark.StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
788
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].sr_exit = (int unsigned)(mode_lib->mp.Watermark.StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
789
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].sr_enter_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterEnterPlusExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
790
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].sr_exit_z8 = (int unsigned)(mode_lib->mp.Watermark.Z8StutterExitWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
791
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].temp_read_or_ppt = (int unsigned)(mode_lib->mp.Watermark.temp_read_or_ppt_watermark_us * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
792
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].uclk_pstate = (int unsigned)(mode_lib->mp.Watermark.DRAMClockChangeWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
793
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].urgent = (int unsigned)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
794
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].usr = (int unsigned)(mode_lib->mp.Watermark.USRRetrainingWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
795
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].refcyc_per_trip_to_mem = (unsigned int)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
796
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].refcyc_per_meta_trip_to_mem = (unsigned int)(mode_lib->mp.Watermark.UrgentWatermark * refclk_freq_in_mhz);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
797
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].frac_urg_bw_flip = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthImmediateFlip * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
798
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].frac_urg_bw_nom = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidth * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml21/src/dml2_dpmm/dml2_dpmm_dcn4.c
799
dchubbub_regs->wm_regs[DML2_DCHUB_WATERMARK_SET_B].frac_urg_bw_mall = (unsigned int)(mode_lib->mp.FractionOfUrgentBandwidthMALL * 1000);
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_utils.c
285
context->bw_ctx.bw.dcn.clk.dcfclk_deep_sleep_khz = (unsigned int)in_ctx->v20.dml_core_ctx.mp.DCFCLKDeepSleep * 1000;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
429
memset(&dml2->v20.dml_core_ctx.mp, 0, sizeof(struct mode_program_st));
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
464
out_clks.dispclk_khz = (unsigned int)dml2->v20.dml_core_ctx.mp.Dispclk_calculated * 1000;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
495
context->bw_ctx.dml.vba.StutterPeriod = context->bw_ctx.dml2->v20.dml_core_ctx.mp.StutterPeriod;
drivers/gpu/drm/amd/display/dc/dml2_0/dml2_wrapper_fpu.c
523
memset(&dml2->v20.dml_core_ctx.mp, 0, sizeof(struct mode_program_st));
drivers/gpu/drm/arm/malidp_crtc.c
271
struct malidp_plane *mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_crtc.c
274
if (!(mp->layer->id & scaling))
drivers/gpu/drm/arm/malidp_crtc.c
321
s->plane_src_id = mp->layer->id;
drivers/gpu/drm/arm/malidp_crtc.c
392
struct malidp_plane *mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_crtc.c
404
if ((mp->layer->id != DE_VIDEO1) ||
drivers/gpu/drm/arm/malidp_planes.c
262
static int malidp_se_check_scaling(struct malidp_plane *mp,
drivers/gpu/drm/arm/malidp_planes.c
291
mc->scaled_planes_mask &= ~(mp->layer->id);
drivers/gpu/drm/arm/malidp_planes.c
295
if (mp->layer->id & (DE_SMART | DE_GRAPHICS2))
drivers/gpu/drm/arm/malidp_planes.c
298
mc->scaled_planes_mask |= mp->layer->id;
drivers/gpu/drm/arm/malidp_planes.c
303
static u32 malidp_get_pgsize_bitmap(struct malidp_plane *mp)
drivers/gpu/drm/arm/malidp_planes.c
307
mmu_dom = iommu_get_domain_for_dev(mp->base.dev->dev);
drivers/gpu/drm/arm/malidp_planes.c
480
static void malidp_de_prefetch_settings(struct malidp_plane *mp,
drivers/gpu/drm/arm/malidp_planes.c
483
if (!mp->layer->mmu_ctrl_offset)
drivers/gpu/drm/arm/malidp_planes.c
487
ms->mmu_prefetch_pgsize = malidp_get_pgsize_bitmap(mp);
drivers/gpu/drm/arm/malidp_planes.c
497
struct malidp_plane *mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_planes.c
510
ms->format = malidp_hw_get_format_id(&mp->hwdev->hw->map,
drivers/gpu/drm/arm/malidp_planes.c
511
mp->layer->id, fb->format->format,
drivers/gpu/drm/arm/malidp_planes.c
518
u8 alignment = malidp_hw_get_pitch_align(mp->hwdev, rotated);
drivers/gpu/drm/arm/malidp_planes.c
539
if ((new_plane_state->crtc_w > mp->hwdev->max_line_size) ||
drivers/gpu/drm/arm/malidp_planes.c
540
(new_plane_state->crtc_h > mp->hwdev->max_line_size) ||
drivers/gpu/drm/arm/malidp_planes.c
541
(new_plane_state->crtc_w < mp->hwdev->min_line_size) ||
drivers/gpu/drm/arm/malidp_planes.c
542
(new_plane_state->crtc_h < mp->hwdev->min_line_size))
drivers/gpu/drm/arm/malidp_planes.c
551
!(mp->hwdev->hw->features & MALIDP_DEVICE_LV_HAS_3_STRIDES) &&
drivers/gpu/drm/arm/malidp_planes.c
555
ret = malidp_se_check_scaling(mp, new_plane_state);
drivers/gpu/drm/arm/malidp_planes.c
561
if (mp->layer->rot == ROTATE_NONE)
drivers/gpu/drm/arm/malidp_planes.c
563
if ((mp->layer->rot == ROTATE_COMPRESSED) && !(fb->modifier))
drivers/gpu/drm/arm/malidp_planes.c
575
if (mp->layer->id == DE_SMART && fb->modifier) {
drivers/gpu/drm/arm/malidp_planes.c
584
val = mp->hwdev->hw->rotmem_required(mp->hwdev, new_plane_state->crtc_w,
drivers/gpu/drm/arm/malidp_planes.c
600
malidp_de_prefetch_settings(mp, ms);
drivers/gpu/drm/arm/malidp_planes.c
605
static void malidp_de_set_plane_pitches(struct malidp_plane *mp,
drivers/gpu/drm/arm/malidp_planes.c
611
if (!mp->layer->stride_offset)
drivers/gpu/drm/arm/malidp_planes.c
615
num_strides = (mp->hwdev->hw->features &
drivers/gpu/drm/arm/malidp_planes.c
624
unsigned int block_h = drm_format_info_block_height(mp->base.state->fb->format, i);
drivers/gpu/drm/arm/malidp_planes.c
626
malidp_hw_write(mp->hwdev, pitches[i] * block_h,
drivers/gpu/drm/arm/malidp_planes.c
627
mp->layer->base +
drivers/gpu/drm/arm/malidp_planes.c
628
mp->layer->stride_offset + i * 4);
drivers/gpu/drm/arm/malidp_planes.c
686
static void malidp_de_set_mmu_control(struct malidp_plane *mp,
drivers/gpu/drm/arm/malidp_planes.c
692
if (!mp->layer->mmu_ctrl_offset)
drivers/gpu/drm/arm/malidp_planes.c
700
malidp_hw_write(mp->hwdev, mmu_ctrl,
drivers/gpu/drm/arm/malidp_planes.c
701
mp->layer->base + mp->layer->mmu_ctrl_offset);
drivers/gpu/drm/arm/malidp_planes.c
705
struct malidp_plane *mp,
drivers/gpu/drm/arm/malidp_planes.c
710
struct drm_plane *plane = &mp->base;
drivers/gpu/drm/arm/malidp_planes.c
713
ptr = mp->layer->ptr + (plane_index << 4);
drivers/gpu/drm/arm/malidp_planes.c
735
malidp_hw_write(mp->hwdev, lower_32_bits(dma_addr), ptr);
drivers/gpu/drm/arm/malidp_planes.c
736
malidp_hw_write(mp->hwdev, upper_32_bits(dma_addr), ptr + 4);
drivers/gpu/drm/arm/malidp_planes.c
741
struct malidp_plane *mp;
drivers/gpu/drm/arm/malidp_planes.c
745
mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_planes.c
748
if (!mp->layer->afbc_decoder_offset)
drivers/gpu/drm/arm/malidp_planes.c
752
malidp_hw_write(mp->hwdev, 0, mp->layer->afbc_decoder_offset);
drivers/gpu/drm/arm/malidp_planes.c
764
malidp_hw_write(mp->hwdev, val,
drivers/gpu/drm/arm/malidp_planes.c
765
mp->layer->afbc_decoder_offset + MALIDP_AD_CROP_H);
drivers/gpu/drm/arm/malidp_planes.c
769
malidp_hw_write(mp->hwdev, val,
drivers/gpu/drm/arm/malidp_planes.c
770
mp->layer->afbc_decoder_offset + MALIDP_AD_CROP_V);
drivers/gpu/drm/arm/malidp_planes.c
778
malidp_hw_write(mp->hwdev, val, mp->layer->afbc_decoder_offset);
drivers/gpu/drm/arm/malidp_planes.c
786
struct malidp_plane *mp;
drivers/gpu/drm/arm/malidp_planes.c
796
mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_planes.c
814
val = malidp_hw_read(mp->hwdev, mp->layer->base);
drivers/gpu/drm/arm/malidp_planes.c
816
malidp_hw_write(mp->hwdev, val, mp->layer->base);
drivers/gpu/drm/arm/malidp_planes.c
819
malidp_set_plane_base_addr(fb, mp, i);
drivers/gpu/drm/arm/malidp_planes.c
821
malidp_de_set_mmu_control(mp, ms);
drivers/gpu/drm/arm/malidp_planes.c
823
malidp_de_set_plane_pitches(mp, ms->n_planes,
drivers/gpu/drm/arm/malidp_planes.c
828
malidp_de_set_color_encoding(mp, plane->state->color_encoding,
drivers/gpu/drm/arm/malidp_planes.c
831
malidp_hw_write(mp->hwdev, LAYER_H_VAL(src_w) | LAYER_V_VAL(src_h),
drivers/gpu/drm/arm/malidp_planes.c
832
mp->layer->base + MALIDP_LAYER_SIZE);
drivers/gpu/drm/arm/malidp_planes.c
834
malidp_hw_write(mp->hwdev, LAYER_H_VAL(dest_w) | LAYER_V_VAL(dest_h),
drivers/gpu/drm/arm/malidp_planes.c
835
mp->layer->base + MALIDP_LAYER_COMP_SIZE);
drivers/gpu/drm/arm/malidp_planes.c
837
malidp_hw_write(mp->hwdev, LAYER_H_VAL(new_state->crtc_x) |
drivers/gpu/drm/arm/malidp_planes.c
839
mp->layer->base + MALIDP_LAYER_OFFSET);
drivers/gpu/drm/arm/malidp_planes.c
841
if (mp->layer->id == DE_SMART) {
drivers/gpu/drm/arm/malidp_planes.c
846
malidp_hw_write(mp->hwdev, 1,
drivers/gpu/drm/arm/malidp_planes.c
847
mp->layer->base + MALIDP550_LS_ENABLE);
drivers/gpu/drm/arm/malidp_planes.c
848
malidp_hw_write(mp->hwdev,
drivers/gpu/drm/arm/malidp_planes.c
850
mp->layer->base + MALIDP550_LS_R1_IN_SIZE);
drivers/gpu/drm/arm/malidp_planes.c
856
val = malidp_hw_read(mp->hwdev, mp->layer->base + MALIDP_LAYER_CONTROL);
drivers/gpu/drm/arm/malidp_planes.c
891
m->scaler_config.plane_src_id == mp->layer->id)
drivers/gpu/drm/arm/malidp_planes.c
898
malidp_hw_write(mp->hwdev, val,
drivers/gpu/drm/arm/malidp_planes.c
899
mp->layer->base + MALIDP_LAYER_CONTROL);
drivers/gpu/drm/arm/malidp_planes.c
905
struct malidp_plane *mp = to_malidp_plane(plane);
drivers/gpu/drm/arm/malidp_planes.c
907
malidp_hw_clearbits(mp->hwdev,
drivers/gpu/drm/arm/malidp_planes.c
909
mp->layer->base + MALIDP_LAYER_CONTROL);
drivers/hid/hid-wiimote-core.c
1045
wdata->state.mp > 0 && (flags & WIIPROTO_FLAG_MP_USED)) {
drivers/hid/hid-wiimote-core.c
1104
bool mp;
drivers/hid/hid-wiimote-core.c
1124
mp = false;
drivers/hid/hid-wiimote-core.c
1127
mp = wiimote_cmd_read_mp(wdata, mpdata);
drivers/hid/hid-wiimote-core.c
1154
if (mp) {
drivers/hid/hid-wiimote-core.c
1155
if (!wdata->state.mp) {
drivers/hid/hid-wiimote-core.c
1159
} else if (wdata->state.mp) {
drivers/hid/hid-wiimote-core.c
1165
mp = false;
drivers/hid/hid-wiimote-core.c
1168
if (mp) {
drivers/hid/hid-wiimote-core.c
1185
if (mp) {
drivers/hid/hid-wiimote-core.c
1205
wdata->state.mp, wdata->state.exttype);
drivers/hid/hid-wiimote-core.c
774
wdata->state.mp = mode;
drivers/hid/hid-wiimote-core.c
783
if (wdata->state.mp < 2)
drivers/hid/hid-wiimote-core.c
789
wdata->state.mp = 0;
drivers/hid/hid-wiimote-core.c
915
bool mp;
drivers/hid/hid-wiimote-core.c
920
mp = wiimote_cmd_read_mp(wdata, mpdata);
drivers/hid/hid-wiimote-core.c
924
if (mp) {
drivers/hid/hid-wiimote-core.c
925
if (!wdata->state.mp) {
drivers/hid/hid-wiimote-core.c
929
} else if (wdata->state.mp) {
drivers/hid/hid-wiimote-core.c
966
wdata->state.mp > 0 && (flags & WIIPROTO_FLAG_MP_USED)) {
drivers/hid/hid-wiimote-modules.c
2765
input_report_abs(wdata->mp, ABS_RX, x);
drivers/hid/hid-wiimote-modules.c
2766
input_report_abs(wdata->mp, ABS_RY, y);
drivers/hid/hid-wiimote-modules.c
2767
input_report_abs(wdata->mp, ABS_RZ, z);
drivers/hid/hid-wiimote-modules.c
2768
input_sync(wdata->mp);
drivers/hid/hid-wiimote-modules.c
2802
wdata->mp = input_allocate_device();
drivers/hid/hid-wiimote-modules.c
2803
if (!wdata->mp)
drivers/hid/hid-wiimote-modules.c
2806
input_set_drvdata(wdata->mp, wdata);
drivers/hid/hid-wiimote-modules.c
2807
wdata->mp->open = wiimod_mp_open;
drivers/hid/hid-wiimote-modules.c
2808
wdata->mp->close = wiimod_mp_close;
drivers/hid/hid-wiimote-modules.c
2809
wdata->mp->dev.parent = &wdata->hdev->dev;
drivers/hid/hid-wiimote-modules.c
2810
wdata->mp->id.bustype = wdata->hdev->bus;
drivers/hid/hid-wiimote-modules.c
2811
wdata->mp->id.vendor = wdata->hdev->vendor;
drivers/hid/hid-wiimote-modules.c
2812
wdata->mp->id.product = wdata->hdev->product;
drivers/hid/hid-wiimote-modules.c
2813
wdata->mp->id.version = wdata->hdev->version;
drivers/hid/hid-wiimote-modules.c
2814
wdata->mp->name = WIIMOTE_NAME " Motion Plus";
drivers/hid/hid-wiimote-modules.c
2816
set_bit(EV_ABS, wdata->mp->evbit);
drivers/hid/hid-wiimote-modules.c
2817
set_bit(ABS_RX, wdata->mp->absbit);
drivers/hid/hid-wiimote-modules.c
2818
set_bit(ABS_RY, wdata->mp->absbit);
drivers/hid/hid-wiimote-modules.c
2819
set_bit(ABS_RZ, wdata->mp->absbit);
drivers/hid/hid-wiimote-modules.c
2820
input_set_abs_params(wdata->mp,
drivers/hid/hid-wiimote-modules.c
2822
input_set_abs_params(wdata->mp,
drivers/hid/hid-wiimote-modules.c
2824
input_set_abs_params(wdata->mp,
drivers/hid/hid-wiimote-modules.c
2827
ret = input_register_device(wdata->mp);
drivers/hid/hid-wiimote-modules.c
2834
input_free_device(wdata->mp);
drivers/hid/hid-wiimote-modules.c
2835
wdata->mp = NULL;
drivers/hid/hid-wiimote-modules.c
2842
if (!wdata->mp)
drivers/hid/hid-wiimote-modules.c
2845
input_unregister_device(wdata->mp);
drivers/hid/hid-wiimote-modules.c
2846
wdata->mp = NULL;
drivers/hid/hid-wiimote.h
123
__u8 mp;
drivers/hid/hid-wiimote.h
153
struct input_dev *mp;
drivers/hid/hid-wiimote.h
202
void (*in_mp) (struct wiimote_data *wdata, const __u8 *mp);
drivers/infiniband/core/mad.c
589
static size_t mad_priv_size(const struct ib_mad_private *mp)
drivers/infiniband/core/mad.c
591
return sizeof(struct ib_mad_private) + mp->mad_size;
drivers/infiniband/core/mad.c
610
static size_t mad_priv_dma_size(const struct ib_mad_private *mp)
drivers/infiniband/core/mad.c
612
return sizeof(struct ib_grh) + mp->mad_size;
drivers/infiniband/hw/hfi1/chip.c
1059
static int qos_rmt_entries(unsigned int n_krcv_queues, unsigned int *mp,
drivers/infiniband/hw/hfi1/chip.c
14293
static int qos_rmt_entries(unsigned int n_krcv_queues, unsigned int *mp,
drivers/infiniband/hw/hfi1/chip.c
14321
if (mp)
drivers/infiniband/hw/hfi1/chip.c
14322
*mp = m;
drivers/infiniband/hw/hfi1/chip.c
14329
if (mp)
drivers/infiniband/hw/hfi1/chip.c
14330
*mp = 0;
drivers/infiniband/hw/mlx5/main.c
340
spin_lock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
341
mpi = ibdev->port[ib_port_num - 1].mp.mpi;
drivers/infiniband/hw/mlx5/main.c
350
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
367
spin_lock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
368
mpi = ibdev->port[port_num - 1].mp.mpi;
drivers/infiniband/hw/mlx5/main.c
376
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3886
spin_lock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3888
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3894
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3899
spin_lock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3905
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3910
spin_lock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3914
port->mp.mpi = NULL;
drivers/infiniband/hw/mlx5/main.c
3916
spin_unlock(&port->mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3940
spin_lock(&ibdev->port[port_num].mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3941
if (ibdev->port[port_num].mp.mpi) {
drivers/infiniband/hw/mlx5/main.c
3944
spin_unlock(&ibdev->port[port_num].mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
3948
ibdev->port[port_num].mp.mpi = mpi;
drivers/infiniband/hw/mlx5/main.c
3951
spin_unlock(&ibdev->port[port_num].mp.mpi_lock);
drivers/infiniband/hw/mlx5/main.c
4055
dev->port[i].mp.mpi = mpi;
drivers/infiniband/hw/mlx5/main.c
4099
if (dev->port[i].mp.mpi) {
drivers/infiniband/hw/mlx5/main.c
4102
kfree(dev->port[i].mp.mpi);
drivers/infiniband/hw/mlx5/main.c
4103
dev->port[i].mp.mpi = NULL;
drivers/infiniband/hw/mlx5/main.c
4107
list_add_tail(&dev->port[i].mp.mpi->list,
drivers/infiniband/hw/mlx5/main.c
4110
dev->port[i].mp.mpi);
drivers/infiniband/hw/mlx5/main.c
4441
spin_lock_init(&dev->port[i].mp.mpi_lock);
drivers/infiniband/hw/mlx5/mlx5_ib.h
932
struct mlx5_ib_multiport mp;
drivers/isdn/capi/capi.c
1038
struct capiminor *mp = capiminor_get(tty->index);
drivers/isdn/capi/capi.c
1042
tty->driver_data = mp;
drivers/isdn/capi/capi.c
1044
capiminor_put(mp);
drivers/isdn/capi/capi.c
1050
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1052
capiminor_put(mp);
drivers/isdn/capi/capi.c
1057
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1060
err = tty_port_open(&mp->port, tty, filp);
drivers/isdn/capi/capi.c
1064
handle_minor_recv(mp);
drivers/isdn/capi/capi.c
1070
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1072
tty_port_close(&mp->port, tty, filp);
drivers/isdn/capi/capi.c
1078
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1083
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1084
skb = mp->outskb;
drivers/isdn/capi/capi.c
1086
mp->outskb = NULL;
drivers/isdn/capi/capi.c
1087
__skb_queue_tail(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
1088
mp->outbytes += skb->len;
drivers/isdn/capi/capi.c
1094
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1101
__skb_queue_tail(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
1102
mp->outbytes += skb->len;
drivers/isdn/capi/capi.c
1103
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1105
handle_minor_send(mp);
drivers/isdn/capi/capi.c
1112
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1119
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1120
skb = mp->outskb;
drivers/isdn/capi/capi.c
1126
mp->outskb = NULL;
drivers/isdn/capi/capi.c
1127
__skb_queue_tail(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
1128
mp->outbytes += skb->len;
drivers/isdn/capi/capi.c
1136
mp->outskb = skb;
drivers/isdn/capi/capi.c
1143
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1146
handle_minor_send(mp);
drivers/isdn/capi/capi.c
1153
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1156
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1157
skb = mp->outskb;
drivers/isdn/capi/capi.c
1159
mp->outskb = NULL;
drivers/isdn/capi/capi.c
1160
__skb_queue_tail(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
1161
mp->outbytes += skb->len;
drivers/isdn/capi/capi.c
1162
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1164
handle_minor_send(mp);
drivers/isdn/capi/capi.c
1166
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
1168
handle_minor_recv(mp);
drivers/isdn/capi/capi.c
1173
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1176
room = CAPINC_MAX_SENDQUEUE-skb_queue_len(&mp->outqueue);
drivers/isdn/capi/capi.c
1184
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1187
mp->outbytes, mp->nack,
drivers/isdn/capi/capi.c
1188
skb_queue_len(&mp->outqueue),
drivers/isdn/capi/capi.c
1189
skb_queue_len(&mp->inqueue));
drivers/isdn/capi/capi.c
1190
return mp->outbytes;
drivers/isdn/capi/capi.c
1195
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1196
mp->ttyinstop = 1;
drivers/isdn/capi/capi.c
1201
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1203
mp->ttyinstop = 0;
drivers/isdn/capi/capi.c
1204
handle_minor_recv(mp);
drivers/isdn/capi/capi.c
1209
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1211
mp->ttyoutstop = 1;
drivers/isdn/capi/capi.c
1216
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1218
mp->ttyoutstop = 0;
drivers/isdn/capi/capi.c
1219
handle_minor_send(mp);
drivers/isdn/capi/capi.c
1224
struct capiminor *mp = tty->driver_data;
drivers/isdn/capi/capi.c
1226
tty_port_hangup(&mp->port);
drivers/isdn/capi/capi.c
147
static int capiminor_add_ack(struct capiminor *mp, u16 datahandle)
drivers/isdn/capi/capi.c
158
spin_lock_bh(&mp->ackqlock);
drivers/isdn/capi/capi.c
159
list_add_tail(&n->list, &mp->ackqueue);
drivers/isdn/capi/capi.c
160
mp->nack++;
drivers/isdn/capi/capi.c
161
spin_unlock_bh(&mp->ackqlock);
drivers/isdn/capi/capi.c
165
static int capiminor_del_ack(struct capiminor *mp, u16 datahandle)
drivers/isdn/capi/capi.c
169
spin_lock_bh(&mp->ackqlock);
drivers/isdn/capi/capi.c
170
list_for_each_entry_safe(p, tmp, &mp->ackqueue, list) {
drivers/isdn/capi/capi.c
173
mp->nack--;
drivers/isdn/capi/capi.c
174
spin_unlock_bh(&mp->ackqlock);
drivers/isdn/capi/capi.c
179
spin_unlock_bh(&mp->ackqlock);
drivers/isdn/capi/capi.c
183
static void capiminor_del_all_ack(struct capiminor *mp)
drivers/isdn/capi/capi.c
187
list_for_each_entry_safe(p, tmp, &mp->ackqueue, list) {
drivers/isdn/capi/capi.c
190
mp->nack--;
drivers/isdn/capi/capi.c
199
struct capiminor *mp = container_of(port, struct capiminor, port);
drivers/isdn/capi/capi.c
201
kfree_skb(mp->outskb);
drivers/isdn/capi/capi.c
202
skb_queue_purge(&mp->inqueue);
drivers/isdn/capi/capi.c
203
skb_queue_purge(&mp->outqueue);
drivers/isdn/capi/capi.c
204
capiminor_del_all_ack(mp);
drivers/isdn/capi/capi.c
205
kfree(mp);
drivers/isdn/capi/capi.c
214
struct capiminor *mp;
drivers/isdn/capi/capi.c
218
mp = kzalloc_obj(*mp);
drivers/isdn/capi/capi.c
219
if (!mp) {
drivers/isdn/capi/capi.c
224
mp->ap = ap;
drivers/isdn/capi/capi.c
225
mp->ncci = ncci;
drivers/isdn/capi/capi.c
226
INIT_LIST_HEAD(&mp->ackqueue);
drivers/isdn/capi/capi.c
227
spin_lock_init(&mp->ackqlock);
drivers/isdn/capi/capi.c
229
skb_queue_head_init(&mp->inqueue);
drivers/isdn/capi/capi.c
230
skb_queue_head_init(&mp->outqueue);
drivers/isdn/capi/capi.c
231
spin_lock_init(&mp->outlock);
drivers/isdn/capi/capi.c
233
tty_port_init(&mp->port);
drivers/isdn/capi/capi.c
234
mp->port.ops = &capiminor_port_ops;
drivers/isdn/capi/capi.c
240
capiminors[minor] = mp;
drivers/isdn/capi/capi.c
250
mp->minor = minor;
drivers/isdn/capi/capi.c
252
dev = tty_port_register_device(&mp->port, capinc_tty_driver, minor,
drivers/isdn/capi/capi.c
257
return mp;
drivers/isdn/capi/capi.c
265
tty_port_put(&mp->port);
drivers/isdn/capi/capi.c
271
struct capiminor *mp;
drivers/isdn/capi/capi.c
274
mp = capiminors[minor];
drivers/isdn/capi/capi.c
275
if (mp)
drivers/isdn/capi/capi.c
276
tty_port_get(&mp->port);
drivers/isdn/capi/capi.c
279
return mp;
drivers/isdn/capi/capi.c
282
static inline void capiminor_put(struct capiminor *mp)
drivers/isdn/capi/capi.c
284
tty_port_put(&mp->port);
drivers/isdn/capi/capi.c
287
static void capiminor_free(struct capiminor *mp)
drivers/isdn/capi/capi.c
289
tty_unregister_device(capinc_tty_driver, mp->minor);
drivers/isdn/capi/capi.c
292
capiminors[mp->minor] = NULL;
drivers/isdn/capi/capi.c
295
capiminor_put(mp);
drivers/isdn/capi/capi.c
308
struct capiminor *mp = np->minorp;
drivers/isdn/capi/capi.c
310
if (mp) {
drivers/isdn/capi/capi.c
311
tty_port_tty_vhangup(&mp->port);
drivers/isdn/capi/capi.c
312
capiminor_free(mp);
drivers/isdn/capi/capi.c
318
struct capiminor *mp = np->minorp;
drivers/isdn/capi/capi.c
322
if (mp) {
drivers/isdn/capi/capi.c
323
tty = tty_port_tty_get(&mp->port);
drivers/isdn/capi/capi.c
383
gen_data_b3_resp_for(struct capiminor *mp, struct sk_buff *skb)
drivers/isdn/capi/capi.c
391
capimsg_setu16(s, 2, mp->ap->applid);
drivers/isdn/capi/capi.c
394
capimsg_setu16(s, 6, atomic_inc_return(&mp->msgid));
drivers/isdn/capi/capi.c
395
capimsg_setu32(s, 8, mp->ncci);
drivers/isdn/capi/capi.c
401
static int handle_recv_skb(struct capiminor *mp, struct sk_buff *skb)
drivers/isdn/capi/capi.c
410
tty = tty_port_tty_get(&mp->port);
drivers/isdn/capi/capi.c
429
if (mp->ttyinstop) {
drivers/isdn/capi/capi.c
439
nskb = gen_data_b3_resp_for(mp, skb);
drivers/isdn/capi/capi.c
447
errcode = capi20_put_message(mp->ap, nskb);
drivers/isdn/capi/capi.c
475
static void handle_minor_recv(struct capiminor *mp)
drivers/isdn/capi/capi.c
479
while ((skb = skb_dequeue(&mp->inqueue)) != NULL)
drivers/isdn/capi/capi.c
480
if (handle_recv_skb(mp, skb) < 0) {
drivers/isdn/capi/capi.c
481
skb_queue_head(&mp->inqueue, skb);
drivers/isdn/capi/capi.c
486
static void handle_minor_send(struct capiminor *mp)
drivers/isdn/capi/capi.c
494
tty = tty_port_tty_get(&mp->port);
drivers/isdn/capi/capi.c
498
if (mp->ttyoutstop) {
drivers/isdn/capi/capi.c
505
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
506
skb = __skb_dequeue(&mp->outqueue);
drivers/isdn/capi/capi.c
508
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
512
mp->outbytes -= len;
drivers/isdn/capi/capi.c
513
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
515
datahandle = atomic_inc_return(&mp->datahandle);
drivers/isdn/capi/capi.c
519
capimsg_setu16(skb->data, 2, mp->ap->applid);
drivers/isdn/capi/capi.c
522
capimsg_setu16(skb->data, 6, atomic_inc_return(&mp->msgid));
drivers/isdn/capi/capi.c
523
capimsg_setu32(skb->data, 8, mp->ncci); /* NCCI */
drivers/isdn/capi/capi.c
529
if (capiminor_add_ack(mp, datahandle) < 0) {
drivers/isdn/capi/capi.c
532
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
533
__skb_queue_head(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
534
mp->outbytes += len;
drivers/isdn/capi/capi.c
535
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
539
errcode = capi20_put_message(mp->ap, skb);
drivers/isdn/capi/capi.c
545
capiminor_del_ack(mp, datahandle);
drivers/isdn/capi/capi.c
550
spin_lock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
551
__skb_queue_head(&mp->outqueue, skb);
drivers/isdn/capi/capi.c
552
mp->outbytes += len;
drivers/isdn/capi/capi.c
553
spin_unlock_bh(&mp->outlock);
drivers/isdn/capi/capi.c
572
struct capiminor *mp;
drivers/isdn/capi/capi.c
607
mp = np->minorp;
drivers/isdn/capi/capi.c
608
if (!mp) {
drivers/isdn/capi/capi.c
617
skb_queue_tail(&mp->inqueue, skb);
drivers/isdn/capi/capi.c
619
handle_minor_recv(mp);
drivers/isdn/capi/capi.c
628
capiminor_del_ack(mp, datahandle);
drivers/isdn/capi/capi.c
629
tty_port_tty_wakeup(&mp->port);
drivers/isdn/capi/capi.c
630
handle_minor_send(mp);
drivers/isdn/capi/capi.c
916
struct capiminor *mp;
drivers/isdn/capi/capi.c
926
mp = nccip->minorp;
drivers/isdn/capi/capi.c
927
if (mp)
drivers/isdn/capi/capi.c
928
unit = mp->minor;
drivers/isdn/hardware/mISDN/mISDNisar.c
179
u8 *mp;
drivers/isdn/hardware/mISDN/mISDNisar.c
242
mp = isar->buf;
drivers/isdn/hardware/mISDN/mISDNisar.c
244
*mp++ = blk_head.sadr >> 8;
drivers/isdn/hardware/mISDN/mISDNisar.c
245
*mp++ = blk_head.sadr & 0xFF;
drivers/isdn/hardware/mISDN/mISDNisar.c
248
*mp++ = noc;
drivers/isdn/hardware/mISDN/mISDNisar.c
254
*mp++ = val >> 8;
drivers/isdn/hardware/mISDN/mISDNisar.c
255
*mp++ = val & 0xFF;
drivers/media/pci/bt8xx/bt878.c
320
bt878_device_control(struct bt878 *bt, unsigned int cmd, union dst_gpio_packet *mp)
drivers/media/pci/bt8xx/bt878.c
332
mp->enb.mask,
drivers/media/pci/bt8xx/bt878.c
333
mp->enb.enable);
drivers/media/pci/bt8xx/bt878.c
338
mp->outp.mask,
drivers/media/pci/bt8xx/bt878.c
339
mp->outp.highvals);
drivers/media/pci/bt8xx/bt878.c
344
retval = bttv_read_gpio(bt->bttv_nr, &mp->rd.value);
drivers/media/pci/bt8xx/bt878.c
349
bt->TS_Size = mp->psize;
drivers/media/pci/bt8xx/dst_priv.h
36
int bt878_device_control(struct bt878 *bt, unsigned int cmd, union dst_gpio_packet *mp);
drivers/media/platform/renesas/vsp1/vsp1_uds.c
59
unsigned int mp;
drivers/media/platform/renesas/vsp1/vsp1_uds.c
61
mp = ratio / 4096;
drivers/media/platform/renesas/vsp1/vsp1_uds.c
62
mp = mp < 4 ? 1 : (mp < 8 ? 2 : 4);
drivers/media/platform/renesas/vsp1/vsp1_uds.c
64
return (input - 1) / mp * mp * 4096 / ratio + 1;
drivers/media/platform/renesas/vsp1/vsp1_uds.c
92
unsigned int mp;
drivers/media/platform/renesas/vsp1/vsp1_uds.c
94
mp = ratio / 4096;
drivers/media/platform/renesas/vsp1/vsp1_uds.c
95
mp = mp < 4 ? 1 : (mp < 8 ? 2 : 4);
drivers/media/platform/renesas/vsp1/vsp1_uds.c
97
return 64 * 4096 * mp / ratio;
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1517
struct rkisp1_capture *mp = &rkisp1->capture_devs[RKISP1_MAINPATH];
drivers/media/platform/rockchip/rkisp1/rkisp1-capture.c
1520
rkisp1_unregister_capture(mp);
drivers/media/platform/rockchip/rkisp1/rkisp1-resizer.c
761
struct rkisp1_resizer *mp = &rkisp1->resizer_devs[RKISP1_MAINPATH];
drivers/media/platform/rockchip/rkisp1/rkisp1-resizer.c
764
rkisp1_rsz_unregister(mp);
drivers/media/platform/samsung/exynos4-is/fimc-capture.c
1188
ret = video_device_pipeline_start(&vc->ve.vdev, &vc->ve.pipe->mp);
drivers/media/platform/samsung/exynos4-is/fimc-isp-video.c
497
ret = video_device_pipeline_start(&ve->vdev, &ve->pipe->mp);
drivers/media/platform/samsung/exynos4-is/fimc-lite.c
819
ret = video_device_pipeline_start(&fimc->ve.vdev, &fimc->ve.pipe->mp);
drivers/media/platform/ti/vpe/csc.c
146
const struct v4l2_pix_format_mplane *mp;
drivers/media/platform/ti/vpe/csc.c
153
mp = &src_fmt->fmt.pix_mp;
drivers/media/platform/ti/vpe/csc.c
154
src_pixelformat = mp->pixelformat;
drivers/media/platform/ti/vpe/csc.c
155
src_ycbcr_enc = mp->ycbcr_enc;
drivers/media/platform/ti/vpe/csc.c
156
src_quantization = mp->quantization;
drivers/media/platform/ti/vpe/csc.c
165
mp = &dst_fmt->fmt.pix_mp;
drivers/media/platform/ti/vpe/csc.c
166
dst_pixelformat = mp->pixelformat;
drivers/media/platform/ti/vpe/csc.c
167
dst_ycbcr_enc = mp->ycbcr_enc;
drivers/media/platform/ti/vpe/csc.c
168
dst_quantization = mp->quantization;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
527
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
530
mp->width = dev->fmt_cap_rect.width;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
531
mp->height = dev->fmt_cap_rect.height;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
532
mp->field = dev->field_cap;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
533
mp->pixelformat = dev->fmt_cap->fourcc;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
534
mp->colorspace = vivid_colorspace_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
535
mp->xfer_func = vivid_xfer_func_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
537
mp->hsv_enc = vivid_hsv_enc_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
539
mp->ycbcr_enc = vivid_ycbcr_enc_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
540
mp->quantization = vivid_quantization_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
541
mp->num_planes = dev->fmt_cap->buffers;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
542
for (p = 0; p < mp->num_planes; p++) {
drivers/media/test-drivers/vivid/vivid-vid-cap.c
543
mp->plane_fmt[p].bytesperline = tpg_g_bytesperline(&dev->tpg, p);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
544
mp->plane_fmt[p].sizeimage =
drivers/media/test-drivers/vivid/vivid-vid-cap.c
545
(tpg_g_line_width(&dev->tpg, p) * mp->height) /
drivers/media/test-drivers/vivid/vivid-vid-cap.c
555
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
556
struct v4l2_plane_pix_format *pfmt = mp->plane_fmt;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
563
bool user_set_csc = !!(mp->flags & V4L2_PIX_FMT_FLAG_SET_CSC);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
565
fmt = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
568
mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
569
mp->pixelformat = V4L2_PIX_FMT_YUYV;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
570
fmt = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
573
mp->field = vivid_field_cap(dev, mp->field);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
578
height, mp->width, mp->height);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
589
if (V4L2_FIELD_HAS_T_OR_B(mp->field))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
593
mp->width = w;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
594
mp->height = h / factor;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
596
struct v4l2_rect r = { 0, 0, mp->width, mp->height * factor };
drivers/media/test-drivers/vivid/vivid-vid-cap.c
609
mp->width = r.width;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
610
mp->height = r.height / factor;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
615
mp->num_planes = fmt->buffers;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
618
bytesperline = (mp->width * fmt->bit_depth[p]) >> 3;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
627
pfmt[p].sizeimage = (pfmt[p].bytesperline * mp->height) /
drivers/media/test-drivers/vivid/vivid-vid-cap.c
633
pfmt[0].sizeimage += (pfmt[0].bytesperline * mp->height *
drivers/media/test-drivers/vivid/vivid-vid-cap.c
637
if (!user_set_csc || !v4l2_is_colorspace_valid(mp->colorspace))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
638
mp->colorspace = vivid_colorspace_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
640
if (!user_set_csc || !v4l2_is_xfer_func_valid(mp->xfer_func))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
641
mp->xfer_func = vivid_xfer_func_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
644
if (!user_set_csc || !v4l2_is_hsv_enc_valid(mp->hsv_enc))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
645
mp->hsv_enc = vivid_hsv_enc_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
647
if (!user_set_csc || !v4l2_is_ycbcr_enc_valid(mp->ycbcr_enc))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
648
mp->ycbcr_enc = vivid_ycbcr_enc_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
650
mp->ycbcr_enc = vivid_ycbcr_enc_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
655
if (!user_set_csc || !v4l2_is_quant_valid(mp->quantization))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
656
mp->quantization = vivid_quantization_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
658
mp->quantization = vivid_quantization_cap(dev);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
661
memset(mp->reserved, 0, sizeof(mp->reserved));
drivers/media/test-drivers/vivid/vivid-vid-cap.c
668
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
686
dev->fmt_cap = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
687
if (V4L2_FIELD_HAS_T_OR_B(mp->field))
drivers/media/test-drivers/vivid/vivid-vid-cap.c
694
struct v4l2_rect r = { 0, 0, mp->width, mp->height };
drivers/media/test-drivers/vivid/vivid-vid-cap.c
757
if (webcam_sizes[i].width == mp->width &&
drivers/media/test-drivers/vivid/vivid-vid-cap.c
758
webcam_sizes[i].height == mp->height)
drivers/media/test-drivers/vivid/vivid-vid-cap.c
765
struct v4l2_rect r = { 0, 0, mp->width, mp->height };
drivers/media/test-drivers/vivid/vivid-vid-cap.c
772
dev->fmt_cap_rect.width = mp->width;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
773
dev->fmt_cap_rect.height = mp->height;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
774
tpg_s_buf_height(&dev->tpg, mp->height);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
777
tpg_s_bytesperline(&dev->tpg, p, mp->plane_fmt[p].bytesperline);
drivers/media/test-drivers/vivid/vivid-vid-cap.c
778
dev->field_cap = mp->field;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
785
dev->tv_field_cap = mp->field;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
787
dev->tpg.colorspace = mp->colorspace;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
788
dev->tpg.xfer_func = mp->xfer_func;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
790
dev->tpg.ycbcr_enc = mp->ycbcr_enc;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
792
dev->tpg.hsv_enc = mp->hsv_enc;
drivers/media/test-drivers/vivid/vivid-vid-cap.c
793
dev->tpg.quantization = mp->quantization;
drivers/media/test-drivers/vivid/vivid-vid-common.c
865
struct v4l2_pix_format_mplane *mp = &mp_fmt->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-common.c
866
struct v4l2_plane_pix_format *ppix = &mp->plane_fmt[0];
drivers/media/test-drivers/vivid/vivid-vid-common.c
870
memset(mp->reserved, 0, sizeof(mp->reserved));
drivers/media/test-drivers/vivid/vivid-vid-common.c
873
mp->width = pix->width;
drivers/media/test-drivers/vivid/vivid-vid-common.c
874
mp->height = pix->height;
drivers/media/test-drivers/vivid/vivid-vid-common.c
875
mp->pixelformat = pix->pixelformat;
drivers/media/test-drivers/vivid/vivid-vid-common.c
876
mp->field = pix->field;
drivers/media/test-drivers/vivid/vivid-vid-common.c
877
mp->colorspace = pix->colorspace;
drivers/media/test-drivers/vivid/vivid-vid-common.c
878
mp->xfer_func = pix->xfer_func;
drivers/media/test-drivers/vivid/vivid-vid-common.c
880
mp->ycbcr_enc = pix->ycbcr_enc;
drivers/media/test-drivers/vivid/vivid-vid-common.c
881
mp->quantization = pix->quantization;
drivers/media/test-drivers/vivid/vivid-vid-common.c
882
mp->num_planes = 1;
drivers/media/test-drivers/vivid/vivid-vid-common.c
883
mp->flags = pix->flags;
drivers/media/test-drivers/vivid/vivid-vid-common.c
893
struct v4l2_pix_format_mplane *mp = &fmt.fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-common.c
894
struct v4l2_plane_pix_format *ppix = &mp->plane_fmt[0];
drivers/media/test-drivers/vivid/vivid-vid-common.c
903
pix->width = mp->width;
drivers/media/test-drivers/vivid/vivid-vid-common.c
904
pix->height = mp->height;
drivers/media/test-drivers/vivid/vivid-vid-common.c
905
pix->pixelformat = mp->pixelformat;
drivers/media/test-drivers/vivid/vivid-vid-common.c
906
pix->field = mp->field;
drivers/media/test-drivers/vivid/vivid-vid-common.c
907
pix->colorspace = mp->colorspace;
drivers/media/test-drivers/vivid/vivid-vid-common.c
908
pix->xfer_func = mp->xfer_func;
drivers/media/test-drivers/vivid/vivid-vid-common.c
910
pix->ycbcr_enc = mp->ycbcr_enc;
drivers/media/test-drivers/vivid/vivid-vid-common.c
911
pix->quantization = mp->quantization;
drivers/media/test-drivers/vivid/vivid-vid-common.c
914
pix->flags = mp->flags;
drivers/media/test-drivers/vivid/vivid-vid-out.c
315
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-out.c
319
mp->width = dev->fmt_out_rect.width;
drivers/media/test-drivers/vivid/vivid-vid-out.c
320
mp->height = dev->fmt_out_rect.height;
drivers/media/test-drivers/vivid/vivid-vid-out.c
321
mp->field = dev->field_out;
drivers/media/test-drivers/vivid/vivid-vid-out.c
322
mp->pixelformat = fmt->fourcc;
drivers/media/test-drivers/vivid/vivid-vid-out.c
323
mp->colorspace = dev->colorspace_out;
drivers/media/test-drivers/vivid/vivid-vid-out.c
324
mp->xfer_func = dev->xfer_func_out;
drivers/media/test-drivers/vivid/vivid-vid-out.c
325
mp->ycbcr_enc = dev->ycbcr_enc_out;
drivers/media/test-drivers/vivid/vivid-vid-out.c
326
mp->quantization = dev->quantization_out;
drivers/media/test-drivers/vivid/vivid-vid-out.c
327
mp->num_planes = fmt->buffers;
drivers/media/test-drivers/vivid/vivid-vid-out.c
328
for (p = 0; p < mp->num_planes; p++) {
drivers/media/test-drivers/vivid/vivid-vid-out.c
329
mp->plane_fmt[p].bytesperline = dev->bytesperline_out[p];
drivers/media/test-drivers/vivid/vivid-vid-out.c
330
mp->plane_fmt[p].sizeimage =
drivers/media/test-drivers/vivid/vivid-vid-out.c
331
mp->plane_fmt[p].bytesperline * mp->height /
drivers/media/test-drivers/vivid/vivid-vid-out.c
337
mp->plane_fmt[0].sizeimage +=
drivers/media/test-drivers/vivid/vivid-vid-out.c
338
(stride * mp->height) / fmt->vdownsampling[p];
drivers/media/test-drivers/vivid/vivid-vid-out.c
348
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-out.c
349
struct v4l2_plane_pix_format *pfmt = mp->plane_fmt;
drivers/media/test-drivers/vivid/vivid-vid-out.c
356
fmt = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-out.c
359
mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-out.c
360
mp->pixelformat = V4L2_PIX_FMT_YUYV;
drivers/media/test-drivers/vivid/vivid-vid-out.c
361
fmt = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-out.c
364
mp->field = vivid_field_out(dev, mp->field);
drivers/media/test-drivers/vivid/vivid-vid-out.c
372
if (V4L2_FIELD_HAS_T_OR_B(mp->field))
drivers/media/test-drivers/vivid/vivid-vid-out.c
375
mp->width = w;
drivers/media/test-drivers/vivid/vivid-vid-out.c
376
mp->height = h / factor;
drivers/media/test-drivers/vivid/vivid-vid-out.c
378
struct v4l2_rect r = { 0, 0, mp->width, mp->height * factor };
drivers/media/test-drivers/vivid/vivid-vid-out.c
391
mp->width = r.width;
drivers/media/test-drivers/vivid/vivid-vid-out.c
392
mp->height = r.height / factor;
drivers/media/test-drivers/vivid/vivid-vid-out.c
397
mp->num_planes = fmt->buffers;
drivers/media/test-drivers/vivid/vivid-vid-out.c
400
bytesperline = (mp->width * fmt->bit_depth[p]) >> 3;
drivers/media/test-drivers/vivid/vivid-vid-out.c
409
pfmt[p].sizeimage = (pfmt[p].bytesperline * mp->height) /
drivers/media/test-drivers/vivid/vivid-vid-out.c
415
pfmt[0].sizeimage += (pfmt[0].bytesperline * mp->height *
drivers/media/test-drivers/vivid/vivid-vid-out.c
419
mp->xfer_func = V4L2_XFER_FUNC_DEFAULT;
drivers/media/test-drivers/vivid/vivid-vid-out.c
420
mp->ycbcr_enc = V4L2_YCBCR_ENC_DEFAULT;
drivers/media/test-drivers/vivid/vivid-vid-out.c
421
mp->quantization = V4L2_QUANTIZATION_DEFAULT;
drivers/media/test-drivers/vivid/vivid-vid-out.c
423
mp->colorspace = V4L2_COLORSPACE_SMPTE170M;
drivers/media/test-drivers/vivid/vivid-vid-out.c
425
mp->colorspace = V4L2_COLORSPACE_SRGB;
drivers/media/test-drivers/vivid/vivid-vid-out.c
427
mp->quantization = V4L2_QUANTIZATION_LIM_RANGE;
drivers/media/test-drivers/vivid/vivid-vid-out.c
429
mp->colorspace = V4L2_COLORSPACE_SMPTE170M;
drivers/media/test-drivers/vivid/vivid-vid-out.c
430
} else if (mp->colorspace != V4L2_COLORSPACE_SMPTE170M &&
drivers/media/test-drivers/vivid/vivid-vid-out.c
431
mp->colorspace != V4L2_COLORSPACE_REC709 &&
drivers/media/test-drivers/vivid/vivid-vid-out.c
432
mp->colorspace != V4L2_COLORSPACE_OPRGB &&
drivers/media/test-drivers/vivid/vivid-vid-out.c
433
mp->colorspace != V4L2_COLORSPACE_BT2020 &&
drivers/media/test-drivers/vivid/vivid-vid-out.c
434
mp->colorspace != V4L2_COLORSPACE_SRGB) {
drivers/media/test-drivers/vivid/vivid-vid-out.c
435
mp->colorspace = V4L2_COLORSPACE_REC709;
drivers/media/test-drivers/vivid/vivid-vid-out.c
437
memset(mp->reserved, 0, sizeof(mp->reserved));
drivers/media/test-drivers/vivid/vivid-vid-out.c
444
struct v4l2_pix_format_mplane *mp = &f->fmt.pix_mp;
drivers/media/test-drivers/vivid/vivid-vid-out.c
458
mp->width != dev->fmt_out_rect.width ||
drivers/media/test-drivers/vivid/vivid-vid-out.c
459
mp->height != dev->fmt_out_rect.height ||
drivers/media/test-drivers/vivid/vivid-vid-out.c
460
mp->pixelformat != dev->fmt_out->fourcc ||
drivers/media/test-drivers/vivid/vivid-vid-out.c
461
mp->field != dev->field_out)) {
drivers/media/test-drivers/vivid/vivid-vid-out.c
474
dev->fmt_out = vivid_get_format(dev, mp->pixelformat);
drivers/media/test-drivers/vivid/vivid-vid-out.c
475
if (V4L2_FIELD_HAS_T_OR_B(mp->field))
drivers/media/test-drivers/vivid/vivid-vid-out.c
479
struct v4l2_rect r = { 0, 0, mp->width, mp->height };
drivers/media/test-drivers/vivid/vivid-vid-out.c
538
struct v4l2_rect r = { 0, 0, mp->width, mp->height };
drivers/media/test-drivers/vivid/vivid-vid-out.c
545
dev->fmt_out_rect.width = mp->width;
drivers/media/test-drivers/vivid/vivid-vid-out.c
546
dev->fmt_out_rect.height = mp->height;
drivers/media/test-drivers/vivid/vivid-vid-out.c
547
for (p = 0; p < mp->num_planes; p++)
drivers/media/test-drivers/vivid/vivid-vid-out.c
548
dev->bytesperline_out[p] = mp->plane_fmt[p].bytesperline;
drivers/media/test-drivers/vivid/vivid-vid-out.c
553
dev->field_out = mp->field;
drivers/media/test-drivers/vivid/vivid-vid-out.c
555
dev->tv_field_out = mp->field;
drivers/media/test-drivers/vivid/vivid-vid-out.c
558
dev->colorspace_out = mp->colorspace;
drivers/media/test-drivers/vivid/vivid-vid-out.c
559
dev->xfer_func_out = mp->xfer_func;
drivers/media/test-drivers/vivid/vivid-vid-out.c
560
dev->ycbcr_enc_out = mp->ycbcr_enc;
drivers/media/test-drivers/vivid/vivid-vid-out.c
561
dev->quantization_out = mp->quantization;
drivers/media/usb/pvrusb2/pvrusb2-context.c
101
static void pvr2_context_check(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
105
"pvr2_context %p (notify)", mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
106
if (!mp->initialized_flag && !mp->disconnect_flag) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
107
mp->initialized_flag = !0;
drivers/media/usb/pvrusb2/pvrusb2-context.c
109
"pvr2_context %p (initialize)", mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
111
if (pvr2_hdw_initialize(mp->hdw, pvr2_context_notify, mp)) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
112
mp->video_stream.stream =
drivers/media/usb/pvrusb2/pvrusb2-context.c
113
pvr2_hdw_get_video_stream(mp->hdw);
drivers/media/usb/pvrusb2/pvrusb2-context.c
117
if (mp->setup_func) mp->setup_func(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
121
mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
130
for (ch1 = mp->mc_first; ch1; ch1 = ch2) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
135
if (mp->disconnect_flag && !mp->mc_first) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
137
pvr2_context_destroy(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
151
struct pvr2_context *mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
156
while ((mp = pvr2_context_notify_first) != NULL) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
157
pvr2_context_set_notify(mp, 0);
drivers/media/usb/pvrusb2/pvrusb2-context.c
158
pvr2_context_check(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
206
struct pvr2_context *mp = NULL;
drivers/media/usb/pvrusb2/pvrusb2-context.c
207
mp = kzalloc_obj(*mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
208
if (!mp) goto done;
drivers/media/usb/pvrusb2/pvrusb2-context.c
209
pvr2_trace(PVR2_TRACE_CTXT,"pvr2_context %p (create)",mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
210
mp->setup_func = setup_func;
drivers/media/usb/pvrusb2/pvrusb2-context.c
211
mutex_init(&mp->mutex);
drivers/media/usb/pvrusb2/pvrusb2-context.c
213
mp->exist_prev = pvr2_context_exist_last;
drivers/media/usb/pvrusb2/pvrusb2-context.c
214
mp->exist_next = NULL;
drivers/media/usb/pvrusb2/pvrusb2-context.c
215
pvr2_context_exist_last = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
216
if (mp->exist_prev) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
217
mp->exist_prev->exist_next = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
219
pvr2_context_exist_first = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
222
mp->hdw = pvr2_hdw_create(intf,devid);
drivers/media/usb/pvrusb2/pvrusb2-context.c
223
if (!mp->hdw) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
224
pvr2_context_destroy(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
225
mp = NULL;
drivers/media/usb/pvrusb2/pvrusb2-context.c
228
pvr2_context_set_notify(mp, !0);
drivers/media/usb/pvrusb2/pvrusb2-context.c
230
return mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
234
static void pvr2_context_reset_input_limits(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
238
struct pvr2_hdw *hdw = mp->hdw;
drivers/media/usb/pvrusb2/pvrusb2-context.c
241
for (cp = mp->mc_first; cp; cp = cp->mc_next) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
250
static void pvr2_context_enter(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
252
mutex_lock(&mp->mutex);
drivers/media/usb/pvrusb2/pvrusb2-context.c
256
static void pvr2_context_exit(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
259
if (!(mp->mc_first || !mp->disconnect_flag)) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
262
mutex_unlock(&mp->mutex);
drivers/media/usb/pvrusb2/pvrusb2-context.c
263
if (destroy_flag) pvr2_context_notify(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
267
void pvr2_context_disconnect(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
269
pvr2_hdw_disconnect(mp->hdw);
drivers/media/usb/pvrusb2/pvrusb2-context.c
271
pvr2_context_notify(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
272
mp->disconnect_flag = !0;
drivers/media/usb/pvrusb2/pvrusb2-context.c
276
void pvr2_channel_init(struct pvr2_channel *cp,struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
278
pvr2_context_enter(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
279
cp->hdw = mp->hdw;
drivers/media/usb/pvrusb2/pvrusb2-context.c
280
cp->mc_head = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
282
cp->mc_prev = mp->mc_last;
drivers/media/usb/pvrusb2/pvrusb2-context.c
283
if (mp->mc_last) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
284
mp->mc_last->mc_next = cp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
286
mp->mc_first = cp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
288
mp->mc_last = cp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
289
pvr2_context_exit(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
30
static void pvr2_context_set_notify(struct pvr2_context *mp, int fl)
drivers/media/usb/pvrusb2/pvrusb2-context.c
304
struct pvr2_context *mp = cp->mc_head;
drivers/media/usb/pvrusb2/pvrusb2-context.c
305
pvr2_context_enter(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
308
pvr2_context_reset_input_limits(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
312
mp->mc_last = cp->mc_prev;
drivers/media/usb/pvrusb2/pvrusb2-context.c
317
mp->mc_first = cp->mc_next;
drivers/media/usb/pvrusb2/pvrusb2-context.c
320
pvr2_context_exit(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
35
if (!mp->notify_flag) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
37
mp->notify_prev = pvr2_context_notify_last;
drivers/media/usb/pvrusb2/pvrusb2-context.c
38
mp->notify_next = NULL;
drivers/media/usb/pvrusb2/pvrusb2-context.c
39
pvr2_context_notify_last = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
40
if (mp->notify_prev) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
41
mp->notify_prev->notify_next = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
43
pvr2_context_notify_first = mp;
drivers/media/usb/pvrusb2/pvrusb2-context.c
45
mp->notify_flag = !0;
drivers/media/usb/pvrusb2/pvrusb2-context.c
48
if (mp->notify_flag) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
49
mp->notify_flag = 0;
drivers/media/usb/pvrusb2/pvrusb2-context.c
50
if (mp->notify_next) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
51
mp->notify_next->notify_prev = mp->notify_prev;
drivers/media/usb/pvrusb2/pvrusb2-context.c
53
pvr2_context_notify_last = mp->notify_prev;
drivers/media/usb/pvrusb2/pvrusb2-context.c
55
if (mp->notify_prev) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
56
mp->notify_prev->notify_next = mp->notify_next;
drivers/media/usb/pvrusb2/pvrusb2-context.c
58
pvr2_context_notify_first = mp->notify_next;
drivers/media/usb/pvrusb2/pvrusb2-context.c
67
static void pvr2_context_destroy(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-context.c
69
pvr2_trace(PVR2_TRACE_CTXT,"pvr2_context %p (destroy)",mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
70
pvr2_hdw_destroy(mp->hdw);
drivers/media/usb/pvrusb2/pvrusb2-context.c
71
pvr2_context_set_notify(mp, 0);
drivers/media/usb/pvrusb2/pvrusb2-context.c
73
if (mp->exist_next) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
74
mp->exist_next->exist_prev = mp->exist_prev;
drivers/media/usb/pvrusb2/pvrusb2-context.c
76
pvr2_context_exist_last = mp->exist_prev;
drivers/media/usb/pvrusb2/pvrusb2-context.c
78
if (mp->exist_prev) {
drivers/media/usb/pvrusb2/pvrusb2-context.c
79
mp->exist_prev->exist_next = mp->exist_next;
drivers/media/usb/pvrusb2/pvrusb2-context.c
81
pvr2_context_exist_first = mp->exist_next;
drivers/media/usb/pvrusb2/pvrusb2-context.c
89
kfree(mp);
drivers/media/usb/pvrusb2/pvrusb2-context.c
95
struct pvr2_context *mp = ptr;
drivers/media/usb/pvrusb2/pvrusb2-context.c
97
pvr2_context_set_notify(mp,!0);
drivers/media/usb/pvrusb2/pvrusb2-sysfs.c
748
void pvr2_sysfs_create(struct pvr2_context *mp)
drivers/media/usb/pvrusb2/pvrusb2-sysfs.c
755
pvr2_channel_init(&sfp->channel,mp);
drivers/media/usb/pvrusb2/pvrusb2-sysfs.h
16
void pvr2_sysfs_create(struct pvr2_context *mp);
drivers/media/usb/pvrusb2/pvrusb2-sysfs.h
20
static inline void pvr2_sysfs_create(struct pvr2_context *mp) { }
drivers/media/v4l2-core/v4l2-ioctl.c
274
const struct v4l2_pix_format_mplane *mp;
drivers/media/v4l2-core/v4l2-ioctl.c
297
mp = &p->fmt.pix_mp;
drivers/media/v4l2-core/v4l2-ioctl.c
298
pixelformat = mp->pixelformat;
drivers/media/v4l2-core/v4l2-ioctl.c
300
mp->width, mp->height, &pixelformat,
drivers/media/v4l2-core/v4l2-ioctl.c
301
prt_names(mp->field, v4l2_field_names),
drivers/media/v4l2-core/v4l2-ioctl.c
302
mp->colorspace, mp->num_planes, mp->flags,
drivers/media/v4l2-core/v4l2-ioctl.c
303
mp->ycbcr_enc, mp->quantization, mp->xfer_func);
drivers/media/v4l2-core/v4l2-ioctl.c
304
planes = min_t(u32, mp->num_planes, VIDEO_MAX_PLANES);
drivers/media/v4l2-core/v4l2-ioctl.c
307
mp->plane_fmt[i].bytesperline,
drivers/media/v4l2-core/v4l2-ioctl.c
308
mp->plane_fmt[i].sizeimage);
drivers/mtd/devices/ms02-nv.c
107
struct ms02nv_private *mp;
drivers/mtd/devices/ms02-nv.c
144
mp = kzalloc_obj(*mp);
drivers/mtd/devices/ms02-nv.c
145
if (!mp)
drivers/mtd/devices/ms02-nv.c
148
mtd->priv = mp;
drivers/mtd/devices/ms02-nv.c
149
mp->resource.module = mod_res;
drivers/mtd/devices/ms02-nv.c
162
mp->resource.diag_ram = diag_res;
drivers/mtd/devices/ms02-nv.c
175
mp->resource.user_ram = user_res;
drivers/mtd/devices/ms02-nv.c
188
mp->resource.csr = csr_res;
drivers/mtd/devices/ms02-nv.c
190
mp->addr = phys_to_virt(addr);
drivers/mtd/devices/ms02-nv.c
191
mp->size = size;
drivers/mtd/devices/ms02-nv.c
199
mp->uaddr = phys_to_virt(fixaddr);
drivers/mtd/devices/ms02-nv.c
220
mp->next = root_ms02nv_mtd;
drivers/mtd/devices/ms02-nv.c
236
kfree(mp);
drivers/mtd/devices/ms02-nv.c
249
struct ms02nv_private *mp = mtd->priv;
drivers/mtd/devices/ms02-nv.c
251
root_ms02nv_mtd = mp->next;
drivers/mtd/devices/ms02-nv.c
255
release_resource(mp->resource.csr);
drivers/mtd/devices/ms02-nv.c
256
kfree(mp->resource.csr);
drivers/mtd/devices/ms02-nv.c
257
release_resource(mp->resource.user_ram);
drivers/mtd/devices/ms02-nv.c
258
kfree(mp->resource.user_ram);
drivers/mtd/devices/ms02-nv.c
259
release_resource(mp->resource.diag_ram);
drivers/mtd/devices/ms02-nv.c
260
kfree(mp->resource.diag_ram);
drivers/mtd/devices/ms02-nv.c
261
release_resource(mp->resource.module);
drivers/mtd/devices/ms02-nv.c
262
kfree(mp->resource.module);
drivers/mtd/devices/ms02-nv.c
263
kfree(mp);
drivers/mtd/devices/ms02-nv.c
56
struct ms02nv_private *mp = mtd->priv;
drivers/mtd/devices/ms02-nv.c
58
memcpy(buf, mp->uaddr + from, len);
drivers/mtd/devices/ms02-nv.c
66
struct ms02nv_private *mp = mtd->priv;
drivers/mtd/devices/ms02-nv.c
68
memcpy(mp->uaddr + to, buf, len);
drivers/net/dsa/dsa_loop.c
93
enum dsa_tag_protocol mp)
drivers/net/dsa/hirschmann/hellcreek.c
242
enum dsa_tag_protocol mp)
drivers/net/dsa/ks8995.c
533
enum dsa_tag_protocol mp)
drivers/net/dsa/lan9303-core.c
939
enum dsa_tag_protocol mp)
drivers/net/dsa/lantiq/lantiq_gswip_common.c
751
enum dsa_tag_protocol mp)
drivers/net/dsa/microchip/ksz_common.c
3532
enum dsa_tag_protocol mp)
drivers/net/dsa/mt7530.c
2087
enum dsa_tag_protocol mp)
drivers/net/dsa/ocelot/felix.c
803
enum dsa_tag_protocol mp)
drivers/net/dsa/qca/qca8k-8xxx.c
1747
enum dsa_tag_protocol mp)
drivers/net/dsa/realtek/rtl8365mb.c
853
enum dsa_tag_protocol mp)
drivers/net/dsa/realtek/rtl8366rb.c
1026
enum dsa_tag_protocol mp)
drivers/net/dsa/rzn1_a5psw.c
97
enum dsa_tag_protocol mp)
drivers/net/dsa/sja1105/sja1105_main.c
2357
enum dsa_tag_protocol mp)
drivers/net/dsa/vitesse-vsc73xx-core.c
688
enum dsa_tag_protocol mp)
drivers/net/ethernet/apple/mace.c
112
struct mace_data *mp;
drivers/net/ethernet/apple/mace.c
155
mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
156
mp->mdev = mdev;
drivers/net/ethernet/apple/mace.c
160
mp->mace = ioremap(dev->base_addr, 0x1000);
drivers/net/ethernet/apple/mace.c
161
if (mp->mace == NULL) {
drivers/net/ethernet/apple/mace.c
173
mp->chipid = (in_8(&mp->mace->chipid_hi) << 8) |
drivers/net/ethernet/apple/mace.c
174
in_8(&mp->mace->chipid_lo);
drivers/net/ethernet/apple/mace.c
177
mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
178
mp->maccc = ENXMT | ENRCV;
drivers/net/ethernet/apple/mace.c
180
mp->tx_dma = ioremap(macio_resource_start(mdev, 1), 0x1000);
drivers/net/ethernet/apple/mace.c
181
if (mp->tx_dma == NULL) {
drivers/net/ethernet/apple/mace.c
186
mp->tx_dma_intr = macio_irq(mdev, 1);
drivers/net/ethernet/apple/mace.c
188
mp->rx_dma = ioremap(macio_resource_start(mdev, 2), 0x1000);
drivers/net/ethernet/apple/mace.c
189
if (mp->rx_dma == NULL) {
drivers/net/ethernet/apple/mace.c
194
mp->rx_dma_intr = macio_irq(mdev, 2);
drivers/net/ethernet/apple/mace.c
196
mp->tx_cmds = (volatile struct dbdma_cmd *) DBDMA_ALIGN(mp + 1);
drivers/net/ethernet/apple/mace.c
197
mp->rx_cmds = mp->tx_cmds + NCMDS_TX * N_TX_RING + 1;
drivers/net/ethernet/apple/mace.c
199
memset((char *) mp->tx_cmds, 0,
drivers/net/ethernet/apple/mace.c
201
timer_setup(&mp->tx_timeout, mace_tx_timeout, 0);
drivers/net/ethernet/apple/mace.c
202
spin_lock_init(&mp->lock);
drivers/net/ethernet/apple/mace.c
203
mp->timeout_active = 0;
drivers/net/ethernet/apple/mace.c
206
mp->port_aaui = port_aaui;
drivers/net/ethernet/apple/mace.c
210
mp->port_aaui = 1;
drivers/net/ethernet/apple/mace.c
213
mp->port_aaui = 1;
drivers/net/ethernet/apple/mace.c
215
mp->port_aaui = 0;
drivers/net/ethernet/apple/mace.c
232
rc = request_irq(mp->tx_dma_intr, mace_txdma_intr, 0, "MACE-txdma", dev);
drivers/net/ethernet/apple/mace.c
234
printk(KERN_ERR "MACE: can't get irq %d\n", mp->tx_dma_intr);
drivers/net/ethernet/apple/mace.c
237
rc = request_irq(mp->rx_dma_intr, mace_rxdma_intr, 0, "MACE-rxdma", dev);
drivers/net/ethernet/apple/mace.c
239
printk(KERN_ERR "MACE: can't get irq %d\n", mp->rx_dma_intr);
drivers/net/ethernet/apple/mace.c
251
mp->chipid >> 8, mp->chipid & 0xff);
drivers/net/ethernet/apple/mace.c
262
iounmap(mp->rx_dma);
drivers/net/ethernet/apple/mace.c
264
iounmap(mp->tx_dma);
drivers/net/ethernet/apple/mace.c
266
iounmap(mp->mace);
drivers/net/ethernet/apple/mace.c
278
struct mace_data *mp;
drivers/net/ethernet/apple/mace.c
284
mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
289
free_irq(mp->tx_dma_intr, dev);
drivers/net/ethernet/apple/mace.c
290
free_irq(mp->rx_dma_intr, dev);
drivers/net/ethernet/apple/mace.c
292
iounmap(mp->rx_dma);
drivers/net/ethernet/apple/mace.c
293
iounmap(mp->tx_dma);
drivers/net/ethernet/apple/mace.c
294
iounmap(mp->mace);
drivers/net/ethernet/apple/mace.c
318
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
319
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
351
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
362
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
365
if (mp->port_aaui)
drivers/net/ethernet/apple/mace.c
373
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
374
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
380
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
392
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
398
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
399
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
402
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
407
out_8(&mb->maccc, mp->maccc);
drivers/net/ethernet/apple/mace.c
409
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
413
static inline void mace_clean_rings(struct mace_data *mp)
drivers/net/ethernet/apple/mace.c
419
if (mp->rx_bufs[i] != NULL) {
drivers/net/ethernet/apple/mace.c
420
dev_kfree_skb(mp->rx_bufs[i]);
drivers/net/ethernet/apple/mace.c
421
mp->rx_bufs[i] = NULL;
drivers/net/ethernet/apple/mace.c
424
for (i = mp->tx_empty; i != mp->tx_fill; ) {
drivers/net/ethernet/apple/mace.c
425
dev_kfree_skb(mp->tx_bufs[i]);
drivers/net/ethernet/apple/mace.c
433
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
434
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
435
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
436
volatile struct dbdma_regs __iomem *td = mp->tx_dma;
drivers/net/ethernet/apple/mace.c
446
mace_clean_rings(mp);
drivers/net/ethernet/apple/mace.c
447
memset((char *)mp->rx_cmds, 0, N_RX_RING * sizeof(struct dbdma_cmd));
drivers/net/ethernet/apple/mace.c
448
cp = mp->rx_cmds;
drivers/net/ethernet/apple/mace.c
457
mp->rx_bufs[i] = skb;
drivers/net/ethernet/apple/mace.c
464
mp->rx_bufs[i] = NULL;
drivers/net/ethernet/apple/mace.c
466
mp->rx_fill = i;
drivers/net/ethernet/apple/mace.c
467
mp->rx_empty = 0;
drivers/net/ethernet/apple/mace.c
472
cp->cmd_dep = cpu_to_le32(virt_to_bus(mp->rx_cmds));
drivers/net/ethernet/apple/mace.c
476
out_le32(&rd->cmdptr, virt_to_bus(mp->rx_cmds));
drivers/net/ethernet/apple/mace.c
480
cp = mp->tx_cmds + NCMDS_TX * N_TX_RING;
drivers/net/ethernet/apple/mace.c
482
cp->cmd_dep = cpu_to_le32(virt_to_bus(mp->tx_cmds));
drivers/net/ethernet/apple/mace.c
486
out_le32(&td->cmdptr, virt_to_bus(mp->tx_cmds));
drivers/net/ethernet/apple/mace.c
487
mp->tx_fill = 0;
drivers/net/ethernet/apple/mace.c
488
mp->tx_empty = 0;
drivers/net/ethernet/apple/mace.c
489
mp->tx_fullup = 0;
drivers/net/ethernet/apple/mace.c
490
mp->tx_active = 0;
drivers/net/ethernet/apple/mace.c
491
mp->tx_bad_runt = 0;
drivers/net/ethernet/apple/mace.c
494
out_8(&mb->maccc, mp->maccc);
drivers/net/ethernet/apple/mace.c
503
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
504
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
505
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
506
volatile struct dbdma_regs __iomem *td = mp->tx_dma;
drivers/net/ethernet/apple/mace.c
516
mace_clean_rings(mp);
drivers/net/ethernet/apple/mace.c
523
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
525
if (mp->timeout_active)
drivers/net/ethernet/apple/mace.c
526
timer_delete(&mp->tx_timeout);
drivers/net/ethernet/apple/mace.c
527
mp->tx_timeout.expires = jiffies + TX_TIMEOUT;
drivers/net/ethernet/apple/mace.c
528
add_timer(&mp->tx_timeout);
drivers/net/ethernet/apple/mace.c
529
mp->timeout_active = 1;
drivers/net/ethernet/apple/mace.c
534
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
535
volatile struct dbdma_regs __iomem *td = mp->tx_dma;
drivers/net/ethernet/apple/mace.c
541
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
542
fill = mp->tx_fill;
drivers/net/ethernet/apple/mace.c
546
if (next == mp->tx_empty) {
drivers/net/ethernet/apple/mace.c
548
mp->tx_fullup = 1;
drivers/net/ethernet/apple/mace.c
549
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
552
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
560
mp->tx_bufs[fill] = skb;
drivers/net/ethernet/apple/mace.c
561
cp = mp->tx_cmds + NCMDS_TX * fill;
drivers/net/ethernet/apple/mace.c
565
np = mp->tx_cmds + NCMDS_TX * next;
drivers/net/ethernet/apple/mace.c
569
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
570
mp->tx_fill = next;
drivers/net/ethernet/apple/mace.c
571
if (!mp->tx_bad_runt && mp->tx_active < MAX_TX_ACTIVE) {
drivers/net/ethernet/apple/mace.c
575
++mp->tx_active;
drivers/net/ethernet/apple/mace.c
580
if (next == mp->tx_empty)
drivers/net/ethernet/apple/mace.c
582
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
589
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
590
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
595
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
596
mp->maccc &= ~PROM;
drivers/net/ethernet/apple/mace.c
598
mp->maccc |= PROM;
drivers/net/ethernet/apple/mace.c
622
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
631
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/mace.c
635
out_8(&mb->maccc, mp->maccc);
drivers/net/ethernet/apple/mace.c
636
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
639
static void mace_handle_misc_intrs(struct mace_data *mp, int intr, struct net_device *dev)
drivers/net/ethernet/apple/mace.c
641
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
663
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
664
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
665
volatile struct dbdma_regs __iomem *td = mp->tx_dma;
drivers/net/ethernet/apple/mace.c
672
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
675
mace_handle_misc_intrs(mp, intr, dev);
drivers/net/ethernet/apple/mace.c
677
i = mp->tx_empty;
drivers/net/ethernet/apple/mace.c
679
timer_delete(&mp->tx_timeout);
drivers/net/ethernet/apple/mace.c
680
mp->timeout_active = 0;
drivers/net/ethernet/apple/mace.c
688
mace_handle_misc_intrs(mp, intr, dev);
drivers/net/ethernet/apple/mace.c
689
if (mp->tx_bad_runt) {
drivers/net/ethernet/apple/mace.c
691
mp->tx_bad_runt = 0;
drivers/net/ethernet/apple/mace.c
727
cp = mp->tx_cmds + NCMDS_TX * i;
drivers/net/ethernet/apple/mace.c
738
mp->tx_bad_runt = 1;
drivers/net/ethernet/apple/mace.c
755
if (i == mp->tx_fill) {
drivers/net/ethernet/apple/mace.c
768
dev->stats.tx_bytes += mp->tx_bufs[i]->len;
drivers/net/ethernet/apple/mace.c
771
dev_consume_skb_irq(mp->tx_bufs[i]);
drivers/net/ethernet/apple/mace.c
772
--mp->tx_active;
drivers/net/ethernet/apple/mace.c
781
if (i != mp->tx_empty) {
drivers/net/ethernet/apple/mace.c
782
mp->tx_fullup = 0;
drivers/net/ethernet/apple/mace.c
785
mp->tx_empty = i;
drivers/net/ethernet/apple/mace.c
786
i += mp->tx_active;
drivers/net/ethernet/apple/mace.c
789
if (!mp->tx_bad_runt && i != mp->tx_fill && mp->tx_active < MAX_TX_ACTIVE) {
drivers/net/ethernet/apple/mace.c
792
cp = mp->tx_cmds + NCMDS_TX * i;
drivers/net/ethernet/apple/mace.c
795
++mp->tx_active;
drivers/net/ethernet/apple/mace.c
798
} while (i != mp->tx_fill && mp->tx_active < MAX_TX_ACTIVE);
drivers/net/ethernet/apple/mace.c
802
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
808
struct mace_data *mp = timer_container_of(mp, t, tx_timeout);
drivers/net/ethernet/apple/mace.c
809
struct net_device *dev = macio_get_drvdata(mp->mdev);
drivers/net/ethernet/apple/mace.c
810
volatile struct mace __iomem *mb = mp->mace;
drivers/net/ethernet/apple/mace.c
811
volatile struct dbdma_regs __iomem *td = mp->tx_dma;
drivers/net/ethernet/apple/mace.c
812
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
817
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
818
mp->timeout_active = 0;
drivers/net/ethernet/apple/mace.c
819
if (mp->tx_active == 0 && !mp->tx_bad_runt)
drivers/net/ethernet/apple/mace.c
823
mace_handle_misc_intrs(mp, in_8(&mb->ir), dev);
drivers/net/ethernet/apple/mace.c
825
cp = mp->tx_cmds + NCMDS_TX * mp->tx_empty;
drivers/net/ethernet/apple/mace.c
841
i = mp->tx_empty;
drivers/net/ethernet/apple/mace.c
842
mp->tx_active = 0;
drivers/net/ethernet/apple/mace.c
844
if (mp->tx_bad_runt) {
drivers/net/ethernet/apple/mace.c
845
mp->tx_bad_runt = 0;
drivers/net/ethernet/apple/mace.c
846
} else if (i != mp->tx_fill) {
drivers/net/ethernet/apple/mace.c
847
dev_kfree_skb_irq(mp->tx_bufs[i]);
drivers/net/ethernet/apple/mace.c
850
mp->tx_empty = i;
drivers/net/ethernet/apple/mace.c
852
mp->tx_fullup = 0;
drivers/net/ethernet/apple/mace.c
854
if (i != mp->tx_fill) {
drivers/net/ethernet/apple/mace.c
855
cp = mp->tx_cmds + NCMDS_TX * i;
drivers/net/ethernet/apple/mace.c
860
++mp->tx_active;
drivers/net/ethernet/apple/mace.c
866
out_8(&mb->maccc, mp->maccc);
drivers/net/ethernet/apple/mace.c
869
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
880
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/mace.c
881
volatile struct dbdma_regs __iomem *rd = mp->rx_dma;
drivers/net/ethernet/apple/mace.c
890
spin_lock_irqsave(&mp->lock, flags);
drivers/net/ethernet/apple/mace.c
891
for (i = mp->rx_empty; i != mp->rx_fill; ) {
drivers/net/ethernet/apple/mace.c
892
cp = mp->rx_cmds + i;
drivers/net/ethernet/apple/mace.c
898
np = mp->rx_cmds + next;
drivers/net/ethernet/apple/mace.c
899
if (next != mp->rx_fill &&
drivers/net/ethernet/apple/mace.c
909
skb = mp->rx_bufs[i];
drivers/net/ethernet/apple/mace.c
91
static inline void mace_clean_rings(struct mace_data *mp);
drivers/net/ethernet/apple/mace.c
936
mp->rx_bufs[i] = NULL;
drivers/net/ethernet/apple/mace.c
948
mp->rx_empty = i;
drivers/net/ethernet/apple/mace.c
950
i = mp->rx_fill;
drivers/net/ethernet/apple/mace.c
955
if (next == mp->rx_empty)
drivers/net/ethernet/apple/mace.c
957
cp = mp->rx_cmds + i;
drivers/net/ethernet/apple/mace.c
958
skb = mp->rx_bufs[i];
drivers/net/ethernet/apple/mace.c
963
mp->rx_bufs[i] = skb;
drivers/net/ethernet/apple/mace.c
980
if (i != mp->rx_fill) {
drivers/net/ethernet/apple/mace.c
982
mp->rx_fill = i;
drivers/net/ethernet/apple/mace.c
984
spin_unlock_irqrestore(&mp->lock, flags);
drivers/net/ethernet/apple/macmace.c
103
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
106
psc_write_long(PSC_ENETRD_ADDR + set, (u32) mp->rx_ring_phys);
drivers/net/ethernet/apple/macmace.c
109
mp->rx_tail = 0;
drivers/net/ethernet/apple/macmace.c
118
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
119
volatile struct mace *mace = mp->mace;
drivers/net/ethernet/apple/macmace.c
133
mp->rx_slot = 0;
drivers/net/ethernet/apple/macmace.c
145
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
146
volatile struct mace *mace = mp->mace;
drivers/net/ethernet/apple/macmace.c
154
mp->tx_slot = mp->tx_sloti = 0;
drivers/net/ethernet/apple/macmace.c
155
mp->tx_count = N_TX_RING;
drivers/net/ethernet/apple/macmace.c
196
struct mace_data *mp;
drivers/net/ethernet/apple/macmace.c
207
mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
209
mp->device = &pdev->dev;
drivers/net/ethernet/apple/macmace.c
214
mp->mace = MACE_BASE;
drivers/net/ethernet/apple/macmace.c
217
mp->dma_intr = IRQ_MAC_MACE_DMA;
drivers/net/ethernet/apple/macmace.c
219
mp->chipid = mp->mace->chipid_hi << 8 | mp->mace->chipid_lo;
drivers/net/ethernet/apple/macmace.c
249
dev->dev_addr, mp->chipid);
drivers/net/ethernet/apple/macmace.c
265
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
266
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
299
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
310
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
322
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
323
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
329
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
339
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
345
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
346
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
370
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
371
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
380
if (request_irq(mp->dma_intr, mace_dma_intr, 0, dev->name, dev)) {
drivers/net/ethernet/apple/macmace.c
381
printk(KERN_ERR "%s: can't get irq %d\n", dev->name, mp->dma_intr);
drivers/net/ethernet/apple/macmace.c
388
mp->tx_ring = dma_alloc_coherent(mp->device,
drivers/net/ethernet/apple/macmace.c
390
&mp->tx_ring_phys, GFP_KERNEL);
drivers/net/ethernet/apple/macmace.c
391
if (mp->tx_ring == NULL)
drivers/net/ethernet/apple/macmace.c
394
mp->rx_ring = dma_alloc_coherent(mp->device,
drivers/net/ethernet/apple/macmace.c
396
&mp->rx_ring_phys, GFP_KERNEL);
drivers/net/ethernet/apple/macmace.c
397
if (mp->rx_ring == NULL)
drivers/net/ethernet/apple/macmace.c
419
dma_free_coherent(mp->device, N_TX_RING * MACE_BUFF_SIZE,
drivers/net/ethernet/apple/macmace.c
420
mp->tx_ring, mp->tx_ring_phys);
drivers/net/ethernet/apple/macmace.c
423
free_irq(mp->dma_intr, dev);
drivers/net/ethernet/apple/macmace.c
433
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
434
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
449
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
456
if (!mp->tx_count) {
drivers/net/ethernet/apple/macmace.c
461
mp->tx_count--;
drivers/net/ethernet/apple/macmace.c
468
skb_copy_from_linear_data(skb, mp->tx_ring, skb->len);
drivers/net/ethernet/apple/macmace.c
472
psc_write_long(PSC_ENETWR_ADDR + mp->tx_slot, (u32) mp->tx_ring_phys);
drivers/net/ethernet/apple/macmace.c
473
psc_write_long(PSC_ENETWR_LEN + mp->tx_slot, skb->len);
drivers/net/ethernet/apple/macmace.c
474
psc_write_word(PSC_ENETWR_CMD + mp->tx_slot, 0x9800);
drivers/net/ethernet/apple/macmace.c
476
mp->tx_slot ^= 0x10;
drivers/net/ethernet/apple/macmace.c
485
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
486
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
517
if (mp->chipid == BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
526
if (mp->chipid != BROKEN_ADDRCHG_REV)
drivers/net/ethernet/apple/macmace.c
536
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
537
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
559
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
560
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
581
if (!mp->tx_count) {
drivers/net/ethernet/apple/macmace.c
599
if (mp->tx_count)
drivers/net/ethernet/apple/macmace.c
609
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
610
volatile struct mace *mb = mp->mace;
drivers/net/ethernet/apple/macmace.c
624
mp->tx_count = N_TX_RING;
drivers/net/ethernet/apple/macmace.c
679
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
698
psc_write_word(PSC_ENETRD_CMD + mp->rx_slot, 0x1100);
drivers/net/ethernet/apple/macmace.c
700
left = psc_read_long(PSC_ENETRD_LEN + mp->rx_slot);
drivers/net/ethernet/apple/macmace.c
705
while (mp->rx_tail < head) {
drivers/net/ethernet/apple/macmace.c
706
mace_dma_rx_frame(dev, (struct mace_frame*) (mp->rx_ring
drivers/net/ethernet/apple/macmace.c
707
+ (mp->rx_tail * MACE_BUFF_SIZE)));
drivers/net/ethernet/apple/macmace.c
708
mp->rx_tail++;
drivers/net/ethernet/apple/macmace.c
715
mace_load_rxdma_base(dev, mp->rx_slot);
drivers/net/ethernet/apple/macmace.c
716
mp->rx_slot ^= 0x10;
drivers/net/ethernet/apple/macmace.c
718
psc_write_word(PSC_ENETRD_CMD + mp->rx_slot, 0x9800);
drivers/net/ethernet/apple/macmace.c
731
psc_write_word(PSC_ENETWR_CMD + mp->tx_sloti, 0x0100);
drivers/net/ethernet/apple/macmace.c
732
mp->tx_sloti ^= 0x10;
drivers/net/ethernet/apple/macmace.c
733
mp->tx_count++;
drivers/net/ethernet/apple/macmace.c
745
struct mace_data *mp = netdev_priv(dev);
drivers/net/ethernet/apple/macmace.c
752
dma_free_coherent(mp->device, N_RX_RING * MACE_BUFF_SIZE,
drivers/net/ethernet/apple/macmace.c
753
mp->rx_ring, mp->rx_ring_phys);
drivers/net/ethernet/apple/macmace.c
754
dma_free_coherent(mp->device, N_TX_RING * MACE_BUFF_SIZE,
drivers/net/ethernet/apple/macmace.c
755
mp->tx_ring, mp->tx_ring_phys);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
329
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
331
WARN_ON(!mp);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
333
return mp->get_entry(mp, offset);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
338
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
340
WARN_ON(!mp);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
342
return mp->get(mp, 1);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
365
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
368
if (!mp->get(mp, 1))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
372
mp->put(mp, 1);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
381
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
383
return mp->put_entry(mp, offset);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
388
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
390
return mp->put(mp, 1);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
409
struct bnx2x_credit_pool_obj *mp = o->macs_pool;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
412
if (!mp->put(mp, 1))
drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c
416
mp->get(mp, 1);
drivers/net/ethernet/marvell/mv643xx_eth.c
1001
txq = mp->txq + queue;
drivers/net/ethernet/marvell/mv643xx_eth.c
1034
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
1035
struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
1041
if (rdlp(mp, TXQ_COMMAND) & (1 << txq->index))
drivers/net/ethernet/marvell/mv643xx_eth.c
1044
hw_desc_ptr = rdlp(mp, TXQ_CURRENT_DESC_PTR(txq->index));
drivers/net/ethernet/marvell/mv643xx_eth.c
1054
mp->work_tx_end &= ~(1 << txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
1059
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
1060
struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
1094
dma_unmap_page(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
1099
dma_unmap_single(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
1113
netdev_info(mp->dev, "tx error\n");
drivers/net/ethernet/marvell/mv643xx_eth.c
1114
mp->dev->stats.tx_errors++;
drivers/net/ethernet/marvell/mv643xx_eth.c
1122
mp->work_tx &= ~(1 << txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
1133
static void tx_set_rate(struct mv643xx_eth_private *mp, int rate, int burst)
drivers/net/ethernet/marvell/mv643xx_eth.c
1139
token_rate = ((rate / 1000) * 64) / (mp->t_clk / 1000);
drivers/net/ethernet/marvell/mv643xx_eth.c
1143
mtu = (mp->dev->mtu + 255) >> 8;
drivers/net/ethernet/marvell/mv643xx_eth.c
1151
switch (mp->shared->tx_bw_control) {
drivers/net/ethernet/marvell/mv643xx_eth.c
1153
wrlp(mp, TX_BW_RATE, token_rate);
drivers/net/ethernet/marvell/mv643xx_eth.c
1154
wrlp(mp, TX_BW_MTU, mtu);
drivers/net/ethernet/marvell/mv643xx_eth.c
1155
wrlp(mp, TX_BW_BURST, bucket_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
1158
wrlp(mp, TX_BW_RATE_MOVED, token_rate);
drivers/net/ethernet/marvell/mv643xx_eth.c
1159
wrlp(mp, TX_BW_MTU_MOVED, mtu);
drivers/net/ethernet/marvell/mv643xx_eth.c
1160
wrlp(mp, TX_BW_BURST_MOVED, bucket_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
1167
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
1171
token_rate = ((rate / 1000) * 64) / (mp->t_clk / 1000);
drivers/net/ethernet/marvell/mv643xx_eth.c
1179
wrlp(mp, TXQ_BW_TOKENS(txq->index), token_rate << 14);
drivers/net/ethernet/marvell/mv643xx_eth.c
1180
wrlp(mp, TXQ_BW_CONF(txq->index), (bucket_size << 10) | token_rate);
drivers/net/ethernet/marvell/mv643xx_eth.c
1185
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
1193
switch (mp->shared->tx_bw_control) {
drivers/net/ethernet/marvell/mv643xx_eth.c
1203
val = rdlp(mp, off);
drivers/net/ethernet/marvell/mv643xx_eth.c
1205
wrlp(mp, off, val);
drivers/net/ethernet/marvell/mv643xx_eth.c
1213
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1214
u32 pscr = rdlp(mp, PORT_SERIAL_CONTROL);
drivers/net/ethernet/marvell/mv643xx_eth.c
1248
wrlp(mp, PORT_SERIAL_CONTROL, pscr);
drivers/net/ethernet/marvell/mv643xx_eth.c
1254
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1261
for (i = 0; i < mp->txq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
1262
struct tx_queue *txq = mp->txq + i;
drivers/net/ethernet/marvell/mv643xx_eth.c
1276
static inline u32 mib_read(struct mv643xx_eth_private *mp, int offset)
drivers/net/ethernet/marvell/mv643xx_eth.c
1278
return rdl(mp, MIB_COUNTERS(mp->port_num) + offset);
drivers/net/ethernet/marvell/mv643xx_eth.c
1281
static void mib_counters_clear(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
1286
mib_read(mp, i);
drivers/net/ethernet/marvell/mv643xx_eth.c
1289
rdlp(mp, RX_DISCARD_FRAME_CNT);
drivers/net/ethernet/marvell/mv643xx_eth.c
1290
rdlp(mp, RX_OVERRUN_FRAME_CNT);
drivers/net/ethernet/marvell/mv643xx_eth.c
1293
static void mib_counters_update(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
1295
struct mib_counters *p = &mp->mib_counters;
drivers/net/ethernet/marvell/mv643xx_eth.c
1297
spin_lock_bh(&mp->mib_counters_lock);
drivers/net/ethernet/marvell/mv643xx_eth.c
1298
p->good_octets_received += mib_read(mp, 0x00);
drivers/net/ethernet/marvell/mv643xx_eth.c
1299
p->bad_octets_received += mib_read(mp, 0x08);
drivers/net/ethernet/marvell/mv643xx_eth.c
1300
p->internal_mac_transmit_err += mib_read(mp, 0x0c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1301
p->good_frames_received += mib_read(mp, 0x10);
drivers/net/ethernet/marvell/mv643xx_eth.c
1302
p->bad_frames_received += mib_read(mp, 0x14);
drivers/net/ethernet/marvell/mv643xx_eth.c
1303
p->broadcast_frames_received += mib_read(mp, 0x18);
drivers/net/ethernet/marvell/mv643xx_eth.c
1304
p->multicast_frames_received += mib_read(mp, 0x1c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1305
p->frames_64_octets += mib_read(mp, 0x20);
drivers/net/ethernet/marvell/mv643xx_eth.c
1306
p->frames_65_to_127_octets += mib_read(mp, 0x24);
drivers/net/ethernet/marvell/mv643xx_eth.c
1307
p->frames_128_to_255_octets += mib_read(mp, 0x28);
drivers/net/ethernet/marvell/mv643xx_eth.c
1308
p->frames_256_to_511_octets += mib_read(mp, 0x2c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1309
p->frames_512_to_1023_octets += mib_read(mp, 0x30);
drivers/net/ethernet/marvell/mv643xx_eth.c
1310
p->frames_1024_to_max_octets += mib_read(mp, 0x34);
drivers/net/ethernet/marvell/mv643xx_eth.c
1311
p->good_octets_sent += mib_read(mp, 0x38);
drivers/net/ethernet/marvell/mv643xx_eth.c
1312
p->good_frames_sent += mib_read(mp, 0x40);
drivers/net/ethernet/marvell/mv643xx_eth.c
1313
p->excessive_collision += mib_read(mp, 0x44);
drivers/net/ethernet/marvell/mv643xx_eth.c
1314
p->multicast_frames_sent += mib_read(mp, 0x48);
drivers/net/ethernet/marvell/mv643xx_eth.c
1315
p->broadcast_frames_sent += mib_read(mp, 0x4c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1316
p->unrec_mac_control_received += mib_read(mp, 0x50);
drivers/net/ethernet/marvell/mv643xx_eth.c
1317
p->fc_sent += mib_read(mp, 0x54);
drivers/net/ethernet/marvell/mv643xx_eth.c
1318
p->good_fc_received += mib_read(mp, 0x58);
drivers/net/ethernet/marvell/mv643xx_eth.c
1319
p->bad_fc_received += mib_read(mp, 0x5c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1320
p->undersize_received += mib_read(mp, 0x60);
drivers/net/ethernet/marvell/mv643xx_eth.c
1321
p->fragments_received += mib_read(mp, 0x64);
drivers/net/ethernet/marvell/mv643xx_eth.c
1322
p->oversize_received += mib_read(mp, 0x68);
drivers/net/ethernet/marvell/mv643xx_eth.c
1323
p->jabber_received += mib_read(mp, 0x6c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1324
p->mac_receive_error += mib_read(mp, 0x70);
drivers/net/ethernet/marvell/mv643xx_eth.c
1325
p->bad_crc_event += mib_read(mp, 0x74);
drivers/net/ethernet/marvell/mv643xx_eth.c
1326
p->collision += mib_read(mp, 0x78);
drivers/net/ethernet/marvell/mv643xx_eth.c
1327
p->late_collision += mib_read(mp, 0x7c);
drivers/net/ethernet/marvell/mv643xx_eth.c
1329
p->rx_discard += rdlp(mp, RX_DISCARD_FRAME_CNT);
drivers/net/ethernet/marvell/mv643xx_eth.c
1330
p->rx_overrun += rdlp(mp, RX_OVERRUN_FRAME_CNT);
drivers/net/ethernet/marvell/mv643xx_eth.c
1331
spin_unlock_bh(&mp->mib_counters_lock);
drivers/net/ethernet/marvell/mv643xx_eth.c
1336
struct mv643xx_eth_private *mp = timer_container_of(mp, t,
drivers/net/ethernet/marvell/mv643xx_eth.c
1338
mib_counters_update(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
1339
mod_timer(&mp->mib_counters_timer, jiffies + 30 * HZ);
drivers/net/ethernet/marvell/mv643xx_eth.c
1355
static unsigned int get_rx_coal(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
1357
u32 val = rdlp(mp, SDMA_CONFIG);
drivers/net/ethernet/marvell/mv643xx_eth.c
1360
if (mp->shared->extended_rx_coal_limit)
drivers/net/ethernet/marvell/mv643xx_eth.c
1366
temp += mp->t_clk / 2;
drivers/net/ethernet/marvell/mv643xx_eth.c
1367
do_div(temp, mp->t_clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
1372
static void set_rx_coal(struct mv643xx_eth_private *mp, unsigned int usec)
drivers/net/ethernet/marvell/mv643xx_eth.c
1377
temp = (u64)usec * mp->t_clk;
drivers/net/ethernet/marvell/mv643xx_eth.c
1381
val = rdlp(mp, SDMA_CONFIG);
drivers/net/ethernet/marvell/mv643xx_eth.c
1382
if (mp->shared->extended_rx_coal_limit) {
drivers/net/ethernet/marvell/mv643xx_eth.c
1394
wrlp(mp, SDMA_CONFIG, val);
drivers/net/ethernet/marvell/mv643xx_eth.c
1397
static unsigned int get_tx_coal(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
1401
temp = (rdlp(mp, TX_FIFO_URGENT_THRESHOLD) & 0x3fff0) >> 4;
drivers/net/ethernet/marvell/mv643xx_eth.c
1403
temp += mp->t_clk / 2;
drivers/net/ethernet/marvell/mv643xx_eth.c
1404
do_div(temp, mp->t_clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
1409
static void set_tx_coal(struct mv643xx_eth_private *mp, unsigned int usec)
drivers/net/ethernet/marvell/mv643xx_eth.c
1413
temp = (u64)usec * mp->t_clk;
drivers/net/ethernet/marvell/mv643xx_eth.c
1420
wrlp(mp, TX_FIFO_URGENT_THRESHOLD, temp << 4);
drivers/net/ethernet/marvell/mv643xx_eth.c
1484
mv643xx_eth_get_link_ksettings_phy(struct mv643xx_eth_private *mp,
drivers/net/ethernet/marvell/mv643xx_eth.c
1487
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
1503
mv643xx_eth_get_link_ksettings_phyless(struct mv643xx_eth_private *mp,
drivers/net/ethernet/marvell/mv643xx_eth.c
1509
port_status = rdlp(mp, PORT_STATUS);
drivers/net/ethernet/marvell/mv643xx_eth.c
1571
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1574
return mv643xx_eth_get_link_ksettings_phy(mp, cmd);
drivers/net/ethernet/marvell/mv643xx_eth.c
1576
return mv643xx_eth_get_link_ksettings_phyless(mp, cmd);
drivers/net/ethernet/marvell/mv643xx_eth.c
1621
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1623
ec->rx_coalesce_usecs = get_rx_coal(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
1624
ec->tx_coalesce_usecs = get_tx_coal(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
1634
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1636
set_rx_coal(mp, ec->rx_coalesce_usecs);
drivers/net/ethernet/marvell/mv643xx_eth.c
1637
set_tx_coal(mp, ec->tx_coalesce_usecs);
drivers/net/ethernet/marvell/mv643xx_eth.c
1647
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1652
er->rx_pending = mp->rx_ring_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
1653
er->tx_pending = mp->tx_ring_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
1661
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1666
mp->rx_ring_size = min(er->rx_pending, 4096U);
drivers/net/ethernet/marvell/mv643xx_eth.c
1667
mp->tx_ring_size = clamp_t(unsigned int, er->tx_pending,
drivers/net/ethernet/marvell/mv643xx_eth.c
1669
if (mp->tx_ring_size != er->tx_pending)
drivers/net/ethernet/marvell/mv643xx_eth.c
1671
mp->tx_ring_size, er->tx_pending);
drivers/net/ethernet/marvell/mv643xx_eth.c
1689
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1692
wrlp(mp, PORT_CONFIG, rx_csum ? 0x02000000 : 0x00000000);
drivers/net/ethernet/marvell/mv643xx_eth.c
1711
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1715
mib_counters_update(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
1724
p = ((void *)mp->dev) + stat->netdev_off;
drivers/net/ethernet/marvell/mv643xx_eth.c
1726
p = ((void *)mp) + stat->mp_off;
drivers/net/ethernet/marvell/mv643xx_eth.c
1762
static void uc_addr_get(struct mv643xx_eth_private *mp, unsigned char *addr)
drivers/net/ethernet/marvell/mv643xx_eth.c
1764
unsigned int mac_h = rdlp(mp, MAC_ADDR_HIGH);
drivers/net/ethernet/marvell/mv643xx_eth.c
1765
unsigned int mac_l = rdlp(mp, MAC_ADDR_LOW);
drivers/net/ethernet/marvell/mv643xx_eth.c
1775
static void uc_addr_set(struct mv643xx_eth_private *mp, const u8 *addr)
drivers/net/ethernet/marvell/mv643xx_eth.c
1777
wrlp(mp, MAC_ADDR_HIGH,
drivers/net/ethernet/marvell/mv643xx_eth.c
1779
wrlp(mp, MAC_ADDR_LOW, (addr[4] << 8) | addr[5]);
drivers/net/ethernet/marvell/mv643xx_eth.c
1805
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1810
uc_addr_set(mp, dev->dev_addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
1812
port_config = rdlp(mp, PORT_CONFIG) & ~UNICAST_PROMISCUOUS_MODE;
drivers/net/ethernet/marvell/mv643xx_eth.c
1821
int off = UNICAST_TABLE(mp->port_num) + i;
drivers/net/ethernet/marvell/mv643xx_eth.c
1835
wrl(mp, off, v);
drivers/net/ethernet/marvell/mv643xx_eth.c
1838
wrlp(mp, PORT_CONFIG, port_config);
drivers/net/ethernet/marvell/mv643xx_eth.c
1861
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
1893
wrl(mp, SPECIAL_MCAST_TABLE(mp->port_num) + i * sizeof(u32),
drivers/net/ethernet/marvell/mv643xx_eth.c
1895
wrl(mp, OTHER_MCAST_TABLE(mp->port_num) + i * sizeof(u32),
drivers/net/ethernet/marvell/mv643xx_eth.c
1904
wrl(mp, SPECIAL_MCAST_TABLE(mp->port_num) + i * sizeof(u32),
drivers/net/ethernet/marvell/mv643xx_eth.c
1906
wrl(mp, OTHER_MCAST_TABLE(mp->port_num) + i * sizeof(u32),
drivers/net/ethernet/marvell/mv643xx_eth.c
1935
static int rxq_init(struct mv643xx_eth_private *mp, int index)
drivers/net/ethernet/marvell/mv643xx_eth.c
1937
struct rx_queue *rxq = mp->rxq + index;
drivers/net/ethernet/marvell/mv643xx_eth.c
1944
rxq->rx_ring_size = mp->rx_ring_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
1952
if (index == 0 && size <= mp->rx_desc_sram_size) {
drivers/net/ethernet/marvell/mv643xx_eth.c
1953
rxq->rx_desc_area = ioremap(mp->rx_desc_sram_addr,
drivers/net/ethernet/marvell/mv643xx_eth.c
1954
mp->rx_desc_sram_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
1955
rxq->rx_desc_dma = mp->rx_desc_sram_addr;
drivers/net/ethernet/marvell/mv643xx_eth.c
1957
rxq->rx_desc_area = dma_alloc_coherent(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
1963
netdev_err(mp->dev,
drivers/net/ethernet/marvell/mv643xx_eth.c
1990
if (index == 0 && size <= mp->rx_desc_sram_size)
drivers/net/ethernet/marvell/mv643xx_eth.c
1993
dma_free_coherent(mp->dev->dev.parent, size,
drivers/net/ethernet/marvell/mv643xx_eth.c
2003
struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
drivers/net/ethernet/marvell/mv643xx_eth.c
2016
netdev_err(mp->dev, "error freeing rx ring -- %d skbs stuck\n",
drivers/net/ethernet/marvell/mv643xx_eth.c
2021
rxq->rx_desc_area_size <= mp->rx_desc_sram_size)
drivers/net/ethernet/marvell/mv643xx_eth.c
2024
dma_free_coherent(mp->dev->dev.parent, rxq->rx_desc_area_size,
drivers/net/ethernet/marvell/mv643xx_eth.c
2030
static int txq_init(struct mv643xx_eth_private *mp, int index)
drivers/net/ethernet/marvell/mv643xx_eth.c
2032
struct tx_queue *txq = mp->txq + index;
drivers/net/ethernet/marvell/mv643xx_eth.c
2040
txq->tx_ring_size = mp->tx_ring_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
2055
if (index == 0 && size <= mp->tx_desc_sram_size) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2056
txq->tx_desc_area = ioremap(mp->tx_desc_sram_addr,
drivers/net/ethernet/marvell/mv643xx_eth.c
2057
mp->tx_desc_sram_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
2058
txq->tx_desc_dma = mp->tx_desc_sram_addr;
drivers/net/ethernet/marvell/mv643xx_eth.c
2060
txq->tx_desc_area = dma_alloc_coherent(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
2066
netdev_err(mp->dev,
drivers/net/ethernet/marvell/mv643xx_eth.c
2096
txq->tso_hdrs = dma_alloc_coherent(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
2110
if (index == 0 && size <= mp->tx_desc_sram_size)
drivers/net/ethernet/marvell/mv643xx_eth.c
2113
dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size,
drivers/net/ethernet/marvell/mv643xx_eth.c
2120
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
2128
txq->tx_desc_area_size <= mp->tx_desc_sram_size)
drivers/net/ethernet/marvell/mv643xx_eth.c
2131
dma_free_coherent(mp->dev->dev.parent, txq->tx_desc_area_size,
drivers/net/ethernet/marvell/mv643xx_eth.c
2136
dma_free_coherent(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
2143
static int mv643xx_eth_collect_events(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2148
int_cause = rdlp(mp, INT_CAUSE) & mp->int_mask;
drivers/net/ethernet/marvell/mv643xx_eth.c
2155
int_cause_ext = rdlp(mp, INT_CAUSE_EXT);
drivers/net/ethernet/marvell/mv643xx_eth.c
2159
wrlp(mp, INT_CAUSE, ~int_cause);
drivers/net/ethernet/marvell/mv643xx_eth.c
2160
mp->work_tx_end |= ((int_cause & INT_TX_END) >> 19) &
drivers/net/ethernet/marvell/mv643xx_eth.c
2161
~(rdlp(mp, TXQ_COMMAND) & 0xff);
drivers/net/ethernet/marvell/mv643xx_eth.c
2162
mp->work_rx |= (int_cause & INT_RX) >> 2;
drivers/net/ethernet/marvell/mv643xx_eth.c
2167
wrlp(mp, INT_CAUSE_EXT, ~int_cause_ext);
drivers/net/ethernet/marvell/mv643xx_eth.c
2169
mp->work_link = 1;
drivers/net/ethernet/marvell/mv643xx_eth.c
2170
mp->work_tx |= int_cause_ext & INT_EXT_TX;
drivers/net/ethernet/marvell/mv643xx_eth.c
2179
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2181
if (unlikely(!mv643xx_eth_collect_events(mp)))
drivers/net/ethernet/marvell/mv643xx_eth.c
2184
wrlp(mp, INT_MASK, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
2185
napi_schedule(&mp->napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2190
static void handle_link_event(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2192
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
2198
port_status = rdlp(mp, PORT_STATUS);
drivers/net/ethernet/marvell/mv643xx_eth.c
2207
for (i = 0; i < mp->txq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2208
struct tx_queue *txq = mp->txq + i;
drivers/net/ethernet/marvell/mv643xx_eth.c
2243
struct mv643xx_eth_private *mp;
drivers/net/ethernet/marvell/mv643xx_eth.c
2246
mp = container_of(napi, struct mv643xx_eth_private, napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2248
if (unlikely(mp->oom)) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2249
mp->oom = 0;
drivers/net/ethernet/marvell/mv643xx_eth.c
2250
timer_delete(&mp->rx_oom);
drivers/net/ethernet/marvell/mv643xx_eth.c
2259
if (mp->work_link) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2260
mp->work_link = 0;
drivers/net/ethernet/marvell/mv643xx_eth.c
2261
handle_link_event(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2266
queue_mask = mp->work_tx | mp->work_tx_end | mp->work_rx;
drivers/net/ethernet/marvell/mv643xx_eth.c
2267
if (likely(!mp->oom))
drivers/net/ethernet/marvell/mv643xx_eth.c
2268
queue_mask |= mp->work_rx_refill;
drivers/net/ethernet/marvell/mv643xx_eth.c
2271
if (mv643xx_eth_collect_events(mp))
drivers/net/ethernet/marvell/mv643xx_eth.c
2283
if (mp->work_tx_end & queue_mask) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2284
txq_kick(mp->txq + queue);
drivers/net/ethernet/marvell/mv643xx_eth.c
2285
} else if (mp->work_tx & queue_mask) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2286
work_done += txq_reclaim(mp->txq + queue, work_tbd, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
2287
txq_maybe_wake(mp->txq + queue);
drivers/net/ethernet/marvell/mv643xx_eth.c
2288
} else if (mp->work_rx & queue_mask) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2289
work_done += rxq_process(mp->rxq + queue, work_tbd);
drivers/net/ethernet/marvell/mv643xx_eth.c
2290
} else if (!mp->oom && (mp->work_rx_refill & queue_mask)) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2291
work_done += rxq_refill(mp->rxq + queue, work_tbd);
drivers/net/ethernet/marvell/mv643xx_eth.c
2298
if (mp->oom)
drivers/net/ethernet/marvell/mv643xx_eth.c
2299
mod_timer(&mp->rx_oom, jiffies + (HZ / 10));
drivers/net/ethernet/marvell/mv643xx_eth.c
2301
wrlp(mp, INT_MASK, mp->int_mask);
drivers/net/ethernet/marvell/mv643xx_eth.c
2309
struct mv643xx_eth_private *mp = timer_container_of(mp, t, rx_oom);
drivers/net/ethernet/marvell/mv643xx_eth.c
2311
napi_schedule(&mp->napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2314
static void port_start(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2316
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
2336
pscr = rdlp(mp, PORT_SERIAL_CONTROL);
drivers/net/ethernet/marvell/mv643xx_eth.c
2339
wrlp(mp, PORT_SERIAL_CONTROL, pscr);
drivers/net/ethernet/marvell/mv643xx_eth.c
2344
wrlp(mp, PORT_SERIAL_CONTROL, pscr);
drivers/net/ethernet/marvell/mv643xx_eth.c
2349
tx_set_rate(mp, 1000000000, 16777216);
drivers/net/ethernet/marvell/mv643xx_eth.c
2350
for (i = 0; i < mp->txq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2351
struct tx_queue *txq = mp->txq + i;
drivers/net/ethernet/marvell/mv643xx_eth.c
2363
mv643xx_eth_set_features(mp->dev, mp->dev->features);
drivers/net/ethernet/marvell/mv643xx_eth.c
2368
wrlp(mp, PORT_CONFIG_EXT, 0x00000000);
drivers/net/ethernet/marvell/mv643xx_eth.c
2373
mv643xx_eth_program_unicast_filter(mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2378
for (i = 0; i < mp->rxq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2379
struct rx_queue *rxq = mp->rxq + i;
drivers/net/ethernet/marvell/mv643xx_eth.c
2384
wrlp(mp, RXQ_CURRENT_DESC_PTR(i), addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
2390
static void mv643xx_eth_recalc_skb_size(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2400
skb_size = mp->dev->mtu + 36;
drivers/net/ethernet/marvell/mv643xx_eth.c
2407
mp->skb_size = (skb_size + 7) & ~7;
drivers/net/ethernet/marvell/mv643xx_eth.c
2415
mp->skb_size += SKB_DMA_REALIGN;
drivers/net/ethernet/marvell/mv643xx_eth.c
2420
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2424
wrlp(mp, INT_CAUSE, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
2425
wrlp(mp, INT_CAUSE_EXT, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
2426
rdlp(mp, INT_CAUSE_EXT);
drivers/net/ethernet/marvell/mv643xx_eth.c
2435
mv643xx_eth_recalc_skb_size(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2437
napi_enable(&mp->napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2439
mp->int_mask = INT_EXT;
drivers/net/ethernet/marvell/mv643xx_eth.c
2441
for (i = 0; i < mp->rxq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2442
err = rxq_init(mp, i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2445
rxq_deinit(mp->rxq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2449
rxq_refill(mp->rxq + i, INT_MAX);
drivers/net/ethernet/marvell/mv643xx_eth.c
2450
mp->int_mask |= INT_RX_0 << i;
drivers/net/ethernet/marvell/mv643xx_eth.c
2453
if (mp->oom) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2454
mp->rx_oom.expires = jiffies + (HZ / 10);
drivers/net/ethernet/marvell/mv643xx_eth.c
2455
add_timer(&mp->rx_oom);
drivers/net/ethernet/marvell/mv643xx_eth.c
2458
for (i = 0; i < mp->txq_count; i++) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2459
err = txq_init(mp, i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2462
txq_deinit(mp->txq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2465
mp->int_mask |= INT_TX_END_0 << i;
drivers/net/ethernet/marvell/mv643xx_eth.c
2468
add_timer(&mp->mib_counters_timer);
drivers/net/ethernet/marvell/mv643xx_eth.c
2469
port_start(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2471
wrlp(mp, INT_MASK_EXT, INT_EXT_LINK_PHY | INT_EXT_TX);
drivers/net/ethernet/marvell/mv643xx_eth.c
2472
wrlp(mp, INT_MASK, mp->int_mask);
drivers/net/ethernet/marvell/mv643xx_eth.c
2478
for (i = 0; i < mp->rxq_count; i++)
drivers/net/ethernet/marvell/mv643xx_eth.c
2479
rxq_deinit(mp->rxq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2481
napi_disable(&mp->napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2487
static void port_reset(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2492
for (i = 0; i < mp->rxq_count; i++)
drivers/net/ethernet/marvell/mv643xx_eth.c
2493
rxq_disable(mp->rxq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2494
for (i = 0; i < mp->txq_count; i++)
drivers/net/ethernet/marvell/mv643xx_eth.c
2495
txq_disable(mp->txq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2498
u32 ps = rdlp(mp, PORT_STATUS);
drivers/net/ethernet/marvell/mv643xx_eth.c
2506
data = rdlp(mp, PORT_SERIAL_CONTROL);
drivers/net/ethernet/marvell/mv643xx_eth.c
2510
wrlp(mp, PORT_SERIAL_CONTROL, data);
drivers/net/ethernet/marvell/mv643xx_eth.c
2515
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2518
wrlp(mp, INT_MASK_EXT, 0x00000000);
drivers/net/ethernet/marvell/mv643xx_eth.c
2519
wrlp(mp, INT_MASK, 0x00000000);
drivers/net/ethernet/marvell/mv643xx_eth.c
2520
rdlp(mp, INT_MASK);
drivers/net/ethernet/marvell/mv643xx_eth.c
2522
napi_disable(&mp->napi);
drivers/net/ethernet/marvell/mv643xx_eth.c
2524
timer_delete_sync(&mp->rx_oom);
drivers/net/ethernet/marvell/mv643xx_eth.c
2531
port_reset(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2533
mib_counters_update(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2534
timer_delete_sync(&mp->mib_counters_timer);
drivers/net/ethernet/marvell/mv643xx_eth.c
2536
for (i = 0; i < mp->rxq_count; i++)
drivers/net/ethernet/marvell/mv643xx_eth.c
2537
rxq_deinit(mp->rxq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2538
for (i = 0; i < mp->txq_count; i++)
drivers/net/ethernet/marvell/mv643xx_eth.c
2539
txq_deinit(mp->txq + i);
drivers/net/ethernet/marvell/mv643xx_eth.c
2559
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2562
mv643xx_eth_recalc_skb_size(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2563
tx_set_rate(mp, 1000000000, 16777216);
drivers/net/ethernet/marvell/mv643xx_eth.c
2585
struct mv643xx_eth_private *mp;
drivers/net/ethernet/marvell/mv643xx_eth.c
2587
mp = container_of(ugly, struct mv643xx_eth_private, tx_timeout_task);
drivers/net/ethernet/marvell/mv643xx_eth.c
2588
if (netif_running(mp->dev)) {
drivers/net/ethernet/marvell/mv643xx_eth.c
2589
netif_tx_stop_all_queues(mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2590
port_reset(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2591
port_start(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
2592
netif_tx_wake_all_queues(mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2598
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2602
schedule_work(&mp->tx_timeout_task);
drivers/net/ethernet/marvell/mv643xx_eth.c
2608
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
2610
wrlp(mp, INT_MASK, 0x00000000);
drivers/net/ethernet/marvell/mv643xx_eth.c
2611
rdlp(mp, INT_MASK);
drivers/net/ethernet/marvell/mv643xx_eth.c
2615
wrlp(mp, INT_MASK, mp->int_mask);
drivers/net/ethernet/marvell/mv643xx_eth.c
2904
static void phy_addr_set(struct mv643xx_eth_private *mp, int phy_addr)
drivers/net/ethernet/marvell/mv643xx_eth.c
2906
int addr_shift = 5 * mp->port_num;
drivers/net/ethernet/marvell/mv643xx_eth.c
2909
data = rdl(mp, PHY_ADDR);
drivers/net/ethernet/marvell/mv643xx_eth.c
2912
wrl(mp, PHY_ADDR, data);
drivers/net/ethernet/marvell/mv643xx_eth.c
2915
static int phy_addr_get(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2919
data = rdl(mp, PHY_ADDR);
drivers/net/ethernet/marvell/mv643xx_eth.c
2921
return (data >> (5 * mp->port_num)) & 0x1f;
drivers/net/ethernet/marvell/mv643xx_eth.c
2924
static void set_params(struct mv643xx_eth_private *mp,
drivers/net/ethernet/marvell/mv643xx_eth.c
2927
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
2935
uc_addr_get(mp, addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
2939
mp->rx_ring_size = DEFAULT_RX_QUEUE_SIZE;
drivers/net/ethernet/marvell/mv643xx_eth.c
2941
mp->rx_ring_size = pd->rx_queue_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
2942
mp->rx_desc_sram_addr = pd->rx_sram_addr;
drivers/net/ethernet/marvell/mv643xx_eth.c
2943
mp->rx_desc_sram_size = pd->rx_sram_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
2945
mp->rxq_count = pd->rx_queue_count ? : 1;
drivers/net/ethernet/marvell/mv643xx_eth.c
2951
mp->tx_ring_size = clamp_t(unsigned int, tx_ring_size,
drivers/net/ethernet/marvell/mv643xx_eth.c
2953
if (mp->tx_ring_size != tx_ring_size)
drivers/net/ethernet/marvell/mv643xx_eth.c
2955
mp->tx_ring_size, tx_ring_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
2957
mp->tx_desc_sram_addr = pd->tx_sram_addr;
drivers/net/ethernet/marvell/mv643xx_eth.c
2958
mp->tx_desc_sram_size = pd->tx_sram_size;
drivers/net/ethernet/marvell/mv643xx_eth.c
2960
mp->txq_count = pd->tx_queue_count ? : 1;
drivers/net/ethernet/marvell/mv643xx_eth.c
2963
static int get_phy_mode(struct mv643xx_eth_private *mp)
drivers/net/ethernet/marvell/mv643xx_eth.c
2965
struct device *dev = mp->dev->dev.parent;
drivers/net/ethernet/marvell/mv643xx_eth.c
2980
static struct phy_device *phy_scan(struct mv643xx_eth_private *mp,
drivers/net/ethernet/marvell/mv643xx_eth.c
2990
start = phy_addr_get(mp) & 0x1f;
drivers/net/ethernet/marvell/mv643xx_eth.c
3005
phydev = phy_connect(mp->dev, phy_id, mv643xx_eth_adjust_link,
drivers/net/ethernet/marvell/mv643xx_eth.c
3006
get_phy_mode(mp));
drivers/net/ethernet/marvell/mv643xx_eth.c
3008
phy_addr_set(mp, addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3016
static void phy_init(struct mv643xx_eth_private *mp, int speed, int duplex)
drivers/net/ethernet/marvell/mv643xx_eth.c
3018
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
3037
static void init_pscr(struct mv643xx_eth_private *mp, int speed, int duplex)
drivers/net/ethernet/marvell/mv643xx_eth.c
3039
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
3042
pscr = rdlp(mp, PORT_SERIAL_CONTROL);
drivers/net/ethernet/marvell/mv643xx_eth.c
3045
wrlp(mp, PORT_SERIAL_CONTROL, pscr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3063
wrlp(mp, PORT_SERIAL_CONTROL, pscr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3086
struct mv643xx_eth_private *mp;
drivers/net/ethernet/marvell/mv643xx_eth.c
3108
mp = netdev_priv(dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
3109
platform_set_drvdata(pdev, mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
3111
mp->shared = platform_get_drvdata(pd->shared);
drivers/net/ethernet/marvell/mv643xx_eth.c
3112
mp->base = mp->shared->base + 0x0400 + (pd->port_number << 10);
drivers/net/ethernet/marvell/mv643xx_eth.c
3113
mp->port_num = pd->port_number;
drivers/net/ethernet/marvell/mv643xx_eth.c
3115
mp->dev = dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
3119
psc1r = rdlp(mp, PORT_SERIAL_CONTROL1);
drivers/net/ethernet/marvell/mv643xx_eth.c
3154
wrlp(mp, PORT_SERIAL_CONTROL1, psc1r);
drivers/net/ethernet/marvell/mv643xx_eth.c
3161
mp->t_clk = 133000000;
drivers/net/ethernet/marvell/mv643xx_eth.c
3162
mp->clk = devm_clk_get(&pdev->dev, NULL);
drivers/net/ethernet/marvell/mv643xx_eth.c
3163
if (!IS_ERR(mp->clk)) {
drivers/net/ethernet/marvell/mv643xx_eth.c
3164
clk_prepare_enable(mp->clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
3165
mp->t_clk = clk_get_rate(mp->clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
3166
} else if (!IS_ERR(mp->shared->clk)) {
drivers/net/ethernet/marvell/mv643xx_eth.c
3167
mp->t_clk = clk_get_rate(mp->shared->clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
3170
set_params(mp, pd);
drivers/net/ethernet/marvell/mv643xx_eth.c
3171
netif_set_real_num_tx_queues(dev, mp->txq_count);
drivers/net/ethernet/marvell/mv643xx_eth.c
3172
netif_set_real_num_rx_queues(dev, mp->rxq_count);
drivers/net/ethernet/marvell/mv643xx_eth.c
3176
phydev = of_phy_connect(mp->dev, pd->phy_node,
drivers/net/ethernet/marvell/mv643xx_eth.c
3178
get_phy_mode(mp));
drivers/net/ethernet/marvell/mv643xx_eth.c
3182
phy_addr_set(mp, phydev->mdio.addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3184
phydev = phy_scan(mp, pd->phy_addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3189
phy_init(mp, pd->speed, pd->duplex);
drivers/net/ethernet/marvell/mv643xx_eth.c
3200
init_pscr(mp, pd->speed, pd->duplex);
drivers/net/ethernet/marvell/mv643xx_eth.c
3203
mib_counters_clear(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
3205
timer_setup(&mp->mib_counters_timer, mib_counters_timer_wrapper, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
3206
mp->mib_counters_timer.expires = jiffies + 30 * HZ;
drivers/net/ethernet/marvell/mv643xx_eth.c
3208
spin_lock_init(&mp->mib_counters_lock);
drivers/net/ethernet/marvell/mv643xx_eth.c
3210
INIT_WORK(&mp->tx_timeout_task, tx_timeout_task);
drivers/net/ethernet/marvell/mv643xx_eth.c
3212
netif_napi_add(dev, &mp->napi, mv643xx_eth_poll);
drivers/net/ethernet/marvell/mv643xx_eth.c
3214
timer_setup(&mp->rx_oom, oom_timer_wrapper, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
3242
if (mp->shared->win_protect)
drivers/net/ethernet/marvell/mv643xx_eth.c
3243
wrl(mp, WINDOW_PROTECT(mp->port_num), mp->shared->win_protect);
drivers/net/ethernet/marvell/mv643xx_eth.c
3247
wrlp(mp, SDMA_CONFIG, PORT_SDMA_CONFIG_DEFAULT_VALUE);
drivers/net/ethernet/marvell/mv643xx_eth.c
3249
set_rx_coal(mp, 250);
drivers/net/ethernet/marvell/mv643xx_eth.c
3250
set_tx_coal(mp, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
3257
mp->port_num, dev->dev_addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
3259
if (mp->tx_desc_sram_size > 0)
drivers/net/ethernet/marvell/mv643xx_eth.c
3265
if (!IS_ERR(mp->clk))
drivers/net/ethernet/marvell/mv643xx_eth.c
3266
clk_disable_unprepare(mp->clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
3274
struct mv643xx_eth_private *mp = platform_get_drvdata(pdev);
drivers/net/ethernet/marvell/mv643xx_eth.c
3275
struct net_device *dev = mp->dev;
drivers/net/ethernet/marvell/mv643xx_eth.c
3277
unregister_netdev(mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
3280
cancel_work_sync(&mp->tx_timeout_task);
drivers/net/ethernet/marvell/mv643xx_eth.c
3282
if (!IS_ERR(mp->clk))
drivers/net/ethernet/marvell/mv643xx_eth.c
3283
clk_disable_unprepare(mp->clk);
drivers/net/ethernet/marvell/mv643xx_eth.c
3285
free_netdev(mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
3290
struct mv643xx_eth_private *mp = platform_get_drvdata(pdev);
drivers/net/ethernet/marvell/mv643xx_eth.c
3293
wrlp(mp, INT_MASK, 0);
drivers/net/ethernet/marvell/mv643xx_eth.c
3294
rdlp(mp, INT_MASK);
drivers/net/ethernet/marvell/mv643xx_eth.c
3296
if (netif_running(mp->dev))
drivers/net/ethernet/marvell/mv643xx_eth.c
3297
port_reset(mp);
drivers/net/ethernet/marvell/mv643xx_eth.c
419
static inline u32 rdl(struct mv643xx_eth_private *mp, int offset)
drivers/net/ethernet/marvell/mv643xx_eth.c
421
return readl(mp->shared->base + offset);
drivers/net/ethernet/marvell/mv643xx_eth.c
424
static inline u32 rdlp(struct mv643xx_eth_private *mp, int offset)
drivers/net/ethernet/marvell/mv643xx_eth.c
426
return readl(mp->base + offset);
drivers/net/ethernet/marvell/mv643xx_eth.c
429
static inline void wrl(struct mv643xx_eth_private *mp, int offset, u32 data)
drivers/net/ethernet/marvell/mv643xx_eth.c
431
writel(data, mp->shared->base + offset);
drivers/net/ethernet/marvell/mv643xx_eth.c
434
static inline void wrlp(struct mv643xx_eth_private *mp, int offset, u32 data)
drivers/net/ethernet/marvell/mv643xx_eth.c
436
writel(data, mp->base + offset);
drivers/net/ethernet/marvell/mv643xx_eth.c
453
struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
drivers/net/ethernet/marvell/mv643xx_eth.c
454
wrlp(mp, RXQ_COMMAND, 1 << rxq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
459
struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
drivers/net/ethernet/marvell/mv643xx_eth.c
462
wrlp(mp, RXQ_COMMAND, mask << 8);
drivers/net/ethernet/marvell/mv643xx_eth.c
463
while (rdlp(mp, RXQ_COMMAND) & mask)
drivers/net/ethernet/marvell/mv643xx_eth.c
469
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
474
wrlp(mp, TXQ_CURRENT_DESC_PTR(txq->index), addr);
drivers/net/ethernet/marvell/mv643xx_eth.c
479
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
480
wrlp(mp, TXQ_COMMAND, 1 << txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
485
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
488
wrlp(mp, TXQ_COMMAND, mask << 8);
drivers/net/ethernet/marvell/mv643xx_eth.c
489
while (rdlp(mp, TXQ_COMMAND) & mask)
drivers/net/ethernet/marvell/mv643xx_eth.c
495
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
496
struct netdev_queue *nq = netdev_get_tx_queue(mp->dev, txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
508
struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
drivers/net/ethernet/marvell/mv643xx_eth.c
509
struct net_device_stats *stats = &mp->dev->stats;
drivers/net/ethernet/marvell/mv643xx_eth.c
533
dma_unmap_single(mp->dev->dev.parent, rx_desc->buf_ptr,
drivers/net/ethernet/marvell/mv643xx_eth.c
538
mp->work_rx_refill |= 1 << rxq->index;
drivers/net/ethernet/marvell/mv643xx_eth.c
570
skb->protocol = eth_type_trans(skb, mp->dev);
drivers/net/ethernet/marvell/mv643xx_eth.c
572
napi_gro_receive(&mp->napi, skb);
drivers/net/ethernet/marvell/mv643xx_eth.c
582
netdev_err(mp->dev,
drivers/net/ethernet/marvell/mv643xx_eth.c
593
mp->work_rx &= ~(1 << rxq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
600
struct mv643xx_eth_private *mp = rxq_to_mp(rxq);
drivers/net/ethernet/marvell/mv643xx_eth.c
610
skb = netdev_alloc_skb(mp->dev, mp->skb_size);
drivers/net/ethernet/marvell/mv643xx_eth.c
613
mp->oom = 1;
drivers/net/ethernet/marvell/mv643xx_eth.c
630
rx_desc->buf_ptr = dma_map_single(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
648
mp->work_rx_refill &= ~(1 << rxq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
670
static int skb_tx_csum(struct mv643xx_eth_private *mp, struct sk_buff *skb,
drivers/net/ethernet/marvell/mv643xx_eth.c
686
if (length - hdr_len > mp->shared->tx_csum_limit ||
drivers/net/ethernet/marvell/mv643xx_eth.c
778
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
790
ret = skb_tx_csum(mp, skb, &l4i_chk, &cmd_csum, length);
drivers/net/ethernet/marvell/mv643xx_eth.c
821
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
877
mp->work_tx_end &= ~(1 << txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
893
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
923
desc->buf_ptr = skb_frag_dma_map(mp->dev->dev.parent,
drivers/net/ethernet/marvell/mv643xx_eth.c
932
struct mv643xx_eth_private *mp = txq_to_mp(txq);
drivers/net/ethernet/marvell/mv643xx_eth.c
949
ret = skb_tx_csum(mp, skb, &l4i_chk, &cmd_sts, skb->len);
drivers/net/ethernet/marvell/mv643xx_eth.c
970
desc->buf_ptr = dma_map_single(mp->dev->dev.parent, skb->data,
drivers/net/ethernet/marvell/mv643xx_eth.c
982
mp->work_tx_end &= ~(1 << txq->index);
drivers/net/ethernet/marvell/mv643xx_eth.c
995
struct mv643xx_eth_private *mp = netdev_priv(dev);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
103
struct lag_mp *mp = container_of(nb, struct lag_mp, fib_nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
105
flush_workqueue(mp->wq);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
108
static void mlx5_lag_fib_set(struct lag_mp *mp, struct fib_info *fi, u32 dst, int dst_len)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
110
mp->fib.mfi = fi;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
111
mp->fib.priority = fi->fib_priority;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
112
mp->fib.dst = dst;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
113
mp->fib.dst_len = dst_len;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
161
struct lag_mp *mp = &ldev->lag_mp;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
170
if (mp->fib.mfi == fi)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
171
mp->fib.mfi = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
176
if (mp->fib.mfi &&
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
177
(mp->fib.dst != fen_info->dst || mp->fib.dst_len != fen_info->dst_len) &&
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
178
mp->fib.dst_len <= fen_info->dst_len &&
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
179
!(mp->fib.dst_len == fen_info->dst_len &&
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
180
fi->fib_priority < mp->fib.priority)) {
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
191
if (mp->fib.dst == fen_info->dst && mp->fib.dst_len == fen_info->dst_len)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
192
mp->fib.mfi = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
210
mlx5_lag_fib_set(mp, fi, fen_info->dst, fen_info->dst_len);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
217
if (!mp->fib.mfi && !__mlx5_lag_is_active(ldev)) {
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
225
mlx5_lag_fib_set(mp, fi, fen_info->dst, fen_info->dst_len);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
233
struct lag_mp *mp = &ldev->lag_mp;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
237
if (!mp->fib.mfi || mp->fib.mfi != fi)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
309
struct lag_mp *mp = container_of(nb, struct lag_mp, fib_nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
310
struct mlx5_lag *ldev = container_of(mp, struct mlx5_lag, lag_mp);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
355
queue_work(mp->wq, &fib_work->work);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
370
struct lag_mp *mp = &ldev->lag_mp;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
376
mp->fib.mfi = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
378
if (mp->fib_nb.notifier_call)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
381
mp->wq = create_singlethread_workqueue("mlx5_lag_mp");
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
382
if (!mp->wq)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
385
mp->fib_nb.notifier_call = mlx5_lag_fib_event;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
386
err = register_fib_notifier(&init_net, &mp->fib_nb,
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
389
destroy_workqueue(mp->wq);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
390
mp->fib_nb.notifier_call = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
398
struct lag_mp *mp = &ldev->lag_mp;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
400
if (!mp->fib_nb.notifier_call)
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
403
unregister_fib_notifier(&init_net, &mp->fib_nb);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
404
destroy_workqueue(mp->wq);
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
405
mp->fib_nb.notifier_call = NULL;
drivers/net/ethernet/mellanox/mlx5/core/lag/mp.c
406
mp->fib.mfi = NULL;
drivers/net/ethernet/sun/ldmvsw.c
292
rmac = mdesc_get_property(hp, vdev->mp, remote_macaddr_prop, &len);
drivers/net/ethernet/sun/ldmvsw.c
300
port_id = mdesc_get_property(hp, vdev->mp, id_prop, NULL);
drivers/net/ethernet/sun/ldmvsw.c
309
vp = vsw_get_vnet(hp, vdev->mp, &handle);
drivers/net/ethernet/sun/niu.c
3921
struct niu_xmac_stats *mp = &np->mac_stats.xmac;
drivers/net/ethernet/sun/niu.c
3926
mp->tx_frames += TXMAC_FRM_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3928
mp->tx_bytes += TXMAC_BYTE_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3930
mp->tx_fifo_errors++;
drivers/net/ethernet/sun/niu.c
3932
mp->tx_overflow_errors++;
drivers/net/ethernet/sun/niu.c
3934
mp->tx_max_pkt_size_errors++;
drivers/net/ethernet/sun/niu.c
3936
mp->tx_underflow_errors++;
drivers/net/ethernet/sun/niu.c
3940
mp->rx_local_faults++;
drivers/net/ethernet/sun/niu.c
3942
mp->rx_remote_faults++;
drivers/net/ethernet/sun/niu.c
3944
mp->rx_link_faults += LINK_FAULT_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3946
mp->rx_align_errors += RXMAC_ALIGN_ERR_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3948
mp->rx_frags += RXMAC_FRAG_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3950
mp->rx_mcasts += RXMAC_MC_FRM_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3952
mp->rx_bcasts += RXMAC_BC_FRM_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3954
mp->rx_hist_cnt1 += RXMAC_HIST_CNT1_COUNT;
drivers/net/ethernet/sun/niu.c
3956
mp->rx_hist_cnt2 += RXMAC_HIST_CNT2_COUNT;
drivers/net/ethernet/sun/niu.c
3958
mp->rx_hist_cnt3 += RXMAC_HIST_CNT3_COUNT;
drivers/net/ethernet/sun/niu.c
3960
mp->rx_hist_cnt4 += RXMAC_HIST_CNT4_COUNT;
drivers/net/ethernet/sun/niu.c
3962
mp->rx_hist_cnt5 += RXMAC_HIST_CNT5_COUNT;
drivers/net/ethernet/sun/niu.c
3964
mp->rx_hist_cnt6 += RXMAC_HIST_CNT6_COUNT;
drivers/net/ethernet/sun/niu.c
3966
mp->rx_hist_cnt7 += RXMAC_HIST_CNT7_COUNT;
drivers/net/ethernet/sun/niu.c
3968
mp->rx_octets += RXMAC_BT_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3970
mp->rx_code_violations += RXMAC_CD_VIO_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3972
mp->rx_len_errors += RXMAC_MPSZER_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3974
mp->rx_crc_errors += RXMAC_CRC_ER_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
3976
mp->rx_underflows++;
drivers/net/ethernet/sun/niu.c
3978
mp->rx_overflows++;
drivers/net/ethernet/sun/niu.c
3982
mp->pause_off_state++;
drivers/net/ethernet/sun/niu.c
3984
mp->pause_on_state++;
drivers/net/ethernet/sun/niu.c
3986
mp->pause_received++;
drivers/net/ethernet/sun/niu.c
3991
struct niu_bmac_stats *mp = &np->mac_stats.bmac;
drivers/net/ethernet/sun/niu.c
3996
mp->tx_underflow_errors++;
drivers/net/ethernet/sun/niu.c
3998
mp->tx_max_pkt_size_errors++;
drivers/net/ethernet/sun/niu.c
4000
mp->tx_bytes += BTXMAC_BYTE_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4002
mp->tx_frames += BTXMAC_FRM_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4006
mp->rx_overflows++;
drivers/net/ethernet/sun/niu.c
4008
mp->rx_frames += BRXMAC_FRAME_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4010
mp->rx_align_errors += BRXMAC_ALIGN_ERR_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4012
mp->rx_crc_errors += BRXMAC_ALIGN_ERR_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4014
mp->rx_len_errors += BRXMAC_CODE_VIOL_ERR_CNT_COUNT;
drivers/net/ethernet/sun/niu.c
4018
mp->pause_off_state++;
drivers/net/ethernet/sun/niu.c
4020
mp->pause_on_state++;
drivers/net/ethernet/sun/niu.c
4022
mp->pause_received++;
drivers/net/ethernet/sun/niu.c
6191
struct niu_xmac_stats *mp = &np->mac_stats.xmac;
drivers/net/ethernet/sun/niu.c
6193
mp->tx_frames += nr64_mac(TXMAC_FRM_CNT);
drivers/net/ethernet/sun/niu.c
6194
mp->tx_bytes += nr64_mac(TXMAC_BYTE_CNT);
drivers/net/ethernet/sun/niu.c
6196
mp->rx_link_faults += nr64_mac(LINK_FAULT_CNT);
drivers/net/ethernet/sun/niu.c
6197
mp->rx_align_errors += nr64_mac(RXMAC_ALIGN_ERR_CNT);
drivers/net/ethernet/sun/niu.c
6198
mp->rx_frags += nr64_mac(RXMAC_FRAG_CNT);
drivers/net/ethernet/sun/niu.c
6199
mp->rx_mcasts += nr64_mac(RXMAC_MC_FRM_CNT);
drivers/net/ethernet/sun/niu.c
6200
mp->rx_bcasts += nr64_mac(RXMAC_BC_FRM_CNT);
drivers/net/ethernet/sun/niu.c
6201
mp->rx_hist_cnt1 += nr64_mac(RXMAC_HIST_CNT1);
drivers/net/ethernet/sun/niu.c
6202
mp->rx_hist_cnt2 += nr64_mac(RXMAC_HIST_CNT2);
drivers/net/ethernet/sun/niu.c
6203
mp->rx_hist_cnt3 += nr64_mac(RXMAC_HIST_CNT3);
drivers/net/ethernet/sun/niu.c
6204
mp->rx_hist_cnt4 += nr64_mac(RXMAC_HIST_CNT4);
drivers/net/ethernet/sun/niu.c
6205
mp->rx_hist_cnt5 += nr64_mac(RXMAC_HIST_CNT5);
drivers/net/ethernet/sun/niu.c
6206
mp->rx_hist_cnt6 += nr64_mac(RXMAC_HIST_CNT6);
drivers/net/ethernet/sun/niu.c
6207
mp->rx_hist_cnt7 += nr64_mac(RXMAC_HIST_CNT7);
drivers/net/ethernet/sun/niu.c
6208
mp->rx_octets += nr64_mac(RXMAC_BT_CNT);
drivers/net/ethernet/sun/niu.c
6209
mp->rx_code_violations += nr64_mac(RXMAC_CD_VIO_CNT);
drivers/net/ethernet/sun/niu.c
6210
mp->rx_len_errors += nr64_mac(RXMAC_MPSZER_CNT);
drivers/net/ethernet/sun/niu.c
6211
mp->rx_crc_errors += nr64_mac(RXMAC_CRC_ER_CNT);
drivers/net/ethernet/sun/niu.c
6216
struct niu_bmac_stats *mp = &np->mac_stats.bmac;
drivers/net/ethernet/sun/niu.c
6218
mp->tx_bytes += nr64_mac(BTXMAC_BYTE_CNT);
drivers/net/ethernet/sun/niu.c
6219
mp->tx_frames += nr64_mac(BTXMAC_FRM_CNT);
drivers/net/ethernet/sun/niu.c
6221
mp->rx_frames += nr64_mac(BRXMAC_FRAME_CNT);
drivers/net/ethernet/sun/niu.c
6222
mp->rx_align_errors += nr64_mac(BRXMAC_ALIGN_ERR_CNT);
drivers/net/ethernet/sun/niu.c
6223
mp->rx_crc_errors += nr64_mac(BRXMAC_ALIGN_ERR_CNT);
drivers/net/ethernet/sun/niu.c
6224
mp->rx_len_errors += nr64_mac(BRXMAC_CODE_VIOL_ERR_CNT);
drivers/net/ethernet/sun/sunvnet.c
425
vp = vnet_find_parent(hp, vdev->mp, vdev);
drivers/net/ethernet/sun/sunvnet.c
432
rmac = mdesc_get_property(hp, vdev->mp, remote_macaddr_prop, &len);
drivers/net/ethernet/sun/sunvnet.c
465
if (mdesc_get_property(hp, vdev->mp, "switch-port", NULL))
drivers/net/ethernet/ti/icssg/icssg_config.c
198
const struct map *mp;
drivers/net/ethernet/ti/icssg/icssg_config.c
202
mp = &hwq_map[slice][j];
drivers/net/ethernet/ti/icssg/icssg_config.c
203
if (mp->special) {
drivers/net/ethernet/ti/icssg/icssg_config.c
215
pdword[0] |= mp->flags;
drivers/net/ethernet/ti/icssg/icssg_config.c
216
pdaddr = mp->pd_addr_start + i * pd_size;
drivers/net/ethernet/ti/icssg/icssg_config.c
219
queue = mp->queue;
drivers/net/hamradio/baycom_epp.c
1037
hi.data.mp.iobase = dev->base_addr;
drivers/net/hamradio/baycom_epp.c
1038
hi.data.mp.irq = dev->irq;
drivers/net/hamradio/baycom_epp.c
1039
hi.data.mp.dma = dev->dma;
drivers/net/hamradio/baycom_epp.c
1040
hi.data.mp.dma2 = 0;
drivers/net/hamradio/baycom_epp.c
1041
hi.data.mp.seriobase = 0;
drivers/net/hamradio/baycom_epp.c
1042
hi.data.mp.pariobase = 0;
drivers/net/hamradio/baycom_epp.c
1043
hi.data.mp.midiiobase = 0;
drivers/net/hamradio/baycom_epp.c
1049
dev->base_addr = hi.data.mp.iobase;
drivers/net/hamradio/hdlcdrv.c
528
bi.data.mp.iobase = dev->base_addr;
drivers/net/hamradio/hdlcdrv.c
529
bi.data.mp.irq = dev->irq;
drivers/net/hamradio/hdlcdrv.c
530
bi.data.mp.dma = dev->dma;
drivers/net/hamradio/hdlcdrv.c
531
bi.data.mp.dma2 = s->ptt_out.dma2;
drivers/net/hamradio/hdlcdrv.c
532
bi.data.mp.seriobase = s->ptt_out.seriobase;
drivers/net/hamradio/hdlcdrv.c
533
bi.data.mp.pariobase = s->ptt_out.pariobase;
drivers/net/hamradio/hdlcdrv.c
534
bi.data.mp.midiiobase = s->ptt_out.midiiobase;
drivers/net/hamradio/hdlcdrv.c
540
dev->base_addr = bi.data.mp.iobase;
drivers/net/hamradio/hdlcdrv.c
541
dev->irq = bi.data.mp.irq;
drivers/net/hamradio/hdlcdrv.c
542
dev->dma = bi.data.mp.dma;
drivers/net/hamradio/hdlcdrv.c
543
s->ptt_out.dma2 = bi.data.mp.dma2;
drivers/net/hamradio/hdlcdrv.c
544
s->ptt_out.seriobase = bi.data.mp.seriobase;
drivers/net/hamradio/hdlcdrv.c
545
s->ptt_out.pariobase = bi.data.mp.pariobase;
drivers/net/hamradio/hdlcdrv.c
546
s->ptt_out.midiiobase = bi.data.mp.midiiobase;
drivers/net/wireless/realtek/rtw89/fw.c
643
if (tmp->cv <= rtwdev->hal.cv && !tmp->mp) {
drivers/net/wireless/realtek/rtw89/fw.h
4206
u8 mp;
drivers/pci/controller/dwc/pci-meson.c
107
struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
109
struct dw_pcie *pci = &mp->pci;
drivers/pci/controller/dwc/pci-meson.c
127
mp->cfg_base = devm_platform_ioremap_resource_byname(pdev, "cfg");
drivers/pci/controller/dwc/pci-meson.c
128
if (IS_ERR(mp->cfg_base))
drivers/pci/controller/dwc/pci-meson.c
129
return PTR_ERR(mp->cfg_base);
drivers/pci/controller/dwc/pci-meson.c
134
static int meson_pcie_power_on(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
138
ret = phy_init(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
142
ret = phy_power_on(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
144
phy_exit(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
151
static void meson_pcie_power_off(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
153
phy_power_off(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
154
phy_exit(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
157
static int meson_pcie_reset(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
159
struct meson_pcie_rc_reset *mrst = &mp->mrst;
drivers/pci/controller/dwc/pci-meson.c
162
ret = phy_reset(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
212
static int meson_pcie_probe_clocks(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
214
struct device *dev = mp->pci.dev;
drivers/pci/controller/dwc/pci-meson.c
215
struct meson_pcie_clk_res *res = &mp->clk_res;
drivers/pci/controller/dwc/pci-meson.c
232
static inline u32 meson_cfg_readl(struct meson_pcie *mp, u32 reg)
drivers/pci/controller/dwc/pci-meson.c
234
return readl(mp->cfg_base + reg);
drivers/pci/controller/dwc/pci-meson.c
237
static inline void meson_cfg_writel(struct meson_pcie *mp, u32 val, u32 reg)
drivers/pci/controller/dwc/pci-meson.c
239
writel(val, mp->cfg_base + reg);
drivers/pci/controller/dwc/pci-meson.c
242
static void meson_pcie_assert_reset(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
244
gpiod_set_value_cansleep(mp->reset_gpio, 1);
drivers/pci/controller/dwc/pci-meson.c
246
gpiod_set_value_cansleep(mp->reset_gpio, 0);
drivers/pci/controller/dwc/pci-meson.c
249
static void meson_pcie_ltssm_enable(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
253
val = meson_cfg_readl(mp, PCIE_CFG0);
drivers/pci/controller/dwc/pci-meson.c
255
meson_cfg_writel(mp, val, PCIE_CFG0);
drivers/pci/controller/dwc/pci-meson.c
258
static int meson_size_to_payload(struct meson_pcie *mp, int size)
drivers/pci/controller/dwc/pci-meson.c
260
struct device *dev = mp->pci.dev;
drivers/pci/controller/dwc/pci-meson.c
275
static void meson_set_max_payload(struct meson_pcie *mp, int size)
drivers/pci/controller/dwc/pci-meson.c
277
struct dw_pcie *pci = &mp->pci;
drivers/pci/controller/dwc/pci-meson.c
280
int max_payload_size = meson_size_to_payload(mp, size);
drivers/pci/controller/dwc/pci-meson.c
291
static void meson_set_max_rd_req_size(struct meson_pcie *mp, int size)
drivers/pci/controller/dwc/pci-meson.c
293
struct dw_pcie *pci = &mp->pci;
drivers/pci/controller/dwc/pci-meson.c
296
int max_rd_req_size = meson_size_to_payload(mp, size);
drivers/pci/controller/dwc/pci-meson.c
309
struct meson_pcie *mp = to_meson_pcie(pci);
drivers/pci/controller/dwc/pci-meson.c
311
meson_pcie_ltssm_enable(mp);
drivers/pci/controller/dwc/pci-meson.c
312
meson_pcie_assert_reset(mp);
drivers/pci/controller/dwc/pci-meson.c
351
struct meson_pcie *mp = to_meson_pcie(pci);
drivers/pci/controller/dwc/pci-meson.c
354
state12 = meson_cfg_readl(mp, PCIE_CFG_STATUS12);
drivers/pci/controller/dwc/pci-meson.c
361
struct meson_pcie *mp = to_meson_pcie(pci);
drivers/pci/controller/dwc/pci-meson.c
365
meson_set_max_payload(mp, MAX_PAYLOAD_SIZE);
drivers/pci/controller/dwc/pci-meson.c
366
meson_set_max_rd_req_size(mp, MAX_READ_REQ_SIZE);
drivers/pci/controller/dwc/pci-meson.c
384
struct meson_pcie *mp;
drivers/pci/controller/dwc/pci-meson.c
387
mp = devm_kzalloc(dev, sizeof(*mp), GFP_KERNEL);
drivers/pci/controller/dwc/pci-meson.c
388
if (!mp)
drivers/pci/controller/dwc/pci-meson.c
391
pci = &mp->pci;
drivers/pci/controller/dwc/pci-meson.c
397
mp->phy = devm_phy_get(dev, "pcie");
drivers/pci/controller/dwc/pci-meson.c
398
if (IS_ERR(mp->phy)) {
drivers/pci/controller/dwc/pci-meson.c
399
dev_err(dev, "get phy failed, %ld\n", PTR_ERR(mp->phy));
drivers/pci/controller/dwc/pci-meson.c
400
return PTR_ERR(mp->phy);
drivers/pci/controller/dwc/pci-meson.c
403
mp->reset_gpio = devm_gpiod_get(dev, "reset", GPIOD_OUT_LOW);
drivers/pci/controller/dwc/pci-meson.c
404
if (IS_ERR(mp->reset_gpio)) {
drivers/pci/controller/dwc/pci-meson.c
406
return PTR_ERR(mp->reset_gpio);
drivers/pci/controller/dwc/pci-meson.c
409
ret = meson_pcie_get_resets(mp);
drivers/pci/controller/dwc/pci-meson.c
415
ret = meson_pcie_get_mems(pdev, mp);
drivers/pci/controller/dwc/pci-meson.c
421
ret = meson_pcie_power_on(mp);
drivers/pci/controller/dwc/pci-meson.c
427
ret = meson_pcie_reset(mp);
drivers/pci/controller/dwc/pci-meson.c
433
ret = meson_pcie_probe_clocks(mp);
drivers/pci/controller/dwc/pci-meson.c
439
platform_set_drvdata(pdev, mp);
drivers/pci/controller/dwc/pci-meson.c
450
meson_pcie_power_off(mp);
drivers/pci/controller/dwc/pci-meson.c
74
static struct reset_control *meson_pcie_get_reset(struct meson_pcie *mp,
drivers/pci/controller/dwc/pci-meson.c
78
struct device *dev = mp->pci.dev;
drivers/pci/controller/dwc/pci-meson.c
89
static int meson_pcie_get_resets(struct meson_pcie *mp)
drivers/pci/controller/dwc/pci-meson.c
91
struct meson_pcie_rc_reset *mrst = &mp->mrst;
drivers/pci/controller/dwc/pci-meson.c
93
mrst->port = meson_pcie_get_reset(mp, "port", PCIE_NORMAL_RESET);
drivers/pci/controller/dwc/pci-meson.c
98
mrst->apb = meson_pcie_get_reset(mp, "apb", PCIE_SHARED_RESET);
drivers/platform/x86/intel/pmc/core.c
1002
bool required = lpm_req_regs[mp + (mode * num_maps)] &
drivers/platform/x86/intel/pmc/core.c
934
unsigned int mp, pmc_idx;
drivers/platform/x86/intel/pmc/core.c
962
for (mp = 0; mp < num_maps; mp++) {
drivers/platform/x86/intel/pmc/core.c
976
req_mask |= lpm_req_regs[mp + (mode * num_maps)];
drivers/platform/x86/intel/pmc/core.c
979
lpm_status = pmc_core_reg_read(pmc, sts_offset + (mp * 4));
drivers/platform/x86/intel/pmc/core.c
982
lpm_status_live = pmc_core_reg_read(pmc, sts_offset_live + (mp * 4));
drivers/platform/x86/intel/pmc/core.c
985
map = maps[mp];
drivers/s390/cio/cio.c
386
schib->pmcw.mp = sch->config.mp;
drivers/s390/cio/cio.c
400
(schib->pmcw.mp == sch->config.mp) &&
drivers/s390/cio/cio.h
27
u32 mp : 1; /* multipath mode */
drivers/s390/cio/cio.h
58
u32 mp:1;
drivers/s390/cio/device_pgid.c
44
if (sch->config.mp != mpath) {
drivers/s390/cio/device_pgid.c
45
sch->config.mp = mpath;
drivers/s390/cio/device_pgid.c
584
if (sch->config.mp) {
drivers/s390/cio/device_pgid.c
585
sch->config.mp = 0;
drivers/scsi/aacraid/aachba.c
951
struct aac_driver_ident *mp = aac_get_driver_ident(dev->cardtype);
drivers/scsi/aacraid/aachba.c
953
inqstrcpy (mp->vname, str->vid);
drivers/scsi/aacraid/aachba.c
955
inqstrcpy (mp->model, str->pid);
drivers/scsi/fcoe/fcoe_ctlr.c
1311
struct fip_mac_desc *mp;
drivers/scsi/fcoe/fcoe_ctlr.c
1400
mp = (struct fip_mac_desc *)desc;
drivers/scsi/fcoe/fcoe_ctlr.c
1401
if (dlen < sizeof(*mp))
drivers/scsi/fcoe/fcoe_ctlr.c
1403
if (!ether_addr_equal(mp->fd_mac, fcf->fcf_mac))
drivers/scsi/libfc/fc_exch.c
1039
struct fc_exch_mgr *mp,
drivers/scsi/libfc/fc_exch.c
1057
ep = fc_exch_find(mp, xid);
drivers/scsi/libfc/fc_exch.c
1059
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1086
ep = fc_exch_find(mp, xid);
drivers/scsi/libfc/fc_exch.c
1089
atomic_inc(&mp->stats.xid_busy);
drivers/scsi/libfc/fc_exch.c
1093
ep = fc_exch_resp(lport, mp, fp);
drivers/scsi/libfc/fc_exch.c
1100
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1118
atomic_inc(&mp->stats.seq_not_found);
drivers/scsi/libfc/fc_exch.c
1168
static struct fc_seq *fc_seq_lookup_orig(struct fc_exch_mgr *mp,
drivers/scsi/libfc/fc_exch.c
117
struct fc_exch_mgr *mp;
drivers/scsi/libfc/fc_exch.c
1180
ep = fc_exch_find(mp, xid);
drivers/scsi/libfc/fc_exch.c
1484
fc_seq_lookup_recip(lport, ema->mp, fp) == FC_RJT_NONE)
drivers/scsi/libfc/fc_exch.c
1509
static void fc_exch_recv_req(struct fc_lport *lport, struct fc_exch_mgr *mp,
drivers/scsi/libfc/fc_exch.c
1536
reject = fc_seq_lookup_recip(lport, mp, fp);
drivers/scsi/libfc/fc_exch.c
1571
static void fc_exch_recv_seq_resp(struct fc_exch_mgr *mp, struct fc_frame *fp)
drivers/scsi/libfc/fc_exch.c
1580
ep = fc_exch_find(mp, ntohs(fh->fh_ox_id));
drivers/scsi/libfc/fc_exch.c
1582
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1586
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1592
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1597
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1665
static void fc_exch_recv_resp(struct fc_exch_mgr *mp, struct fc_frame *fp)
drivers/scsi/libfc/fc_exch.c
1669
sp = fc_seq_lookup_orig(mp, fp); /* doesn't hold sequence */
drivers/scsi/libfc/fc_exch.c
1672
atomic_inc(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
1674
atomic_inc(&mp->stats.non_bls_resp);
drivers/scsi/libfc/fc_exch.c
1764
static void fc_exch_recv_bls(struct fc_exch_mgr *mp, struct fc_frame *fp)
drivers/scsi/libfc/fc_exch.c
1774
ep = fc_exch_find(mp, (f_ctl & FC_FC_EX_CTX) ?
drivers/scsi/libfc/fc_exch.c
1987
per_cpu_ptr(ema->mp->pool, cpu),
drivers/scsi/libfc/fc_exch.c
2005
if (ema->mp->min_xid <= xid && xid <= ema->mp->max_xid)
drivers/scsi/libfc/fc_exch.c
2006
return fc_exch_find(ema->mp, xid);
drivers/scsi/libfc/fc_exch.c
2365
struct fc_exch_mgr *mp;
drivers/scsi/libfc/fc_exch.c
2370
mp = ema->mp;
drivers/scsi/libfc/fc_exch.c
2371
st->fc_no_free_exch += atomic_read(&mp->stats.no_free_exch);
drivers/scsi/libfc/fc_exch.c
2373
atomic_read(&mp->stats.no_free_exch_xid);
drivers/scsi/libfc/fc_exch.c
2374
st->fc_xid_not_found += atomic_read(&mp->stats.xid_not_found);
drivers/scsi/libfc/fc_exch.c
2375
st->fc_xid_busy += atomic_read(&mp->stats.xid_busy);
drivers/scsi/libfc/fc_exch.c
2376
st->fc_seq_not_found += atomic_read(&mp->stats.seq_not_found);
drivers/scsi/libfc/fc_exch.c
2377
st->fc_non_bls_resp += atomic_read(&mp->stats.non_bls_resp);
drivers/scsi/libfc/fc_exch.c
2389
struct fc_exch_mgr *mp,
drivers/scsi/libfc/fc_exch.c
2398
ema->mp = mp;
drivers/scsi/libfc/fc_exch.c
2402
kref_get(&mp->kref);
drivers/scsi/libfc/fc_exch.c
2413
struct fc_exch_mgr *mp = container_of(kref, struct fc_exch_mgr, kref);
drivers/scsi/libfc/fc_exch.c
2415
mempool_destroy(mp->ep_pool);
drivers/scsi/libfc/fc_exch.c
2416
free_percpu(mp->pool);
drivers/scsi/libfc/fc_exch.c
2417
kfree(mp);
drivers/scsi/libfc/fc_exch.c
2428
kref_put(&ema->mp->kref, fc_exch_mgr_destroy);
drivers/scsi/libfc/fc_exch.c
2443
if (!fc_exch_mgr_add(dst, ema->mp, ema->match))
drivers/scsi/libfc/fc_exch.c
2467
struct fc_exch_mgr *mp;
drivers/scsi/libfc/fc_exch.c
2483
mp = kzalloc_obj(struct fc_exch_mgr, GFP_ATOMIC);
drivers/scsi/libfc/fc_exch.c
2484
if (!mp)
drivers/scsi/libfc/fc_exch.c
2487
mp->class = class;
drivers/scsi/libfc/fc_exch.c
2488
mp->lport = lport;
drivers/scsi/libfc/fc_exch.c
2490
mp->min_xid = min_xid;
drivers/scsi/libfc/fc_exch.c
2496
mp->max_xid = pool_exch_range * (fc_cpu_mask + 1) +
drivers/scsi/libfc/fc_exch.c
2499
mp->max_xid = max_xid;
drivers/scsi/libfc/fc_exch.c
2500
pool_exch_range = (mp->max_xid - mp->min_xid + 1) /
drivers/scsi/libfc/fc_exch.c
2504
mp->ep_pool = mempool_create_slab_pool(2, fc_em_cachep);
drivers/scsi/libfc/fc_exch.c
2505
if (!mp->ep_pool)
drivers/scsi/libfc/fc_exch.c
2513
mp->pool_max_index = pool_exch_range - 1;
drivers/scsi/libfc/fc_exch.c
2519
mp->pool = __alloc_percpu(pool_size, __alignof__(struct fc_exch_pool));
drivers/scsi/libfc/fc_exch.c
2520
if (!mp->pool)
drivers/scsi/libfc/fc_exch.c
2523
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_exch.c
2531
kref_init(&mp->kref);
drivers/scsi/libfc/fc_exch.c
2532
if (!fc_exch_mgr_add(lport, mp, match)) {
drivers/scsi/libfc/fc_exch.c
2533
free_percpu(mp->pool);
drivers/scsi/libfc/fc_exch.c
2542
kref_put(&mp->kref, fc_exch_mgr_destroy);
drivers/scsi/libfc/fc_exch.c
2543
return mp;
drivers/scsi/libfc/fc_exch.c
2546
mempool_destroy(mp->ep_pool);
drivers/scsi/libfc/fc_exch.c
2548
kfree(mp);
drivers/scsi/libfc/fc_exch.c
2591
if ((xid >= ema->mp->min_xid) &&
drivers/scsi/libfc/fc_exch.c
2592
(xid <= ema->mp->max_xid))
drivers/scsi/libfc/fc_exch.c
2639
fc_exch_recv_bls(ema->mp, fp);
drivers/scsi/libfc/fc_exch.c
2642
fc_exch_recv_seq_resp(ema->mp, fp);
drivers/scsi/libfc/fc_exch.c
2644
fc_exch_recv_resp(ema->mp, fp);
drivers/scsi/libfc/fc_exch.c
2646
fc_exch_recv_req(lport, ema->mp, fp);
drivers/scsi/libfc/fc_exch.c
313
struct fc_exch_mgr *mp;
drivers/scsi/libfc/fc_exch.c
316
mp = ep->em;
drivers/scsi/libfc/fc_exch.c
320
mempool_free(ep, mp->ep_pool);
drivers/scsi/libfc/fc_exch.c
813
struct fc_exch_mgr *mp)
drivers/scsi/libfc/fc_exch.c
821
ep = mempool_alloc(mp->ep_pool, GFP_ATOMIC);
drivers/scsi/libfc/fc_exch.c
823
atomic_inc(&mp->stats.no_free_exch);
drivers/scsi/libfc/fc_exch.c
829
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_exch.c
851
index = index == mp->pool_max_index ? 0 : index + 1;
drivers/scsi/libfc/fc_exch.c
855
pool->next_index = index == mp->pool_max_index ? 0 : index + 1;
drivers/scsi/libfc/fc_exch.c
875
ep->oxid = ep->xid = (index << fc_cpu_order | cpu) + mp->min_xid;
drivers/scsi/libfc/fc_exch.c
876
ep->em = mp;
drivers/scsi/libfc/fc_exch.c
881
ep->class = mp->class;
drivers/scsi/libfc/fc_exch.c
889
atomic_inc(&mp->stats.no_free_exch_xid);
drivers/scsi/libfc/fc_exch.c
890
mempool_free(ep, mp->ep_pool);
drivers/scsi/libfc/fc_exch.c
913
ep = fc_exch_em_alloc(lport, ema->mp);
drivers/scsi/libfc/fc_exch.c
926
static struct fc_exch *fc_exch_find(struct fc_exch_mgr *mp, u16 xid)
drivers/scsi/libfc/fc_exch.c
928
struct fc_lport *lport = mp->lport;
drivers/scsi/libfc/fc_exch.c
942
if ((xid >= mp->min_xid) && (xid <= mp->max_xid)) {
drivers/scsi/libfc/fc_exch.c
943
pool = per_cpu_ptr(mp->pool, cpu);
drivers/scsi/libfc/fc_exch.c
945
ep = fc_exch_ptr_get(pool, (xid - mp->min_xid) >> fc_cpu_order);
drivers/scsi/libfc/fc_exch.c
991
struct fc_exch_mgr *mp,
drivers/scsi/libfc/fc_lport.c
392
__be32 *mp;
drivers/scsi/libfc/fc_lport.c
394
mp = &lport->fcts.ff_type_map[type / FC_NS_BPW];
drivers/scsi/libfc/fc_lport.c
395
*mp = htonl(ntohl(*mp) | 1UL << (type % FC_NS_BPW));
drivers/scsi/lpfc/lpfc_bsg.c
163
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_bsg.c
177
mp = kmalloc_obj(struct lpfc_dmabuf);
drivers/scsi/lpfc/lpfc_bsg.c
178
if (!mp) {
drivers/scsi/lpfc/lpfc_bsg.c
184
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_bsg.c
185
mp->virt = lpfc_mbuf_alloc(phba, MEM_PRI, &(mp->phys));
drivers/scsi/lpfc/lpfc_bsg.c
187
if (!mp->virt) {
drivers/scsi/lpfc/lpfc_bsg.c
188
kfree(mp);
drivers/scsi/lpfc/lpfc_bsg.c
196
mlist = mp;
drivers/scsi/lpfc/lpfc_bsg.c
198
list_add_tail(&mp->list, &mlist->list);
drivers/scsi/lpfc/lpfc_bsg.c
205
bpl->addrLow = le32_to_cpu(putPaddrLow(mp->phys));
drivers/scsi/lpfc/lpfc_bsg.c
206
bpl->addrHigh = le32_to_cpu(putPaddrHigh(mp->phys));
drivers/scsi/lpfc/lpfc_bsg.c
223
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_bsg.c
244
list_for_each_entry(mp, &temp_list, list) {
drivers/scsi/lpfc/lpfc_bsg.c
248
dma_address = mp->virt + dma_offset;
drivers/scsi/lpfc/lpfc_bsg.c
2901
struct lpfc_dmabuf *mp[2] = {NULL, NULL};
drivers/scsi/lpfc/lpfc_bsg.c
2936
mp[i] = list_entry(curr, struct lpfc_dmabuf, list);
drivers/scsi/lpfc/lpfc_bsg.c
2940
mp[i]->buffer_tag = lpfc_sli_get_buffer_tag(phba);
drivers/scsi/lpfc/lpfc_bsg.c
2942
putPaddrHigh(mp[i]->phys);
drivers/scsi/lpfc/lpfc_bsg.c
2944
putPaddrLow(mp[i]->phys);
drivers/scsi/lpfc/lpfc_bsg.c
2946
((struct lpfc_dmabufext *)mp[i])->size;
drivers/scsi/lpfc/lpfc_bsg.c
2947
cmd->un.quexri64cx.buff.buffer_tag = mp[i]->buffer_tag;
drivers/scsi/lpfc/lpfc_bsg.c
2955
cmd->un.cont64[i].addrHigh = putPaddrHigh(mp[i]->phys);
drivers/scsi/lpfc/lpfc_bsg.c
2956
cmd->un.cont64[i].addrLow = putPaddrLow(mp[i]->phys);
drivers/scsi/lpfc/lpfc_bsg.c
2958
((struct lpfc_dmabufext *)mp[i])->size;
drivers/scsi/lpfc/lpfc_bsg.c
2975
(struct lpfc_dmabufext *)mp[0]);
drivers/scsi/lpfc/lpfc_bsg.c
2976
if (mp[1])
drivers/scsi/lpfc/lpfc_bsg.c
2978
(struct lpfc_dmabufext *)mp[1]);
drivers/scsi/lpfc/lpfc_bsg.c
2984
lpfc_sli_ringpostbuf_put(phba, pring, mp[0]);
drivers/scsi/lpfc/lpfc_bsg.c
2985
if (mp[1]) {
drivers/scsi/lpfc/lpfc_bsg.c
2986
lpfc_sli_ringpostbuf_put(phba, pring, mp[1]);
drivers/scsi/lpfc/lpfc_bsg.c
2987
mp[1] = NULL;
drivers/scsi/lpfc/lpfc_crtn.h
442
void lpfc_rq_buf_free(struct lpfc_hba *phba, struct lpfc_dmabuf *mp);
drivers/scsi/lpfc/lpfc_ct.c
100
struct lpfc_dmabuf *mp, uint32_t size)
drivers/scsi/lpfc/lpfc_ct.c
102
lpfc_ct_ignore_hbq_buffer(phba, piocbq, mp, size);
drivers/scsi/lpfc/lpfc_ct.c
119
struct lpfc_dmabuf *mp, *bmp;
drivers/scsi/lpfc/lpfc_ct.c
125
mp = cmdiocb->rsp_dmabuf;
drivers/scsi/lpfc/lpfc_ct.c
127
if (mp) {
drivers/scsi/lpfc/lpfc_ct.c
128
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
129
kfree(mp);
drivers/scsi/lpfc/lpfc_ct.c
162
struct lpfc_dmabuf *mp = NULL;
drivers/scsi/lpfc/lpfc_ct.c
168
mp = kmalloc_obj(*mp);
drivers/scsi/lpfc/lpfc_ct.c
169
if (!mp) {
drivers/scsi/lpfc/lpfc_ct.c
174
mp->virt = lpfc_mbuf_alloc(phba, MEM_PRI, &mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
175
if (!mp->virt) {
drivers/scsi/lpfc/lpfc_ct.c
1883
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_ct.c
1886
mp = cmdiocb->cmd_dmabuf;
drivers/scsi/lpfc/lpfc_ct.c
1887
if (mp == NULL)
drivers/scsi/lpfc/lpfc_ct.c
1889
CtReq = (struct lpfc_sli_ct_request *)mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
1909
struct lpfc_dmabuf *mp, *bmp;
drivers/scsi/lpfc/lpfc_ct.c
1927
mp = kmalloc_obj(struct lpfc_dmabuf);
drivers/scsi/lpfc/lpfc_ct.c
1928
if (!mp) {
drivers/scsi/lpfc/lpfc_ct.c
193
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_ct.c
1933
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_ct.c
1934
mp->virt = lpfc_mbuf_alloc(phba, MEM_PRI, &(mp->phys));
drivers/scsi/lpfc/lpfc_ct.c
1935
if (!mp->virt) {
drivers/scsi/lpfc/lpfc_ct.c
1962
bpl->addrHigh = le32_to_cpu(putPaddrHigh(mp->phys) );
drivers/scsi/lpfc/lpfc_ct.c
1963
bpl->addrLow = le32_to_cpu(putPaddrLow(mp->phys) );
drivers/scsi/lpfc/lpfc_ct.c
198
bpl->addrHigh = le32_to_cpu(putPaddrHigh(mp->phys));
drivers/scsi/lpfc/lpfc_ct.c
199
bpl->addrLow = le32_to_cpu(putPaddrLow(mp->phys));
drivers/scsi/lpfc/lpfc_ct.c
1991
CtReq = (struct lpfc_sli_ct_request *) mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
204
ct_rsp = (struct lpfc_sli_ct_request *)mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
2168
if (!lpfc_ct_cmd(vport, mp, bmp, ndlp, cmpl, rsp_size, retry)) {
drivers/scsi/lpfc/lpfc_ct.c
2182
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
2184
kfree(mp);
drivers/scsi/lpfc/lpfc_ct.c
235
cmdiocbq->rsp_dmabuf = mp;
drivers/scsi/lpfc/lpfc_ct.c
263
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
265
kfree(mp);
drivers/scsi/lpfc/lpfc_ct.c
340
struct lpfc_dmabuf *mp = NULL;
drivers/scsi/lpfc/lpfc_ct.c
3697
struct lpfc_dmabuf *mp, *bmp;
drivers/scsi/lpfc/lpfc_ct.c
3719
mp = kmalloc_obj(*mp);
drivers/scsi/lpfc/lpfc_ct.c
3720
if (!mp)
drivers/scsi/lpfc/lpfc_ct.c
3723
mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
3724
if (!mp->virt)
drivers/scsi/lpfc/lpfc_ct.c
3736
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_ct.c
3742
ctreq = (struct lpfc_sli_ct_request *)mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
3743
data = mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
3808
bpl->addrHigh = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
3809
bpl->addrLow = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
3816
if (!lpfc_ct_cmd(vport, mp, bmp, ndlp, cmpl, rsp_size, retry))
drivers/scsi/lpfc/lpfc_ct.c
3824
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_ct.c
3826
kfree(mp);
drivers/scsi/lpfc/lpfc_ct.c
440
mp = lpfc_sli_ringpostbuf_get(phba, pring,
drivers/scsi/lpfc/lpfc_ct.c
443
lpfc_ct_unsol_buffer(phba, iocbq, mp, size);
drivers/scsi/lpfc/lpfc_ct.c
444
lpfc_in_buf_free(phba, mp);
drivers/scsi/lpfc/lpfc_ct.c
493
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_ct.c
501
mp = kmalloc_obj(struct lpfc_dmabuf);
drivers/scsi/lpfc/lpfc_ct.c
502
if (!mp) {
drivers/scsi/lpfc/lpfc_ct.c
508
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_ct.c
512
mp->virt = lpfc_mbuf_alloc(phba, MEM_PRI, &(mp->phys));
drivers/scsi/lpfc/lpfc_ct.c
514
mp->virt = lpfc_mbuf_alloc(phba, 0, &(mp->phys));
drivers/scsi/lpfc/lpfc_ct.c
516
if (!mp->virt) {
drivers/scsi/lpfc/lpfc_ct.c
517
kfree(mp);
drivers/scsi/lpfc/lpfc_ct.c
525
mlist = mp;
drivers/scsi/lpfc/lpfc_ct.c
527
list_add_tail(&mp->list, &mlist->list);
drivers/scsi/lpfc/lpfc_ct.c
531
bpl->addrLow = le32_to_cpu(putPaddrLow(mp->phys) );
drivers/scsi/lpfc/lpfc_ct.c
532
bpl->addrHigh = le32_to_cpu(putPaddrHigh(mp->phys) );
drivers/scsi/lpfc/lpfc_ct.c
84
struct lpfc_dmabuf *mp, uint32_t size)
drivers/scsi/lpfc/lpfc_ct.c
844
lpfc_ns_rsp(struct lpfc_vport *vport, struct lpfc_dmabuf *mp, uint8_t fc4_type,
drivers/scsi/lpfc/lpfc_ct.c
848
(struct lpfc_sli_ct_request *) mp->virt;
drivers/scsi/lpfc/lpfc_ct.c
86
if (!mp) {
drivers/scsi/lpfc/lpfc_ct.c
861
list_add_tail(&head, &mp->list);
drivers/scsi/lpfc/lpfc_ct.c
862
list_for_each_entry_safe(mp, next_mp, &head, list) {
drivers/scsi/lpfc/lpfc_ct.c
863
mlast = mp;
drivers/scsi/lpfc/lpfc_els.c
7351
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_els.c
7365
mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_els.c
7366
mpsave = mp;
drivers/scsi/lpfc/lpfc_els.c
7367
virt = mp->virt;
drivers/scsi/lpfc/lpfc_els.c
7388
mbox->u.mqe.un.mem_dump_type3.addr_lo = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_els.c
7389
mbox->u.mqe.un.mem_dump_type3.addr_hi = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_els.c
7400
mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_els.c
7402
mp = mpsave;
drivers/scsi/lpfc/lpfc_els.c
7409
lpfc_sli_bemem_bcopy(mp->virt, &rdp_context->page_a0,
drivers/scsi/lpfc/lpfc_els.c
7413
memset(mp->virt, 0, DMP_SFF_PAGE_A2_SIZE);
drivers/scsi/lpfc/lpfc_els.c
7414
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_els.c
7417
mbox->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_els.c
7447
mbox->u.mqe.un.mem_dump_type3.addr_lo = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_els.c
7448
mbox->u.mqe.un.mem_dump_type3.addr_hi = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_els.c
7461
lpfc_sli_bemem_bcopy(mp->virt, &rdp_context->page_a2,
drivers/scsi/lpfc/lpfc_hbadisc.c
3442
struct lpfc_dmabuf *mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_hbadisc.c
3459
memcpy((uint8_t *) &vport->fc_sparam, (uint8_t *) mp->virt,
drivers/scsi/lpfc/lpfc_hbadisc.c
3747
struct lpfc_dmabuf *mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_hbadisc.c
3768
memcpy(&phba->alpa_map[0], mp->virt, 128);
drivers/scsi/lpfc/lpfc_hbadisc.c
3861
struct lpfc_dmabuf *mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_hbadisc.c
3901
pmb->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_hbadisc.c
4044
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_hbadisc.c
4073
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_hbadisc.c
4074
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_hbadisc.c
4075
kfree(mp);
drivers/scsi/lpfc/lpfc_hbadisc.c
4096
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_hbadisc.c
4101
memcpy(vport_buff + offset, mp->virt, byte_count);
drivers/scsi/lpfc/lpfc_init.c
423
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_init.c
463
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_init.c
469
memcpy(&vport->fc_sparam, mp->virt, sizeof (struct serv_parm));
drivers/scsi/lpfc/lpfc_init.c
470
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_init.c
471
kfree(mp);
drivers/scsi/lpfc/lpfc_init.c
969
struct lpfc_dmabuf *mp, *next_mp;
drivers/scsi/lpfc/lpfc_init.c
983
list_for_each_entry_safe(mp, next_mp, &buflist, list) {
drivers/scsi/lpfc/lpfc_init.c
984
list_del(&mp->list);
drivers/scsi/lpfc/lpfc_init.c
986
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_init.c
987
kfree(mp);
drivers/scsi/lpfc/lpfc_mbox.c
103
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_mbox.c
105
mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
109
if (mp) {
drivers/scsi/lpfc/lpfc_mbox.c
111
__lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
113
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
114
kfree(mp);
drivers/scsi/lpfc/lpfc_mbox.c
137
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_mbox.c
165
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
166
mb->un.varWords[3] = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
167
mb->un.varWords[4] = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2331
struct lpfc_dmabuf *mp = NULL;
drivers/scsi/lpfc/lpfc_mbox.c
2350
mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
2351
mb->un.varWords[3] = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2352
mb->un.varWords[4] = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2379
struct lpfc_dmabuf *mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
2385
lpfc_sli_bemem_bcopy(mp->virt, &rdp_context->page_a2,
drivers/scsi/lpfc/lpfc_mbox.c
2392
mbox->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_mbox.c
2409
struct lpfc_dmabuf *mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
2415
lpfc_sli_bemem_bcopy(mp->virt, &rdp_context->page_a0,
drivers/scsi/lpfc/lpfc_mbox.c
2420
memset(mp->virt, 0, DMP_SFF_PAGE_A2_SIZE);
drivers/scsi/lpfc/lpfc_mbox.c
2421
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_mbox.c
2424
mbox->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_mbox.c
2436
mbox->u.mqe.un.mem_dump_type3.addr_lo = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2437
mbox->u.mqe.un.mem_dump_type3.addr_hi = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2465
struct lpfc_dmabuf *mp = NULL;
drivers/scsi/lpfc/lpfc_mbox.c
2486
mp = mbox->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
2487
mbox->u.mqe.un.mem_dump_type3.addr_lo = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
2488
mbox->u.mqe.un.mem_dump_type3.addr_hi = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
352
struct lpfc_dmabuf *mp)
drivers/scsi/lpfc/lpfc_mbox.c
359
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_mbox.c
362
mb->un.varReadTop.lilpBde64.addrHigh = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
363
mb->un.varReadTop.lilpBde64.addrLow = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
368
pmb->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_mbox.c
65
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_mbox.c
668
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_mbox.c
67
mp = kmalloc_obj(*mp);
drivers/scsi/lpfc/lpfc_mbox.c
68
if (!mp)
drivers/scsi/lpfc/lpfc_mbox.c
682
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
687
mb->un.varRdSparm.un.sp64.addrHigh = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
688
mb->un.varRdSparm.un.sp64.addrLow = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
71
mp->virt = lpfc_mbuf_alloc(phba, 0, &mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
72
if (!mp->virt) {
drivers/scsi/lpfc/lpfc_mbox.c
73
kfree(mp);
drivers/scsi/lpfc/lpfc_mbox.c
77
memset(mp->virt, 0, LPFC_BPL_SIZE);
drivers/scsi/lpfc/lpfc_mbox.c
80
INIT_LIST_HEAD(&mp->list);
drivers/scsi/lpfc/lpfc_mbox.c
81
mbox->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_mbox.c
811
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_mbox.c
836
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_mbox.c
837
sparam = mp->virt;
drivers/scsi/lpfc/lpfc_mbox.c
843
mb->un.varRegLogin.un.sp64.addrHigh = putPaddrHigh(mp->phys);
drivers/scsi/lpfc/lpfc_mbox.c
844
mb->un.varRegLogin.un.sp64.addrLow = putPaddrLow(mp->phys);
drivers/scsi/lpfc/lpfc_mem.c
678
lpfc_in_buf_free(struct lpfc_hba *phba, struct lpfc_dmabuf *mp)
drivers/scsi/lpfc/lpfc_mem.c
683
if (!mp)
drivers/scsi/lpfc/lpfc_mem.c
687
hbq_entry = container_of(mp, struct hbq_dmabuf, dbuf);
drivers/scsi/lpfc/lpfc_mem.c
703
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_mem.c
704
kfree(mp);
drivers/scsi/lpfc/lpfc_mem.c
722
lpfc_rq_buf_free(struct lpfc_hba *phba, struct lpfc_dmabuf *mp)
drivers/scsi/lpfc/lpfc_mem.c
731
if (!mp)
drivers/scsi/lpfc/lpfc_mem.c
734
rqb_entry = container_of(mp, struct rqb_dmabuf, hbuf);
drivers/scsi/lpfc/lpfc_sli.c
12232
struct lpfc_dmabuf *mp)
drivers/scsi/lpfc/lpfc_sli.c
12237
list_add_tail(&mp->list, &pring->postbufq);
drivers/scsi/lpfc/lpfc_sli.c
12287
struct lpfc_dmabuf *mp, *next_mp;
drivers/scsi/lpfc/lpfc_sli.c
12292
list_for_each_entry_safe(mp, next_mp, &pring->postbufq, list) {
drivers/scsi/lpfc/lpfc_sli.c
12293
if (mp->buffer_tag == tag) {
drivers/scsi/lpfc/lpfc_sli.c
12294
list_del_init(&mp->list);
drivers/scsi/lpfc/lpfc_sli.c
12297
return mp;
drivers/scsi/lpfc/lpfc_sli.c
12331
struct lpfc_dmabuf *mp, *next_mp;
drivers/scsi/lpfc/lpfc_sli.c
12336
list_for_each_entry_safe(mp, next_mp, &pring->postbufq, list) {
drivers/scsi/lpfc/lpfc_sli.c
12337
if (mp->phys == phys) {
drivers/scsi/lpfc/lpfc_sli.c
12338
list_del_init(&mp->list);
drivers/scsi/lpfc/lpfc_sli.c
12341
return mp;
drivers/scsi/lpfc/lpfc_sli.c
13700
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
13894
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
13908
pmb->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_sli.c
14345
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
14395
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
14407
pmb->ctx_buf = mp;
drivers/scsi/lpfc/lpfc_sli.c
20761
struct lpfc_dmabuf *mp = NULL;
drivers/scsi/lpfc/lpfc_sli.c
20779
mp = mboxq->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
20790
lpfc_sli_pcimem_bcopy((char *)mp->virt, rgn23_data, data_length);
drivers/scsi/lpfc/lpfc_sli.c
2860
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
2873
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
2874
if (mp) {
drivers/scsi/lpfc/lpfc_sli.c
2876
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_sli.c
2877
kfree(mp);
drivers/scsi/lpfc/lpfc_sli.c
5809
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
5830
mp = mboxq->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
5864
lpfc_parse_fcoe_conf(phba, mp->virt, data_length);
drivers/scsi/lpfc/lpfc_sli.c
8213
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
8296
mp = kmalloc_obj(*mp);
drivers/scsi/lpfc/lpfc_sli.c
8297
if (mp)
drivers/scsi/lpfc/lpfc_sli.c
8298
mp->virt = dma_alloc_coherent
drivers/scsi/lpfc/lpfc_sli.c
8301
&mp->phys, GFP_KERNEL);
drivers/scsi/lpfc/lpfc_sli.c
8302
if (!mp || !mp->virt) {
drivers/scsi/lpfc/lpfc_sli.c
8306
kfree(mp);
drivers/scsi/lpfc/lpfc_sli.c
8311
phba->cgn_i = mp;
drivers/scsi/lpfc/lpfc_sli.c
8531
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_sli.c
8846
mp = mboxq->ctx_buf;
drivers/scsi/lpfc/lpfc_sli.c
8848
memcpy(&vport->fc_sparam, mp->virt, sizeof(struct serv_parm));
drivers/scsi/lpfc/lpfc_sli.c
8857
lpfc_mbuf_free(phba, mp->virt, mp->phys);
drivers/scsi/lpfc/lpfc_sli.c
8858
kfree(mp);
drivers/scsi/lpfc/lpfc_vport.c
122
struct lpfc_dmabuf *mp;
drivers/scsi/lpfc/lpfc_vport.c
169
mp = pmb->ctx_buf;
drivers/scsi/lpfc/lpfc_vport.c
170
memcpy(&vport->fc_sparam, mp->virt, sizeof (struct serv_parm));
drivers/scsi/ncr53c8xx.c
227
static void *___m_alloc(m_pool_s *mp, int size)
drivers/scsi/ncr53c8xx.c
233
m_link_s *h = mp->h;
drivers/scsi/ncr53c8xx.c
246
h[j].next = (m_link_s *)mp->getp(mp);
drivers/scsi/ncr53c8xx.c
270
static void ___m_free(m_pool_s *mp, void *ptr, int size)
drivers/scsi/ncr53c8xx.c
276
m_link_s *h = mp->h;
drivers/scsi/ncr53c8xx.c
295
mp->freep(mp, a);
drivers/scsi/ncr53c8xx.c
318
static void *__m_calloc2(m_pool_s *mp, int size, char *name, int uflags)
drivers/scsi/ncr53c8xx.c
322
p = ___m_alloc(mp, size);
drivers/scsi/ncr53c8xx.c
335
#define __m_calloc(mp, s, n) __m_calloc2(mp, s, n, MEMO_WARN)
drivers/scsi/ncr53c8xx.c
337
static void __m_free(m_pool_s *mp, void *ptr, int size, char *name)
drivers/scsi/ncr53c8xx.c
342
___m_free(mp, ptr, size);
drivers/scsi/ncr53c8xx.c
352
static m_addr_t ___mp0_getp(m_pool_s *mp)
drivers/scsi/ncr53c8xx.c
356
++mp->nump;
drivers/scsi/ncr53c8xx.c
360
static void ___mp0_freep(m_pool_s *mp, m_addr_t m)
drivers/scsi/ncr53c8xx.c
363
--mp->nump;
drivers/scsi/ncr53c8xx.c
376
static m_addr_t ___dma_getp(m_pool_s *mp)
drivers/scsi/ncr53c8xx.c
384
vp = (m_addr_t) dma_alloc_coherent(mp->bush,
drivers/scsi/ncr53c8xx.c
391
vbp->next = mp->vtob[hc];
drivers/scsi/ncr53c8xx.c
392
mp->vtob[hc] = vbp;
drivers/scsi/ncr53c8xx.c
393
++mp->nump;
drivers/scsi/ncr53c8xx.c
402
static void ___dma_freep(m_pool_s *mp, m_addr_t m)
drivers/scsi/ncr53c8xx.c
407
vbpp = &mp->vtob[hc];
drivers/scsi/ncr53c8xx.c
413
dma_free_coherent(mp->bush, PAGE_SIZE<<MEMO_PAGE_ORDER,
drivers/scsi/ncr53c8xx.c
416
--mp->nump;
drivers/scsi/ncr53c8xx.c
422
m_pool_s *mp;
drivers/scsi/ncr53c8xx.c
423
for (mp = mp0.next; mp && mp->bush != bush; mp = mp->next);
drivers/scsi/ncr53c8xx.c
424
return mp;
drivers/scsi/ncr53c8xx.c
429
m_pool_s *mp;
drivers/scsi/ncr53c8xx.c
430
mp = __m_calloc(&mp0, sizeof(*mp), "MPOOL");
drivers/scsi/ncr53c8xx.c
431
if (mp) {
drivers/scsi/ncr53c8xx.c
432
memset(mp, 0, sizeof(*mp));
drivers/scsi/ncr53c8xx.c
433
mp->bush = bush;
drivers/scsi/ncr53c8xx.c
434
mp->getp = ___dma_getp;
drivers/scsi/ncr53c8xx.c
435
mp->freep = ___dma_freep;
drivers/scsi/ncr53c8xx.c
436
mp->next = mp0.next;
drivers/scsi/ncr53c8xx.c
437
mp0.next = mp;
drivers/scsi/ncr53c8xx.c
439
return mp;
drivers/scsi/ncr53c8xx.c
457
struct m_pool *mp;
drivers/scsi/ncr53c8xx.c
461
mp = ___get_dma_pool(bush);
drivers/scsi/ncr53c8xx.c
462
if (!mp)
drivers/scsi/ncr53c8xx.c
463
mp = ___cre_dma_pool(bush);
drivers/scsi/ncr53c8xx.c
464
if (mp)
drivers/scsi/ncr53c8xx.c
465
m = __m_calloc(mp, size, name);
drivers/scsi/ncr53c8xx.c
466
if (mp && !mp->nump)
drivers/scsi/ncr53c8xx.c
467
___del_dma_pool(mp);
drivers/scsi/ncr53c8xx.c
476
struct m_pool *mp;
drivers/scsi/ncr53c8xx.c
479
mp = ___get_dma_pool(bush);
drivers/scsi/ncr53c8xx.c
480
if (mp)
drivers/scsi/ncr53c8xx.c
481
__m_free(mp, m, size, name);
drivers/scsi/ncr53c8xx.c
482
if (mp && !mp->nump)
drivers/scsi/ncr53c8xx.c
483
___del_dma_pool(mp);
drivers/scsi/ncr53c8xx.c
490
m_pool_s *mp;
drivers/scsi/ncr53c8xx.c
496
mp = ___get_dma_pool(bush);
drivers/scsi/ncr53c8xx.c
497
if (mp) {
drivers/scsi/ncr53c8xx.c
498
vp = mp->vtob[hc];
drivers/scsi/qedf/qedf_fip.c
173
struct fip_mac_desc *mp;
drivers/scsi/qedf/qedf_fip.c
237
mp = (struct fip_mac_desc *)desc;
drivers/scsi/qedf/qedf_fip.c
239
"Switch fd_mac=%pM.\n", mp->fd_mac);
drivers/scsi/qedf/qedf_fip.c
240
if (ether_addr_equal(mp->fd_mac,
drivers/scsi/sym53c8xx_2/sym_hipd.h
1156
#define M_GET_MEM_CLUSTER() mp->get_mem_cluster(mp)
drivers/scsi/sym53c8xx_2/sym_hipd.h
1157
#define M_FREE_MEM_CLUSTER(p) mp->free_mem_cluster(mp, p)
drivers/scsi/sym53c8xx_2/sym_hipd.h
1193
static inline void *sym_m_get_dma_mem_cluster(m_pool_p mp, m_vtob_p vbp)
drivers/scsi/sym53c8xx_2/sym_hipd.h
1198
vaddr = dma_alloc_coherent(mp->dev_dmat, SYM_MEM_CLUSTER_SIZE, &baddr,
drivers/scsi/sym53c8xx_2/sym_hipd.h
1207
static inline void sym_m_free_dma_mem_cluster(m_pool_p mp, m_vtob_p vbp)
drivers/scsi/sym53c8xx_2/sym_hipd.h
1209
dma_free_coherent(mp->dev_dmat, SYM_MEM_CLUSTER_SIZE, vbp->vaddr,
drivers/scsi/sym53c8xx_2/sym_malloc.c
145
static void *__sym_calloc2(m_pool_p mp, int size, char *name, int uflags)
drivers/scsi/sym53c8xx_2/sym_malloc.c
149
p = ___sym_malloc(mp, size);
drivers/scsi/sym53c8xx_2/sym_malloc.c
161
#define __sym_calloc(mp, s, n) __sym_calloc2(mp, s, n, SYM_MEM_WARN)
drivers/scsi/sym53c8xx_2/sym_malloc.c
166
static void __sym_mfree(m_pool_p mp, void *ptr, int size, char *name)
drivers/scsi/sym53c8xx_2/sym_malloc.c
171
___sym_mfree(mp, ptr, size);
drivers/scsi/sym53c8xx_2/sym_malloc.c
180
static void *___mp0_get_mem_cluster(m_pool_p mp)
drivers/scsi/sym53c8xx_2/sym_malloc.c
184
++mp->nump;
drivers/scsi/sym53c8xx_2/sym_malloc.c
189
static void ___mp0_free_mem_cluster(m_pool_p mp, void *m)
drivers/scsi/sym53c8xx_2/sym_malloc.c
192
--mp->nump;
drivers/scsi/sym53c8xx_2/sym_malloc.c
210
static void * ___get_dma_mem_cluster(m_pool_p mp)
drivers/scsi/sym53c8xx_2/sym_malloc.c
219
vaddr = sym_m_get_dma_mem_cluster(mp, vbp);
drivers/scsi/sym53c8xx_2/sym_malloc.c
222
vbp->next = mp->vtob[hc];
drivers/scsi/sym53c8xx_2/sym_malloc.c
223
mp->vtob[hc] = vbp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
224
++mp->nump;
drivers/scsi/sym53c8xx_2/sym_malloc.c
233
static void ___free_dma_mem_cluster(m_pool_p mp, void *m)
drivers/scsi/sym53c8xx_2/sym_malloc.c
238
vbpp = &mp->vtob[hc];
drivers/scsi/sym53c8xx_2/sym_malloc.c
244
sym_m_free_dma_mem_cluster(mp, vbp);
drivers/scsi/sym53c8xx_2/sym_malloc.c
246
--mp->nump;
drivers/scsi/sym53c8xx_2/sym_malloc.c
254
m_pool_p mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
255
for (mp = mp0.next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
256
mp && !sym_m_pool_match(mp->dev_dmat, dev_dmat);
drivers/scsi/sym53c8xx_2/sym_malloc.c
257
mp = mp->next);
drivers/scsi/sym53c8xx_2/sym_malloc.c
258
return mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
264
m_pool_p mp = __sym_calloc(&mp0, sizeof(*mp), "MPOOL");
drivers/scsi/sym53c8xx_2/sym_malloc.c
265
if (mp) {
drivers/scsi/sym53c8xx_2/sym_malloc.c
266
mp->dev_dmat = dev_dmat;
drivers/scsi/sym53c8xx_2/sym_malloc.c
267
mp->get_mem_cluster = ___get_dma_mem_cluster;
drivers/scsi/sym53c8xx_2/sym_malloc.c
269
mp->free_mem_cluster = ___free_dma_mem_cluster;
drivers/scsi/sym53c8xx_2/sym_malloc.c
271
mp->next = mp0.next;
drivers/scsi/sym53c8xx_2/sym_malloc.c
272
mp0.next = mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
273
return mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
302
m_pool_p mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
306
mp = ___get_dma_pool(dev_dmat);
drivers/scsi/sym53c8xx_2/sym_malloc.c
307
if (!mp)
drivers/scsi/sym53c8xx_2/sym_malloc.c
308
mp = ___cre_dma_pool(dev_dmat);
drivers/scsi/sym53c8xx_2/sym_malloc.c
309
if (!mp)
drivers/scsi/sym53c8xx_2/sym_malloc.c
311
m = __sym_calloc(mp, size, name);
drivers/scsi/sym53c8xx_2/sym_malloc.c
313
if (!mp->nump)
drivers/scsi/sym53c8xx_2/sym_malloc.c
314
___del_dma_pool(mp);
drivers/scsi/sym53c8xx_2/sym_malloc.c
325
m_pool_p mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
328
mp = ___get_dma_pool(dev_dmat);
drivers/scsi/sym53c8xx_2/sym_malloc.c
329
if (!mp)
drivers/scsi/sym53c8xx_2/sym_malloc.c
331
__sym_mfree(mp, m, size, name);
drivers/scsi/sym53c8xx_2/sym_malloc.c
333
if (!mp->nump)
drivers/scsi/sym53c8xx_2/sym_malloc.c
334
___del_dma_pool(mp);
drivers/scsi/sym53c8xx_2/sym_malloc.c
347
m_pool_p mp;
drivers/scsi/sym53c8xx_2/sym_malloc.c
354
mp = ___get_dma_pool(dev_dmat);
drivers/scsi/sym53c8xx_2/sym_malloc.c
355
if (mp) {
drivers/scsi/sym53c8xx_2/sym_malloc.c
356
vp = mp->vtob[hc];
drivers/scsi/sym53c8xx_2/sym_malloc.c
47
static void *___sym_malloc(m_pool_p mp, int size)
drivers/scsi/sym53c8xx_2/sym_malloc.c
53
m_link_p h = mp->h;
drivers/scsi/sym53c8xx_2/sym_malloc.c
93
static void ___sym_mfree(m_pool_p mp, void *ptr, int size)
drivers/scsi/sym53c8xx_2/sym_malloc.c
99
m_link_p h = mp->h;
drivers/tee/optee/ffa_abi.c
125
u32 attr, const struct optee_msg_param *mp)
drivers/tee/optee/ffa_abi.c
133
p->u.memref.size = mp->u.fmem.size;
drivers/tee/optee/ffa_abi.c
135
if (mp->u.fmem.global_id != OPTEE_MSG_FMEM_INVALID_GLOBAL_ID)
drivers/tee/optee/ffa_abi.c
136
shm = optee_shm_from_ffa_handle(optee, mp->u.fmem.global_id);
drivers/tee/optee/ffa_abi.c
140
offs_low = mp->u.fmem.offs_low;
drivers/tee/optee/ffa_abi.c
141
offs_high = mp->u.fmem.offs_high;
drivers/tee/optee/ffa_abi.c
164
const struct optee_msg_param *mp = msg_params + n;
drivers/tee/optee/ffa_abi.c
165
u32 attr = mp->attr & OPTEE_MSG_ATTR_TYPE_MASK;
drivers/tee/optee/ffa_abi.c
175
optee_from_msg_param_value(p, attr, mp);
drivers/tee/optee/ffa_abi.c
180
from_msg_param_ffa_mem(optee, p, attr, mp);
drivers/tee/optee/ffa_abi.c
190
static int to_msg_param_ffa_mem(struct optee_msg_param *mp,
drivers/tee/optee/ffa_abi.c
195
mp->attr = OPTEE_MSG_ATTR_TYPE_FMEM_INPUT + p->attr -
drivers/tee/optee/ffa_abi.c
201
mp->u.fmem.internal_offs = shm->offset;
drivers/tee/optee/ffa_abi.c
203
mp->u.fmem.offs_low = shm_offs;
drivers/tee/optee/ffa_abi.c
204
mp->u.fmem.offs_high = shm_offs >> 32;
drivers/tee/optee/ffa_abi.c
206
if (mp->u.fmem.offs_high != shm_offs >> 32)
drivers/tee/optee/ffa_abi.c
209
mp->u.fmem.global_id = shm->sec_world_id;
drivers/tee/optee/ffa_abi.c
211
memset(&mp->u, 0, sizeof(mp->u));
drivers/tee/optee/ffa_abi.c
212
mp->u.fmem.global_id = OPTEE_MSG_FMEM_INVALID_GLOBAL_ID;
drivers/tee/optee/ffa_abi.c
214
mp->u.fmem.size = p->u.memref.size;
drivers/tee/optee/ffa_abi.c
237
struct optee_msg_param *mp = msg_params + n;
drivers/tee/optee/ffa_abi.c
241
mp->attr = TEE_IOCTL_PARAM_ATTR_TYPE_NONE;
drivers/tee/optee/ffa_abi.c
242
memset(&mp->u, 0, sizeof(mp->u));
drivers/tee/optee/ffa_abi.c
247
optee_to_msg_param_value(mp, p);
drivers/tee/optee/ffa_abi.c
252
if (to_msg_param_ffa_mem(mp, p))
drivers/tee/optee/optee_private.h
353
const struct optee_msg_param *mp)
drivers/tee/optee/optee_private.h
357
p->u.value.a = mp->u.value.a;
drivers/tee/optee/optee_private.h
358
p->u.value.b = mp->u.value.b;
drivers/tee/optee/optee_private.h
359
p->u.value.c = mp->u.value.c;
drivers/tee/optee/optee_private.h
362
static inline void optee_to_msg_param_value(struct optee_msg_param *mp,
drivers/tee/optee/optee_private.h
365
mp->attr = OPTEE_MSG_ATTR_TYPE_VALUE_INPUT + p->attr -
drivers/tee/optee/optee_private.h
367
mp->u.value.a = p->u.value.a;
drivers/tee/optee/optee_private.h
368
mp->u.value.b = p->u.value.b;
drivers/tee/optee/optee_private.h
369
mp->u.value.c = p->u.value.c;
drivers/tee/optee/smc_abi.c
104
p->u.memref.shm_offs = mp->u.tmem.buf_ptr - pa;
drivers/tee/optee/smc_abi.c
111
const struct optee_msg_param *mp)
drivers/tee/optee/smc_abi.c
117
p->u.memref.size = mp->u.rmem.size;
drivers/tee/optee/smc_abi.c
118
shm = (struct tee_shm *)(unsigned long)mp->u.rmem.shm_ref;
drivers/tee/optee/smc_abi.c
121
p->u.memref.shm_offs = mp->u.rmem.offs;
drivers/tee/optee/smc_abi.c
147
const struct optee_msg_param *mp = msg_params + n;
drivers/tee/optee/smc_abi.c
148
u32 attr = mp->attr & OPTEE_MSG_ATTR_TYPE_MASK;
drivers/tee/optee/smc_abi.c
158
optee_from_msg_param_value(p, attr, mp);
drivers/tee/optee/smc_abi.c
163
rc = from_msg_param_tmp_mem(p, attr, mp);
drivers/tee/optee/smc_abi.c
170
from_msg_param_reg_mem(p, attr, mp);
drivers/tee/optee/smc_abi.c
180
static int to_msg_param_tmp_mem(struct optee_msg_param *mp,
drivers/tee/optee/smc_abi.c
186
mp->attr = OPTEE_MSG_ATTR_TYPE_TMEM_INPUT + p->attr -
drivers/tee/optee/smc_abi.c
189
mp->u.tmem.shm_ref = (unsigned long)p->u.memref.shm;
drivers/tee/optee/smc_abi.c
190
mp->u.tmem.size = p->u.memref.size;
drivers/tee/optee/smc_abi.c
193
mp->u.tmem.buf_ptr = 0;
drivers/tee/optee/smc_abi.c
201
mp->u.tmem.buf_ptr = pa;
drivers/tee/optee/smc_abi.c
202
mp->attr |= OPTEE_MSG_ATTR_CACHE_PREDEFINED <<
drivers/tee/optee/smc_abi.c
208
static int to_msg_param_reg_mem(struct optee_msg_param *mp,
drivers/tee/optee/smc_abi.c
211
mp->attr = OPTEE_MSG_ATTR_TYPE_RMEM_INPUT + p->attr -
drivers/tee/optee/smc_abi.c
214
mp->u.rmem.shm_ref = (unsigned long)p->u.memref.shm;
drivers/tee/optee/smc_abi.c
215
mp->u.rmem.size = p->u.memref.size;
drivers/tee/optee/smc_abi.c
216
mp->u.rmem.offs = p->u.memref.shm_offs;
drivers/tee/optee/smc_abi.c
237
struct optee_msg_param *mp = msg_params + n;
drivers/tee/optee/smc_abi.c
241
mp->attr = TEE_IOCTL_PARAM_ATTR_TYPE_NONE;
drivers/tee/optee/smc_abi.c
242
memset(&mp->u, 0, sizeof(mp->u));
drivers/tee/optee/smc_abi.c
247
optee_to_msg_param_value(mp, p);
drivers/tee/optee/smc_abi.c
253
rc = to_msg_param_reg_mem(mp, p);
drivers/tee/optee/smc_abi.c
255
rc = to_msg_param_tmp_mem(mp, p);
drivers/tee/optee/smc_abi.c
84
const struct optee_msg_param *mp)
drivers/tee/optee/smc_abi.c
92
p->u.memref.size = mp->u.tmem.size;
drivers/tee/optee/smc_abi.c
93
shm = (struct tee_shm *)(unsigned long)mp->u.tmem.shm_ref;
drivers/tty/mxser.c
457
struct mxser_port *mp = container_of(port, struct mxser_port, port);
drivers/tty/mxser.c
459
return inb(mp->ioaddr + UART_MSR) & UART_MSR_DCD;
drivers/tty/mxser.c
464
struct mxser_port *mp = container_of(port, struct mxser_port, port);
drivers/tty/mxser.c
467
guard(spinlock_irqsave)(&mp->slock);
drivers/tty/mxser.c
468
mcr = inb(mp->ioaddr + UART_MCR);
drivers/tty/mxser.c
473
outb(mcr, mp->ioaddr + UART_MCR);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1040
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1042
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1044
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1065
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1071
if (mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1076
mp->enabled = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1091
mp->updating = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1107
mp->enabled = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1115
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1120
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1130
mp->updating = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1131
mp->enabled = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1142
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1152
mp->user_info = *info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1153
mp->user_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1163
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1168
*info = mp->user_info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1208
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1221
if (mp->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1246
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1248
mp->timings = *timings;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1249
mp->extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1256
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1260
if (mp->updating) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1274
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1276
mp->lcd_config = *config;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1277
mp->extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1284
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1288
if (mp->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
132
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1548
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1550
if (mp->framedone_handler)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1553
mp->framedone_handler = handler;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1554
mp->framedone_handler_data = data;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1562
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1564
WARN_ON(mp->framedone_handler != handler ||
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1565
mp->framedone_handler_data != data);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1567
mp->framedone_handler = NULL;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
1568
mp->framedone_handler_data = NULL;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
170
mp = &dss_data.mgr_priv_data_array[OMAP_DSS_CHANNEL_DIGIT];
drivers/video/fbdev/omap2/omapfb/dss/apply.c
172
mp->lcd_config.video_port_width = 24;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
173
mp->lcd_config.clock_info.lck_div = 1;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
174
mp->lcd_config.clock_info.pck_div = 1;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
184
struct mgr_priv_data *mp = get_mgr_priv(ovl->manager);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
186
return mp->lcd_config.stallmode;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
191
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
193
return mp->lcd_config.stallmode;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
204
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
206
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
208
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
211
if (applying && mp->user_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
212
mi = &mp->user_info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
214
mi = &mp->info;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
230
return dss_mgr_check(mgr, mi, &mp->timings, &mp->lcd_config, ois);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
257
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
261
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
263
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
268
if (mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
272
if (mp->busy)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
276
if (mp->info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
280
if (mp->shadow_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
290
if (mp->extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
294
if (mp->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
336
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
339
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
341
if (mp->shadow_info_dirty || mp->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
362
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
365
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
367
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
370
if (!mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
373
if (mp->extra_info_dirty || mp->shadow_extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
471
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
484
if (!mp->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
502
dirty = mp->info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
503
shadow_dirty = mp->shadow_info_dirty;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
542
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
551
mp = get_mgr_priv(ovl->manager);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
560
if (!mp->enabled) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
620
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
630
mp = get_mgr_priv(ovl->manager);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
632
replication = dss_ovl_use_replication(mp->lcd_config, oi->color_mode);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
634
r = dispc_ovl_setup(ovl->id, oi, replication, &mp->timings, false);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
649
if (mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
656
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
669
mp = get_mgr_priv(ovl->manager);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
672
if (mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
678
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
683
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
686
WARN_ON(mp->busy);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
694
if (mp->info_dirty) {
drivers/video/fbdev/omap2/omapfb/dss/apply.c
695
dispc_mgr_setup(mgr->id, &mp->info);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
697
mp->info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
698
if (mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
699
mp->shadow_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
705
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
709
if (!mp->extra_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
712
dispc_mgr_set_timings(mgr->id, &mp->timings);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
716
dispc_mgr_set_lcd_config(mgr->id, &mp->lcd_config);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
718
mp->extra_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
719
if (mp->updating)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
720
mp->shadow_extra_info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
730
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
734
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
736
if (!mp->enabled || mgr_manual_update(mgr) || mp->busy)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
758
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
761
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
763
if (!mp->enabled || mgr_manual_update(mgr) || mp->busy)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
769
mp->busy = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
782
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
785
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
786
mp->shadow_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
787
mp->shadow_extra_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
810
struct mgr_priv_data *mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
816
WARN_ON(mp->updating);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
828
mp->updating = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
889
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
892
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
894
if (!mp->enabled)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
897
mp->updating = dispc_mgr_is_enabled(i);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
900
bool was_busy = mp->busy;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
901
mp->busy = dispc_mgr_go_busy(i);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
903
if (was_busy && !mp->busy)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
918
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
921
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
923
if (!mgr_manual_update(mgr) || !mp->framedone_handler)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
927
mp->framedone_handler(mp->framedone_handler_data);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
952
struct mgr_priv_data *mp;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
954
mp = get_mgr_priv(mgr);
drivers/video/fbdev/omap2/omapfb/dss/apply.c
956
if (!mp->user_info_dirty)
drivers/video/fbdev/omap2/omapfb/dss/apply.c
959
mp->user_info_dirty = false;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
960
mp->info_dirty = true;
drivers/video/fbdev/omap2/omapfb/dss/apply.c
961
mp->info = mp->user_info;
drivers/watchdog/cpwd.c
572
struct miscdevice *mp = &p->devs[i].misc;
drivers/watchdog/cpwd.c
574
mp->minor = WD0_MINOR + i;
drivers/watchdog/cpwd.c
575
mp->name = cpwd_names[i];
drivers/watchdog/cpwd.c
576
mp->fops = &cpwd_fops;
fs/autofs/waitq.c
105
struct autofs_packet_missing *mp = &pkt.v4_pkt.missing;
fs/autofs/waitq.c
107
pktsz = sizeof(*mp);
fs/autofs/waitq.c
109
mp->wait_queue_token = wq->wait_queue_token;
fs/autofs/waitq.c
110
mp->len = wq->name.len;
fs/autofs/waitq.c
111
memcpy(mp->name, wq->name.name, wq->name.len);
fs/autofs/waitq.c
112
mp->name[wq->name.len] = '\0';
fs/gfs2/bmap.c
1017
struct metapath *mp)
fs/gfs2/bmap.c
1062
release_metapath(mp);
fs/gfs2/bmap.c
1064
iomap->length, flags, iomap, mp);
fs/gfs2/bmap.c
1070
ret = __gfs2_iomap_alloc(inode, iomap, mp);
fs/gfs2/bmap.c
1102
struct metapath mp = { .mp_aheight = 1, };
fs/gfs2/bmap.c
1109
ret = __gfs2_iomap_get(inode, pos, length, flags, iomap, &mp);
fs/gfs2/bmap.c
1133
ret = gfs2_iomap_begin_write(inode, pos, length, flags, iomap, &mp);
fs/gfs2/bmap.c
1139
iomap->private = metapath_dibh(&mp);
fs/gfs2/bmap.c
1144
release_metapath(&mp);
fs/gfs2/bmap.c
1428
struct metapath mp = { .mp_aheight = 1, };
fs/gfs2/bmap.c
1431
ret = __gfs2_iomap_get(inode, pos, length, 0, iomap, &mp);
fs/gfs2/bmap.c
1432
release_metapath(&mp);
fs/gfs2/bmap.c
1439
struct metapath mp = { .mp_aheight = 1, };
fs/gfs2/bmap.c
1442
ret = __gfs2_iomap_get(inode, pos, length, IOMAP_WRITE, iomap, &mp);
fs/gfs2/bmap.c
1444
ret = __gfs2_iomap_alloc(inode, iomap, &mp);
fs/gfs2/bmap.c
1445
release_metapath(&mp);
fs/gfs2/bmap.c
1623
static bool mp_eq_to_hgt(struct metapath *mp, __u16 *list, unsigned int h)
fs/gfs2/bmap.c
1625
if (memcmp(mp->mp_list, list, h * sizeof(mp->mp_list[0])))
fs/gfs2/bmap.c
1642
static bool find_nonnull_ptr(struct gfs2_sbd *sdp, struct metapath *mp,
fs/gfs2/bmap.c
1646
struct buffer_head *bh = mp->mp_bh[h];
fs/gfs2/bmap.c
1649
first = metaptr1(h, mp);
fs/gfs2/bmap.c
1650
ptr = first + mp->mp_list[h];
fs/gfs2/bmap.c
1652
if (end_list && mp_eq_to_hgt(mp, end_list, h)) {
fs/gfs2/bmap.c
1659
mp->mp_list[h] = ptr - first;
fs/gfs2/bmap.c
1662
mp->mp_list[h] = 0;
fs/gfs2/bmap.c
1678
metapointer_range(struct metapath *mp, int height,
fs/gfs2/bmap.c
1683
struct buffer_head *bh = mp->mp_bh[height];
fs/gfs2/bmap.c
1686
first = metaptr1(height, mp);
fs/gfs2/bmap.c
1688
if (mp_eq_to_hgt(mp, start_list, height)) {
fs/gfs2/bmap.c
1693
if (end_list && mp_eq_to_hgt(mp, end_list, height)) {
fs/gfs2/bmap.c
1700
struct metapath *mp, int height,
fs/gfs2/bmap.c
1707
if (!mp_eq_to_hgt(mp, end_list, height))
fs/gfs2/bmap.c
1712
return mp->mp_list[height] >= end;
fs/gfs2/bmap.c
1735
struct metapath mp = {};
fs/gfs2/bmap.c
1787
find_metapath(sdp, lend, &mp, ip->i_height);
fs/gfs2/bmap.c
1789
memcpy(end_list, mp.mp_list, sizeof(mp.mp_list));
fs/gfs2/bmap.c
1798
find_metapath(sdp, lblock, &mp, ip->i_height);
fs/gfs2/bmap.c
1799
memcpy(start_list, mp.mp_list, sizeof(start_list));
fs/gfs2/bmap.c
1811
mp.mp_bh[0] = dibh;
fs/gfs2/bmap.c
1812
ret = lookup_metapath(ip, &mp);
fs/gfs2/bmap.c
1817
for (mp_h = 0; mp_h < mp.mp_aheight - 1; mp_h++) {
fs/gfs2/bmap.c
1818
metapointer_range(&mp, mp_h, start_list, start_aligned,
fs/gfs2/bmap.c
1823
if (mp.mp_aheight == ip->i_height)
fs/gfs2/bmap.c
1844
bh = mp.mp_bh[mp_h];
fs/gfs2/bmap.c
1868
metapointer_range(&mp, mp_h, start_list, start_aligned,
fs/gfs2/bmap.c
1871
ret = sweep_bh_for_rgrps(ip, &rd_gh, mp.mp_bh[mp_h],
fs/gfs2/bmap.c
1891
brelse(mp.mp_bh[mp_h]);
fs/gfs2/bmap.c
1892
mp.mp_bh[mp_h] = NULL;
fs/gfs2/bmap.c
1899
memcpy(mp.mp_list, start_list, sizeof(start_list));
fs/gfs2/bmap.c
1904
mp.mp_list[mp_h] = 0;
fs/gfs2/bmap.c
1906
mp.mp_list[mp_h]++;
fs/gfs2/bmap.c
1907
if (walk_done(sdp, &mp, mp_h, end_list, end_aligned))
fs/gfs2/bmap.c
1912
if (find_nonnull_ptr(sdp, &mp, mp_h, end_list, end_aligned)) {
fs/gfs2/bmap.c
1923
ret = fillup_metapath(ip, &mp, mp_h);
fs/gfs2/bmap.c
1928
if (mp.mp_aheight > 1 && strip_h == ip->i_height - 1) {
fs/gfs2/bmap.c
1929
unsigned int height = mp.mp_aheight - 1;
fs/gfs2/bmap.c
1932
if (mp.mp_aheight - 1 == strip_h)
fs/gfs2/bmap.c
1935
for (; height >= mp.mp_aheight - ret; height--) {
fs/gfs2/bmap.c
1936
metapointer_range(&mp, height,
fs/gfs2/bmap.c
1945
if (mp.mp_aheight - 1 == strip_h) {
fs/gfs2/bmap.c
1949
if (mp.mp_aheight < ip->i_height) /* We have a partial height */
fs/gfs2/bmap.c
1950
mp_h = mp.mp_aheight - 1;
fs/gfs2/bmap.c
1955
if (find_nonnull_ptr(sdp, &mp, mp_h, end_list, end_aligned))
fs/gfs2/bmap.c
1991
release_metapath(&mp);
fs/gfs2/bmap.c
234
struct metapath *mp, unsigned int height)
fs/gfs2/bmap.c
238
mp->mp_fheight = height;
fs/gfs2/bmap.c
240
mp->mp_list[i] = do_div(block, sdp->sd_inptrs);
fs/gfs2/bmap.c
243
static inline unsigned int metapath_branch_start(const struct metapath *mp)
fs/gfs2/bmap.c
245
if (mp->mp_list[0] == 0)
fs/gfs2/bmap.c
255
static inline __be64 *metaptr1(unsigned int height, const struct metapath *mp)
fs/gfs2/bmap.c
257
struct buffer_head *bh = mp->mp_bh[height];
fs/gfs2/bmap.c
273
static inline __be64 *metapointer(unsigned int height, const struct metapath *mp)
fs/gfs2/bmap.c
275
__be64 *p = metaptr1(height, mp);
fs/gfs2/bmap.c
276
return p + mp->mp_list[height];
fs/gfs2/bmap.c
279
static inline const __be64 *metaend(unsigned int height, const struct metapath *mp)
fs/gfs2/bmap.c
281
const struct buffer_head *bh = mp->mp_bh[height];
fs/gfs2/bmap.c
285
static void clone_metapath(struct metapath *clone, struct metapath *mp)
fs/gfs2/bmap.c
289
*clone = *mp;
fs/gfs2/bmap.c
290
for (hgt = 0; hgt < mp->mp_aheight; hgt++)
fs/gfs2/bmap.c
319
metapath_dibh(struct metapath *mp)
fs/gfs2/bmap.c
321
return mp->mp_bh[0];
fs/gfs2/bmap.c
324
static int __fillup_metapath(struct gfs2_inode *ip, struct metapath *mp,
fs/gfs2/bmap.c
328
__be64 *ptr = metapointer(x, mp);
fs/gfs2/bmap.c
334
ret = gfs2_meta_buffer(ip, GFS2_METATYPE_IN, dblock, &mp->mp_bh[x + 1]);
fs/gfs2/bmap.c
338
mp->mp_aheight = x + 1;
fs/gfs2/bmap.c
359
static int lookup_metapath(struct gfs2_inode *ip, struct metapath *mp)
fs/gfs2/bmap.c
361
return __fillup_metapath(ip, mp, 0, ip->i_height - 1);
fs/gfs2/bmap.c
375
static int fillup_metapath(struct gfs2_inode *ip, struct metapath *mp, int h)
fs/gfs2/bmap.c
383
if (mp->mp_bh[x])
fs/gfs2/bmap.c
387
ret = __fillup_metapath(ip, mp, x, h);
fs/gfs2/bmap.c
390
return mp->mp_aheight - x - 1;
fs/gfs2/bmap.c
393
static sector_t metapath_to_block(struct gfs2_sbd *sdp, struct metapath *mp)
fs/gfs2/bmap.c
398
for (hgt = mp->mp_fheight - 1; hgt >= 0; hgt--) {
fs/gfs2/bmap.c
399
if (hgt < mp->mp_aheight)
fs/gfs2/bmap.c
400
block += mp->mp_list[hgt] * factor;
fs/gfs2/bmap.c
406
static void release_metapath(struct metapath *mp)
fs/gfs2/bmap.c
411
if (mp->mp_bh[i] == NULL)
fs/gfs2/bmap.c
413
brelse(mp->mp_bh[i]);
fs/gfs2/bmap.c
414
mp->mp_bh[i] = NULL;
fs/gfs2/bmap.c
483
for (hgt = mp->mp_fheight - 1; hgt >= mp->mp_aheight; hgt--) {
fs/gfs2/bmap.c
484
max_len += mp->mp_list[hgt] * factor;
fs/gfs2/bmap.c
485
mp->mp_list[hgt] = 0;
fs/gfs2/bmap.c
490
u16 start = mp->mp_list[hgt];
fs/gfs2/bmap.c
500
status = walker(mp, ptrs);
fs/gfs2/bmap.c
505
BUG_ON(mp->mp_aheight == mp->mp_fheight);
fs/gfs2/bmap.c
506
ptrs = mp->mp_list[hgt] - start;
fs/gfs2/bmap.c
520
brelse(mp->mp_bh[hgt]);
fs/gfs2/bmap.c
521
mp->mp_bh[hgt] = NULL;
fs/gfs2/bmap.c
522
mp->mp_list[hgt] = 0;
fs/gfs2/bmap.c
529
(mp->mp_list[hgt])++;
fs/gfs2/bmap.c
531
if (mp->mp_list[hgt] >= sdp->sd_inptrs)
fs/gfs2/bmap.c
534
if (mp->mp_list[hgt] >= sdp->sd_diptrs)
fs/gfs2/bmap.c
540
ret = fillup_metapath(ip, mp, ip->i_height - 1);
fs/gfs2/bmap.c
546
mp->mp_aheight = hgt + 1;
fs/gfs2/bmap.c
551
static enum walker_status gfs2_hole_walker(struct metapath *mp,
fs/gfs2/bmap.c
557
hgt = mp->mp_aheight - 1;
fs/gfs2/bmap.c
558
start = metapointer(hgt, mp);
fs/gfs2/bmap.c
563
mp->mp_list[hgt] += ptr - start;
fs/gfs2/bmap.c
564
if (mp->mp_aheight == mp->mp_fheight)
fs/gfs2/bmap.c
585
struct metapath *mp, struct iomap *iomap)
fs/gfs2/bmap.c
591
clone_metapath(&clone, mp);
fs/gfs2/bmap.c
608
static inline void gfs2_indirect_init(struct metapath *mp,
fs/gfs2/bmap.c
612
__be64 *ptr = (__be64 *)(mp->mp_bh[i - 1]->b_data +
fs/gfs2/bmap.c
616
BUG_ON(mp->mp_bh[i] != NULL);
fs/gfs2/bmap.c
617
mp->mp_bh[i] = gfs2_meta_new(gl, bn);
fs/gfs2/bmap.c
618
gfs2_trans_add_meta(gl, mp->mp_bh[i]);
fs/gfs2/bmap.c
619
gfs2_metatype_set(mp->mp_bh[i], GFS2_METATYPE_IN, GFS2_FORMAT_IN);
fs/gfs2/bmap.c
620
gfs2_buffer_clear_tail(mp->mp_bh[i], sizeof(struct gfs2_meta_header));
fs/gfs2/bmap.c
662
struct metapath *mp)
fs/gfs2/bmap.c
666
struct buffer_head *dibh = metapath_dibh(mp);
fs/gfs2/bmap.c
670
const unsigned end_of_metadata = mp->mp_fheight - 1;
fs/gfs2/bmap.c
676
BUG_ON(mp->mp_aheight < 1);
fs/gfs2/bmap.c
684
if (mp->mp_fheight == mp->mp_aheight) {
fs/gfs2/bmap.c
689
if (mp->mp_fheight == ip->i_height) {
fs/gfs2/bmap.c
691
iblks = mp->mp_fheight - mp->mp_aheight;
fs/gfs2/bmap.c
696
iblks = mp->mp_fheight - ip->i_height;
fs/gfs2/bmap.c
697
branch_start = metapath_branch_start(mp);
fs/gfs2/bmap.c
698
iblks += (mp->mp_fheight - branch_start);
fs/gfs2/bmap.c
705
i = mp->mp_aheight;
fs/gfs2/bmap.c
722
for (; i - 1 < mp->mp_fheight - ip->i_height && n > 0;
fs/gfs2/bmap.c
724
gfs2_indirect_init(mp, ip->i_gl, i, 0, bn++);
fs/gfs2/bmap.c
725
if (i - 1 == mp->mp_fheight - ip->i_height) {
fs/gfs2/bmap.c
727
gfs2_buffer_copy_tail(mp->mp_bh[i],
fs/gfs2/bmap.c
733
ptr = (__be64 *)(mp->mp_bh[i]->b_data +
fs/gfs2/bmap.c
737
for(i = branch_start; i < mp->mp_fheight; i++) {
fs/gfs2/bmap.c
738
if (mp->mp_bh[i] == NULL)
fs/gfs2/bmap.c
740
brelse(mp->mp_bh[i]);
fs/gfs2/bmap.c
741
mp->mp_bh[i] = NULL;
fs/gfs2/bmap.c
749
if (i > 1 && i < mp->mp_fheight)
fs/gfs2/bmap.c
750
gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[i-1]);
fs/gfs2/bmap.c
751
for (; i < mp->mp_fheight && n > 0; i++, n--)
fs/gfs2/bmap.c
752
gfs2_indirect_init(mp, ip->i_gl, i,
fs/gfs2/bmap.c
753
mp->mp_list[i-1], bn++);
fs/gfs2/bmap.c
754
if (i == mp->mp_fheight)
fs/gfs2/bmap.c
761
BUG_ON(mp->mp_bh[end_of_metadata] == NULL);
fs/gfs2/bmap.c
762
gfs2_trans_add_meta(ip->i_gl, mp->mp_bh[end_of_metadata]);
fs/gfs2/bmap.c
764
ptr = metapointer(end_of_metadata, mp);
fs/gfs2/bmap.c
775
ip->i_height = mp->mp_fheight;
fs/gfs2/bmap.c
795
static u64 gfs2_alloc_size(struct inode *inode, struct metapath *mp, u64 size)
fs/gfs2/bmap.c
809
if (gfs2_is_stuffed(ip) || mp->mp_fheight != mp->mp_aheight) {
fs/gfs2/bmap.c
810
unsigned int maxsize = mp->mp_fheight > 1 ?
fs/gfs2/bmap.c
812
maxsize -= mp->mp_list[mp->mp_fheight - 1];
fs/gfs2/bmap.c
818
first = metapointer(ip->i_height - 1, mp);
fs/gfs2/bmap.c
819
end = metaend(ip->i_height - 1, mp);
fs/gfs2/bmap.c
842
struct metapath *mp)
fs/gfs2/bmap.c
864
mp->mp_bh[0] = dibh;
fs/gfs2/bmap.c
903
find_metapath(sdp, lblock, mp, height);
fs/gfs2/bmap.c
907
ret = lookup_metapath(ip, mp);
fs/gfs2/bmap.c
911
if (mp->mp_aheight != ip->i_height)
fs/gfs2/bmap.c
914
ptr = metapointer(ip->i_height - 1, mp);
fs/gfs2/bmap.c
918
bh = mp->mp_bh[ip->i_height - 1];
fs/gfs2/bmap.c
939
ret = gfs2_hole_size(inode, lblock, len, mp, iomap);
fs/gfs2/bmap.c
948
len = gfs2_alloc_size(inode, mp, len);
fs/gfs2/bmap.c
954
ret = gfs2_hole_size(inode, lblock, len, mp, iomap);
fs/jfs/jfs_btree.h
102
struct metapage *mp; /* 4/8: */
fs/jfs/jfs_btree.h
145
MP = (LEAF)->mp;\
fs/jfs/jfs_btree.h
156
if (! BT_IS_ROOT((BTSTACK)->top->mp))\
fs/jfs/jfs_btree.h
157
release_metapage((BTSTACK)->top->mp);\
fs/jfs/jfs_dmap.c
1026
mp = read_metapage(ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1027
if (mp == NULL) {
fs/jfs/jfs_dmap.c
1032
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
1043
write_metapage(mp);
fs/jfs/jfs_dmap.c
1046
release_metapage(mp);
fs/jfs/jfs_dmap.c
1314
struct metapage *mp;
fs/jfs/jfs_dmap.c
1369
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1370
if (mp == NULL)
fs/jfs/jfs_dmap.c
1372
dcp = (struct dmapctl *) mp->data;
fs/jfs/jfs_dmap.c
1377
release_metapage(mp);
fs/jfs/jfs_dmap.c
1394
release_metapage(mp);
fs/jfs/jfs_dmap.c
1424
release_metapage(mp);
fs/jfs/jfs_dmap.c
1445
release_metapage(mp);
fs/jfs/jfs_dmap.c
1484
release_metapage(mp);
fs/jfs/jfs_dmap.c
157
struct metapage *mp;
fs/jfs/jfs_dmap.c
1687
struct metapage *mp;
fs/jfs/jfs_dmap.c
169
mp = read_metapage(ipbmap,
fs/jfs/jfs_dmap.c
1699
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1700
if (mp == NULL)
fs/jfs/jfs_dmap.c
1702
dcp = (struct dmapctl *) mp->data;
fs/jfs/jfs_dmap.c
1708
release_metapage(mp);
fs/jfs/jfs_dmap.c
172
if (mp == NULL) {
fs/jfs/jfs_dmap.c
1721
release_metapage(mp);
fs/jfs/jfs_dmap.c
178
dbmp_le = (struct dbmap_disk *) mp->data;
fs/jfs/jfs_dmap.c
1804
struct metapage *mp;
fs/jfs/jfs_dmap.c
1813
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1814
if (mp == NULL)
fs/jfs/jfs_dmap.c
1816
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
1819
release_metapage(mp);
fs/jfs/jfs_dmap.c
1827
mark_metapage_dirty(mp);
fs/jfs/jfs_dmap.c
1829
release_metapage(mp);
fs/jfs/jfs_dmap.c
1845
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1846
if (mp == NULL) {
fs/jfs/jfs_dmap.c
1850
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
1855
release_metapage(mp);
fs/jfs/jfs_dmap.c
1869
release_metapage(mp);
fs/jfs/jfs_dmap.c
1875
write_metapage(mp);
fs/jfs/jfs_dmap.c
1898
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
1899
if (mp == NULL) {
fs/jfs/jfs_dmap.c
1907
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
1915
release_metapage(mp);
fs/jfs/jfs_dmap.c
1922
write_metapage(mp);
fs/jfs/jfs_dmap.c
215
release_metapage(mp);
fs/jfs/jfs_dmap.c
231
release_metapage(mp);
fs/jfs/jfs_dmap.c
2472
struct metapage *mp;
fs/jfs/jfs_dmap.c
2483
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
2484
if (mp == NULL)
fs/jfs/jfs_dmap.c
2486
dcp = (struct dmapctl *) mp->data;
fs/jfs/jfs_dmap.c
2490
release_metapage(mp);
fs/jfs/jfs_dmap.c
2528
release_metapage(mp);
fs/jfs/jfs_dmap.c
2537
release_metapage(mp);
fs/jfs/jfs_dmap.c
2583
release_metapage(mp);
fs/jfs/jfs_dmap.c
2602
write_metapage(mp);
fs/jfs/jfs_dmap.c
280
struct metapage *mp;
fs/jfs/jfs_dmap.c
287
mp = read_metapage(ipbmap,
fs/jfs/jfs_dmap.c
290
if (mp == NULL) {
fs/jfs/jfs_dmap.c
295
dbmp_le = (struct dbmap_disk *) mp->data;
fs/jfs/jfs_dmap.c
314
write_metapage(mp);
fs/jfs/jfs_dmap.c
3193
struct metapage *mp;
fs/jfs/jfs_dmap.c
3208
mp = NULL;
fs/jfs/jfs_dmap.c
3211
if (mp) {
fs/jfs/jfs_dmap.c
3212
write_metapage(mp);
fs/jfs/jfs_dmap.c
3217
mp = read_metapage(ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
3218
if (mp == NULL) {
fs/jfs/jfs_dmap.c
3222
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
3231
release_metapage(mp);
fs/jfs/jfs_dmap.c
3238
write_metapage(mp);
fs/jfs/jfs_dmap.c
3374
struct metapage *mp, *l2mp, *l1mp = NULL, *l0mp = NULL;
fs/jfs/jfs_dmap.c
346
struct metapage *mp;
fs/jfs/jfs_dmap.c
3538
mp = read_metapage(ipbmap, p,
fs/jfs/jfs_dmap.c
3540
if (mp == NULL)
fs/jfs/jfs_dmap.c
3545
mp = read_metapage(ipbmap, p,
fs/jfs/jfs_dmap.c
3547
if (mp == NULL)
fs/jfs/jfs_dmap.c
3553
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
3560
write_metapage(mp);
fs/jfs/jfs_dmap.c
376
mp = NULL;
fs/jfs/jfs_dmap.c
379
if (mp) {
fs/jfs/jfs_dmap.c
380
write_metapage(mp);
fs/jfs/jfs_dmap.c
385
mp = read_metapage(ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
386
if (mp == NULL) {
fs/jfs/jfs_dmap.c
390
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
400
release_metapage(mp);
fs/jfs/jfs_dmap.c
407
if (mp)
fs/jfs/jfs_dmap.c
408
write_metapage(mp);
fs/jfs/jfs_dmap.c
447
struct metapage *mp;
fs/jfs/jfs_dmap.c
469
mp = NULL;
fs/jfs/jfs_dmap.c
475
if (mp) {
fs/jfs/jfs_dmap.c
476
write_metapage(mp);
fs/jfs/jfs_dmap.c
479
mp = read_metapage(bmp->db_ipbmap, lblkno, PSIZE,
fs/jfs/jfs_dmap.c
481
if (mp == NULL)
fs/jfs/jfs_dmap.c
483
metapage_wait_for_io(mp);
fs/jfs/jfs_dmap.c
485
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
556
if (mp->lsn != 0) {
fs/jfs/jfs_dmap.c
558
logdiff(diffp, mp->lsn, log);
fs/jfs/jfs_dmap.c
560
mp->lsn = lsn;
fs/jfs/jfs_dmap.c
563
list_move(&mp->synclist, &tblk->synclist);
fs/jfs/jfs_dmap.c
568
logdiff(diffp, mp->clsn, log);
fs/jfs/jfs_dmap.c
570
mp->clsn = tblk->clsn;
fs/jfs/jfs_dmap.c
572
mp->log = log;
fs/jfs/jfs_dmap.c
573
mp->lsn = lsn;
fs/jfs/jfs_dmap.c
577
list_add(&mp->synclist, &tblk->synclist);
fs/jfs/jfs_dmap.c
579
mp->clsn = tblk->clsn;
fs/jfs/jfs_dmap.c
585
if (mp) {
fs/jfs/jfs_dmap.c
586
write_metapage(mp);
fs/jfs/jfs_dmap.c
719
struct metapage *mp;
fs/jfs/jfs_dmap.c
796
mp = read_metapage(ipbmap, lblkno, PSIZE, 0);
fs/jfs/jfs_dmap.c
797
if (mp == NULL)
fs/jfs/jfs_dmap.c
800
dp = (struct dmap *) mp->data;
fs/jfs/jfs_dmap.c
809
mark_metapage_dirty(mp);
fs/jfs/jfs_dmap.c
812
release_metapage(mp);
fs/jfs/jfs_dmap.c
823
release_metapage(mp);
fs/jfs/jfs_dmap.c
835
mark_metapage_dirty(mp);
fs/jfs/jfs_dmap.c
837
release_metapage(mp);
fs/jfs/jfs_dmap.c
847
mark_metapage_dirty(mp);
fs/jfs/jfs_dmap.c
849
release_metapage(mp);
fs/jfs/jfs_dmap.c
853
release_metapage(mp);
fs/jfs/jfs_dmap.c
977
struct metapage *mp;
fs/jfs/jfs_dtree.c
103
struct metapage *mp;
fs/jfs/jfs_dtree.c
1230
split->mp = smp;
fs/jfs/jfs_dtree.c
1335
struct metapage *mp;
fs/jfs/jfs_dtree.c
1353
smp = split->mp;
fs/jfs/jfs_dtree.c
1471
DT_GETPAGE(ip, nextbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
1477
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
1481
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
1483
tlck, ip, mp);
fs/jfs/jfs_dtree.c
1494
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
1580
mp = NULL;
fs/jfs/jfs_dtree.c
1585
rbn, n, &mp, &lblock);
fs/jfs/jfs_dtree.c
1587
if (mp)
fs/jfs/jfs_dtree.c
1588
release_metapage(mp);
fs/jfs/jfs_dtree.c
1643
struct metapage *smp, *pmp, *mp;
fs/jfs/jfs_dtree.c
1664
smp = split->mp;
fs/jfs/jfs_dtree.c
1704
mp = NULL;
fs/jfs/jfs_dtree.c
1711
xaddr, n, &mp, &lblock);
fs/jfs/jfs_dtree.c
1713
if (mp)
fs/jfs/jfs_dtree.c
1714
release_metapage(mp);
fs/jfs/jfs_dtree.c
1895
smp = split->mp;
fs/jfs/jfs_dtree.c
1993
struct metapage *mp = NULL;
fs/jfs/jfs_dtree.c
2000
rbn, n, &mp, &lblock);
fs/jfs/jfs_dtree.c
2002
if (mp)
fs/jfs/jfs_dtree.c
2003
release_metapage(mp);
fs/jfs/jfs_dtree.c
2083
struct metapage *mp, *imp;
fs/jfs/jfs_dtree.c
2106
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
2152
rc = dtDeleteUp(tid, ip, mp, p, &btstack);
fs/jfs/jfs_dtree.c
2160
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2164
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
2216
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2237
struct metapage *mp;
fs/jfs/jfs_dtree.c
2304
DT_GETPAGE(ip, parent->bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2335
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2358
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2368
discard_metapage(mp);
fs/jfs/jfs_dtree.c
238
struct metapage ** mp, s64 *lblock)
fs/jfs/jfs_dtree.c
2380
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2386
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
2422
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2448
struct metapage *mp;
fs/jfs/jfs_dtree.c
2459
DT_GETPAGE(ip, nextbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2463
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2469
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
2471
tlck, ip, mp);
fs/jfs/jfs_dtree.c
2483
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2488
DT_GETPAGE(ip, prevbn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2492
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
2498
tlck = txLock(tid, ip, mp, tlckDTREE | tlckRELINK);
fs/jfs/jfs_dtree.c
2500
tlck, ip, mp);
fs/jfs/jfs_dtree.c
2512
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2623
struct metapage *mp;
fs/jfs/jfs_dtree.c
2632
DT_GETPAGE(inode, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
2638
BT_MARK_DIRTY(mp, inode);
fs/jfs/jfs_dtree.c
264
*mp = NULL;
fs/jfs/jfs_dtree.c
2642
tlck = txLock(tid, inode, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
2643
if (BT_IS_ROOT(mp))
fs/jfs/jfs_dtree.c
2655
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2673
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
272
if (*mp && (*lblock != blkno)) {
fs/jfs/jfs_dtree.c
2726
struct metapage *mp;
fs/jfs/jfs_dtree.c
273
release_metapage(*mp);
fs/jfs/jfs_dtree.c
274
*mp = NULL;
fs/jfs/jfs_dtree.c
276
if (!(*mp)) {
fs/jfs/jfs_dtree.c
278
*mp = read_index_page(ip, blkno);
fs/jfs/jfs_dtree.c
2797
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
280
if (!(*mp)) {
fs/jfs/jfs_dtree.c
2804
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2835
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
286
(struct dir_table_slot *) ((char *) (*mp)->data +
fs/jfs/jfs_dtree.c
2881
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
2892
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
2910
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
292
static inline void lock_index(tid_t tid, struct inode *ip, struct metapage * mp,
fs/jfs/jfs_dtree.c
299
tlck = txLock(tid, ip, mp, tlckDATA);
fs/jfs/jfs_dtree.c
3018
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3041
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3065
struct metapage *mp;
fs/jfs/jfs_dtree.c
3079
DT_GETPAGE(ip, bn, mp, psize, p, rc);
fs/jfs/jfs_dtree.c
3091
btsp->mp = mp;
fs/jfs/jfs_dtree.c
3100
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3112
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3124
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3149
struct metapage *mp;
fs/jfs/jfs_dtree.c
3164
DT_GETSEARCH(ip, btstack->top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
3202
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3217
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3223
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3235
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3244
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3258
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3264
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3276
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
3285
DT_GETPAGE(ip, bn, mp, PSIZE, p, rc);
fs/jfs/jfs_dtree.c
3299
btsp->mp = mp;
fs/jfs/jfs_dtree.c
332
struct metapage *mp;
fs/jfs/jfs_dtree.c
3645
struct metapage *mp = NULL;
fs/jfs/jfs_dtree.c
3761
mp = NULL;
fs/jfs/jfs_dtree.c
3766
&mp, &lblock);
fs/jfs/jfs_dtree.c
3768
if (mp)
fs/jfs/jfs_dtree.c
3769
release_metapage(mp);
fs/jfs/jfs_dtree.c
405
mp = get_index_page(ip, 0);
fs/jfs/jfs_dtree.c
406
if (!mp) {
fs/jfs/jfs_dtree.c
413
tlck = txLock(tid, ip, mp, tlckDATA);
fs/jfs/jfs_dtree.c
422
memcpy(mp->data, temp_table, sizeof(temp_table));
fs/jfs/jfs_dtree.c
424
mark_metapage_dirty(mp);
fs/jfs/jfs_dtree.c
4249
struct metapage *mp;
fs/jfs/jfs_dtree.c
425
release_metapage(mp);
fs/jfs/jfs_dtree.c
4269
DT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
4271
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
4275
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
4296
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
447
if ((mp = get_index_page(ip, blkno)))
fs/jfs/jfs_dtree.c
448
memset(mp->data, 0, PSIZE); /* Just looks better */
fs/jfs/jfs_dtree.c
452
mp = read_index_page(ip, blkno);
fs/jfs/jfs_dtree.c
454
if (!mp) {
fs/jfs/jfs_dtree.c
459
lock_index(tid, ip, mp, index);
fs/jfs/jfs_dtree.c
462
(struct dir_table_slot *) ((char *) mp->data + page_offset);
fs/jfs/jfs_dtree.c
467
mark_metapage_dirty(mp);
fs/jfs/jfs_dtree.c
468
release_metapage(mp);
fs/jfs/jfs_dtree.c
488
struct metapage *mp = NULL;
fs/jfs/jfs_dtree.c
490
dirtab_slot = find_index(ip, index, &mp, &lblock);
fs/jfs/jfs_dtree.c
499
if (mp) {
fs/jfs/jfs_dtree.c
500
lock_index(tid, ip, mp, index);
fs/jfs/jfs_dtree.c
501
mark_metapage_dirty(mp);
fs/jfs/jfs_dtree.c
502
release_metapage(mp);
fs/jfs/jfs_dtree.c
513
int slot, struct metapage ** mp, s64 *lblock)
fs/jfs/jfs_dtree.c
517
dirtab_slot = find_index(ip, index, mp, lblock);
fs/jfs/jfs_dtree.c
525
if (*mp) {
fs/jfs/jfs_dtree.c
526
lock_index(tid, ip, *mp, index);
fs/jfs/jfs_dtree.c
527
mark_metapage_dirty(*mp);
fs/jfs/jfs_dtree.c
541
struct metapage *mp = NULL;
fs/jfs/jfs_dtree.c
544
slot = find_index(ip, index, &mp, &lblock);
fs/jfs/jfs_dtree.c
551
if (mp)
fs/jfs/jfs_dtree.c
552
release_metapage(mp);
fs/jfs/jfs_dtree.c
574
struct metapage *mp;
fs/jfs/jfs_dtree.c
624
DT_GETPAGE(ip, bn, mp, psize, p, rc);
fs/jfs/jfs_dtree.c
701
btsp->mp = mp;
fs/jfs/jfs_dtree.c
750
btsp->mp = mp;
fs/jfs/jfs_dtree.c
789
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
793
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
819
struct metapage *mp; /* meta-page buffer */
fs/jfs/jfs_dtree.c
837
DT_GETSEARCH(ip, btstack->top, bn, mp, p, index);
fs/jfs/jfs_dtree.c
846
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
866
split.mp = mp;
fs/jfs/jfs_dtree.c
880
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_dtree.c
884
tlck = txLock(tid, ip, mp, tlckDTREE | tlckENTRY);
fs/jfs/jfs_dtree.c
909
DT_PUTPAGE(mp);
fs/jfs/jfs_dtree.c
953
smp = split->mp;
fs/jfs/jfs_imap.c
1012
write_metapage(mp);
fs/jfs/jfs_imap.c
111
mp = read_metapage(ipimap,
fs/jfs/jfs_imap.c
114
if (mp == NULL) {
fs/jfs/jfs_imap.c
120
dinom_le = (struct dinomap_disk *) mp->data;
fs/jfs/jfs_imap.c
1241
tlck = txLock(tid, ipimap, mp, tlckINODE | tlckFREE);
fs/jfs/jfs_imap.c
1247
write_metapage(mp);
fs/jfs/jfs_imap.c
1286
release_metapage(mp);
fs/jfs/jfs_imap.c
1329
struct metapage *mp;
fs/jfs/jfs_imap.c
139
release_metapage(mp);
fs/jfs/jfs_imap.c
1392
if ((rc = diIAGRead(imap, iagno, &mp))) {
fs/jfs/jfs_imap.c
1397
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
1440
mark_metapage_dirty(mp);
fs/jfs/jfs_imap.c
1442
release_metapage(mp);
fs/jfs/jfs_imap.c
1501
release_metapage(mp);
fs/jfs/jfs_imap.c
1523
mark_metapage_dirty(mp);
fs/jfs/jfs_imap.c
1525
release_metapage(mp);
fs/jfs/jfs_imap.c
1563
mark_metapage_dirty(mp);
fs/jfs/jfs_imap.c
1565
release_metapage(mp);
fs/jfs/jfs_imap.c
1584
release_metapage(mp);
fs/jfs/jfs_imap.c
1771
struct metapage *mp;
fs/jfs/jfs_imap.c
1784
if ((rc = diIAGRead(imap, iagno, &mp))) {
fs/jfs/jfs_imap.c
1788
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
1795
release_metapage(mp);
fs/jfs/jfs_imap.c
1806
release_metapage(mp);
fs/jfs/jfs_imap.c
1822
release_metapage(mp);
fs/jfs/jfs_imap.c
1833
release_metapage(mp);
fs/jfs/jfs_imap.c
1847
release_metapage(mp);
fs/jfs/jfs_imap.c
1854
write_metapage(mp);
fs/jfs/jfs_imap.c
1895
struct metapage *mp;
fs/jfs/jfs_imap.c
1905
if ((rc = diNewIAG(imap, &iagno, agno, &mp))) {
fs/jfs/jfs_imap.c
1908
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
1918
if ((rc = diIAGRead(imap, iagno, &mp))) {
fs/jfs/jfs_imap.c
1923
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
1930
release_metapage(mp);
fs/jfs/jfs_imap.c
1943
release_metapage(mp);
fs/jfs/jfs_imap.c
1965
write_metapage(mp);
fs/jfs/jfs_imap.c
1973
write_metapage(mp);
fs/jfs/jfs_imap.c
209
struct metapage *mp;
fs/jfs/jfs_imap.c
216
mp = get_metapage(ipimap,
fs/jfs/jfs_imap.c
219
if (mp == NULL) {
fs/jfs/jfs_imap.c
225
dinom_le = (struct dinomap_disk *) mp->data;
fs/jfs/jfs_imap.c
244
write_metapage(mp);
fs/jfs/jfs_imap.c
2447
struct metapage *mp;
fs/jfs/jfs_imap.c
2545
mp = get_metapage(ipimap, blkno, PSIZE, 0);
fs/jfs/jfs_imap.c
2546
if (!mp) {
fs/jfs/jfs_imap.c
2564
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
2584
flush_metapage(mp);
fs/jfs/jfs_imap.c
2620
if ((rc = diIAGRead(imap, iagno, &mp))) {
fs/jfs/jfs_imap.c
2625
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
2633
*mpp = mp;
fs/jfs/jfs_imap.c
2730
struct metapage *mp;
fs/jfs/jfs_imap.c
2748
rc = diIAGRead(imap, iagno, &mp);
fs/jfs/jfs_imap.c
2752
metapage_wait_for_io(mp);
fs/jfs/jfs_imap.c
2753
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
2791
release_metapage(mp);
fs/jfs/jfs_imap.c
2797
release_metapage(mp);
fs/jfs/jfs_imap.c
2811
if (mp->lsn != 0) {
fs/jfs/jfs_imap.c
2814
logdiff(diffp, mp->lsn, log);
fs/jfs/jfs_imap.c
2816
mp->lsn = lsn;
fs/jfs/jfs_imap.c
2818
list_move(&mp->synclist, &tblk->synclist);
fs/jfs/jfs_imap.c
2821
assert(mp->clsn);
fs/jfs/jfs_imap.c
2823
logdiff(diffp, mp->clsn, log);
fs/jfs/jfs_imap.c
2825
mp->clsn = tblk->clsn;
fs/jfs/jfs_imap.c
2827
mp->log = log;
fs/jfs/jfs_imap.c
2828
mp->lsn = lsn;
fs/jfs/jfs_imap.c
2831
list_add(&mp->synclist, &tblk->synclist);
fs/jfs/jfs_imap.c
2832
mp->clsn = tblk->clsn;
fs/jfs/jfs_imap.c
2835
write_metapage(mp);
fs/jfs/jfs_imap.c
2852
struct bmap *mp = JFS_SBI(ipbmap->i_sb)->bmap;
fs/jfs/jfs_imap.c
2902
n = agstart >> mp->db_agl2size;
fs/jfs/jfs_imap.c
2903
iagp->agstart = cpu_to_le64((s64)n << mp->db_agl2size);
fs/jfs/jfs_imap.c
297
struct metapage *mp;
fs/jfs/jfs_imap.c
316
rc = diIAGRead(imap, iagno, &mp);
fs/jfs/jfs_imap.c
323
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
331
release_metapage(mp);
fs/jfs/jfs_imap.c
344
release_metapage(mp);
fs/jfs/jfs_imap.c
367
mp = read_metapage(ipimap, pageno << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
368
if (!mp) {
fs/jfs/jfs_imap.c
374
dp = (struct dinode *) mp->data;
fs/jfs/jfs_imap.c
386
release_metapage(mp);
fs/jfs/jfs_imap.c
424
struct metapage *mp;
fs/jfs/jfs_imap.c
447
mp = read_metapage(ip, address << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
448
if (mp == NULL) {
fs/jfs/jfs_imap.c
455
dp = (struct dinode *) (mp->data);
fs/jfs/jfs_imap.c
464
release_metapage(mp);
fs/jfs/jfs_imap.c
481
release_metapage(mp);
fs/jfs/jfs_imap.c
506
struct metapage *mp;
fs/jfs/jfs_imap.c
518
mp = read_metapage(ip, address << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
519
if (mp == NULL) {
fs/jfs/jfs_imap.c
525
dp = (struct dinode *) (mp->data);
fs/jfs/jfs_imap.c
536
write_metapage(mp);
fs/jfs/jfs_imap.c
591
struct metapage *mp;
fs/jfs/jfs_imap.c
639
mp = read_metapage(ipimap, pageno << sbi->l2nbperpage, PSIZE, 1);
fs/jfs/jfs_imap.c
640
if (!mp)
fs/jfs/jfs_imap.c
644
dp = (struct dinode *) mp->data;
fs/jfs/jfs_imap.c
654
txLock(tid, ipimap, mp, tlckINODE | tlckENTRY)) == NULL)
fs/jfs/jfs_imap.c
683
tlck->mp = mp;
fs/jfs/jfs_imap.c
712
tlck->mp = mp;
fs/jfs/jfs_imap.c
801
write_metapage(mp);
fs/jfs/jfs_imap.c
850
struct metapage *mp, *amp, *bmp, *cmp, *dmp;
fs/jfs/jfs_imap.c
898
if ((rc = diIAGRead(imap, iagno, &mp))) {
fs/jfs/jfs_imap.c
903
iagp = (struct iag *) mp->data;
fs/jfs/jfs_imap.c
918
release_metapage(mp);
fs/jfs/jfs_imap.c
930
release_metapage(mp);
fs/jfs/jfs_imap.c
962
release_metapage(mp);
fs/jfs/jfs_imap.c
97
struct metapage *mp;
fs/jfs/jfs_logmgr.c
1596
struct metapage *mp = (struct metapage *)lp;
fs/jfs/jfs_logmgr.c
1599
mp, sizeof(struct metapage), 0);
fs/jfs/jfs_logmgr.c
1602
sizeof(long), mp->folio,
fs/jfs/jfs_logmgr.c
230
struct metapage *mp = NULL;
fs/jfs/jfs_logmgr.c
244
tlck->type & tlckBTROOT || (mp = tlck->mp) == NULL)
fs/jfs/jfs_logmgr.c
257
if (mp->lsn == 0) {
fs/jfs/jfs_logmgr.c
258
mp->log = log;
fs/jfs/jfs_logmgr.c
259
mp->lsn = lsn;
fs/jfs/jfs_logmgr.c
263
list_add_tail(&mp->synclist, &log->synclist);
fs/jfs/jfs_logmgr.c
286
tblk->lsn = mp->lsn;
fs/jfs/jfs_logmgr.c
290
list_add(&tblk->synclist, &mp->synclist);
fs/jfs/jfs_logmgr.c
297
logdiff(diffp, mp->lsn, log);
fs/jfs/jfs_logmgr.c
301
tblk->lsn = mp->lsn;
fs/jfs/jfs_logmgr.c
304
list_move(&tblk->synclist, &mp->synclist);
fs/jfs/jfs_logmgr.c
380
p = (caddr_t) (tlck->mp->data);
fs/jfs/jfs_metapage.c
108
if (mp) {
fs/jfs/jfs_metapage.c
110
index = (mp->index >> l2mp_blocks) & (MPS_PER_PAGE - 1);
fs/jfs/jfs_metapage.c
112
a->mp[index] = mp;
fs/jfs/jfs_metapage.c
118
static inline void remove_metapage(struct folio *folio, struct metapage *mp)
fs/jfs/jfs_metapage.c
124
index = (mp->index >> l2mp_blocks) & (MPS_PER_PAGE - 1);
fs/jfs/jfs_metapage.c
126
BUG_ON(a->mp[index] != mp);
fs/jfs/jfs_metapage.c
128
a->mp[index] = NULL;
fs/jfs/jfs_metapage.c
162
struct metapage *mp;
fs/jfs/jfs_metapage.c
166
mp = src_anchor->mp[i];
fs/jfs/jfs_metapage.c
167
if (mp && metapage_locked(mp))
fs/jfs/jfs_metapage.c
176
mp = src_anchor->mp[i];
fs/jfs/jfs_metapage.c
177
if (!mp)
fs/jfs/jfs_metapage.c
179
if (unlikely(insert_metapage(dst, mp))) {
fs/jfs/jfs_metapage.c
187
mps[i] = mp;
fs/jfs/jfs_metapage.c
192
mp = mps[i];
fs/jfs/jfs_metapage.c
193
if (mp) {
fs/jfs/jfs_metapage.c
194
int page_offset = mp->data - folio_address(src);
fs/jfs/jfs_metapage.c
196
mp->data = folio_address(dst) + page_offset;
fs/jfs/jfs_metapage.c
197
mp->folio = dst;
fs/jfs/jfs_metapage.c
198
remove_metapage(src, mp);
fs/jfs/jfs_metapage.c
213
static inline int insert_metapage(struct folio *folio, struct metapage *mp)
fs/jfs/jfs_metapage.c
215
if (mp) {
fs/jfs/jfs_metapage.c
216
folio_attach_private(folio, mp);
fs/jfs/jfs_metapage.c
222
static inline void remove_metapage(struct folio *folio, struct metapage *mp)
fs/jfs/jfs_metapage.c
236
struct metapage *mp;
fs/jfs/jfs_metapage.c
240
mp = folio_to_mp(src, 0);
fs/jfs/jfs_metapage.c
241
if (metapage_locked(mp))
fs/jfs/jfs_metapage.c
248
if (unlikely(insert_metapage(dst, mp)))
fs/jfs/jfs_metapage.c
251
page_offset = mp->data - folio_address(src);
fs/jfs/jfs_metapage.c
252
mp->data = folio_address(dst) + page_offset;
fs/jfs/jfs_metapage.c
253
mp->folio = dst;
fs/jfs/jfs_metapage.c
254
remove_metapage(src, mp);
fs/jfs/jfs_metapage.c
264
struct metapage *mp = mempool_alloc(metapage_mempool, gfp_mask);
fs/jfs/jfs_metapage.c
266
if (mp) {
fs/jfs/jfs_metapage.c
267
mp->lid = 0;
fs/jfs/jfs_metapage.c
268
mp->lsn = 0;
fs/jfs/jfs_metapage.c
269
mp->data = NULL;
fs/jfs/jfs_metapage.c
270
mp->clsn = 0;
fs/jfs/jfs_metapage.c
271
mp->log = NULL;
fs/jfs/jfs_metapage.c
272
init_waitqueue_head(&mp->wait);
fs/jfs/jfs_metapage.c
274
return mp;
fs/jfs/jfs_metapage.c
277
static inline void free_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
279
mempool_free(mp, metapage_mempool);
fs/jfs/jfs_metapage.c
309
static inline void drop_metapage(struct folio *folio, struct metapage *mp)
fs/jfs/jfs_metapage.c
311
if (mp->count || mp->nohomeok || test_bit(META_dirty, &mp->flag) ||
fs/jfs/jfs_metapage.c
312
test_bit(META_io, &mp->flag))
fs/jfs/jfs_metapage.c
314
remove_metapage(folio, mp);
fs/jfs/jfs_metapage.c
316
free_metapage(mp);
fs/jfs/jfs_metapage.c
34
#define metapage_locked(mp) test_bit(META_locked, &(mp)->flag)
fs/jfs/jfs_metapage.c
35
#define trylock_metapage(mp) test_and_set_bit_lock(META_locked, &(mp)->flag)
fs/jfs/jfs_metapage.c
365
static void remove_from_logsync(struct metapage *mp)
fs/jfs/jfs_metapage.c
367
struct jfs_log *log = mp->log;
fs/jfs/jfs_metapage.c
37
static inline void unlock_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
377
if (mp->lsn) {
fs/jfs/jfs_metapage.c
378
mp->log = NULL;
fs/jfs/jfs_metapage.c
379
mp->lsn = 0;
fs/jfs/jfs_metapage.c
380
mp->clsn = 0;
fs/jfs/jfs_metapage.c
382
list_del(&mp->synclist);
fs/jfs/jfs_metapage.c
389
struct metapage *mp;
fs/jfs/jfs_metapage.c
39
clear_bit_unlock(META_locked, &mp->flag);
fs/jfs/jfs_metapage.c
399
mp = folio_to_mp(folio, offset);
fs/jfs/jfs_metapage.c
40
wake_up(&mp->wait);
fs/jfs/jfs_metapage.c
400
if (mp && test_bit(META_io, &mp->flag)) {
fs/jfs/jfs_metapage.c
401
if (mp->lsn)
fs/jfs/jfs_metapage.c
402
remove_from_logsync(mp);
fs/jfs/jfs_metapage.c
403
clear_bit(META_io, &mp->flag);
fs/jfs/jfs_metapage.c
43
static inline void __lock_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
432
struct metapage *mp;
fs/jfs/jfs_metapage.c
450
mp = folio_to_mp(folio, offset);
fs/jfs/jfs_metapage.c
452
if (!mp || !test_bit(META_dirty, &mp->flag))
fs/jfs/jfs_metapage.c
455
if (mp->nohomeok && !test_bit(META_forcewrite, &mp->flag)) {
fs/jfs/jfs_metapage.c
461
if (mp->log && !(mp->log->cflag & logGC_PAGEOUT))
fs/jfs/jfs_metapage.c
462
jfs_flush_journal(mp->log, 0);
fs/jfs/jfs_metapage.c
466
clear_bit(META_dirty, &mp->flag);
fs/jfs/jfs_metapage.c
467
set_bit(META_io, &mp->flag);
fs/jfs/jfs_metapage.c
47
add_wait_queue_exclusive(&mp->wait, &wait);
fs/jfs/jfs_metapage.c
50
if (metapage_locked(mp)) {
fs/jfs/jfs_metapage.c
51
folio_unlock(mp->folio);
fs/jfs/jfs_metapage.c
53
folio_lock(mp->folio);
fs/jfs/jfs_metapage.c
55
} while (trylock_metapage(mp));
fs/jfs/jfs_metapage.c
57
remove_wait_queue(&mp->wait, &wait);
fs/jfs/jfs_metapage.c
614
struct metapage *mp;
fs/jfs/jfs_metapage.c
619
mp = folio_to_mp(folio, offset);
fs/jfs/jfs_metapage.c
621
if (!mp)
fs/jfs/jfs_metapage.c
624
jfs_info("metapage_release_folio: mp = 0x%p", mp);
fs/jfs/jfs_metapage.c
625
if (mp->count || mp->nohomeok ||
fs/jfs/jfs_metapage.c
626
test_bit(META_dirty, &mp->flag)) {
fs/jfs/jfs_metapage.c
627
jfs_info("count = %ld, nohomeok = %d", mp->count,
fs/jfs/jfs_metapage.c
628
mp->nohomeok);
fs/jfs/jfs_metapage.c
63
static inline void lock_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
632
if (mp->lsn)
fs/jfs/jfs_metapage.c
633
remove_from_logsync(mp);
fs/jfs/jfs_metapage.c
634
remove_metapage(folio, mp);
fs/jfs/jfs_metapage.c
636
free_metapage(mp);
fs/jfs/jfs_metapage.c
65
if (trylock_metapage(mp))
fs/jfs/jfs_metapage.c
66
__lock_metapage(mp);
fs/jfs/jfs_metapage.c
690
struct metapage *mp = NULL;
fs/jfs/jfs_metapage.c
737
mp = folio_to_mp(folio, page_offset);
fs/jfs/jfs_metapage.c
738
if (mp) {
fs/jfs/jfs_metapage.c
739
if (mp->logical_size != size) {
fs/jfs/jfs_metapage.c
743
mp->logical_size, size);
fs/jfs/jfs_metapage.c
747
mp->count++;
fs/jfs/jfs_metapage.c
748
lock_metapage(mp);
fs/jfs/jfs_metapage.c
749
if (test_bit(META_discard, &mp->flag)) {
fs/jfs/jfs_metapage.c
753
discard_metapage(mp);
fs/jfs/jfs_metapage.c
756
clear_bit(META_discard, &mp->flag);
fs/jfs/jfs_metapage.c
760
mp = alloc_metapage(GFP_NOFS);
fs/jfs/jfs_metapage.c
761
if (!mp)
fs/jfs/jfs_metapage.c
763
mp->folio = folio;
fs/jfs/jfs_metapage.c
764
mp->sb = inode->i_sb;
fs/jfs/jfs_metapage.c
765
mp->flag = 0;
fs/jfs/jfs_metapage.c
766
mp->xflag = COMMIT_PAGE;
fs/jfs/jfs_metapage.c
767
mp->count = 1;
fs/jfs/jfs_metapage.c
768
mp->nohomeok = 0;
fs/jfs/jfs_metapage.c
769
mp->logical_size = size;
fs/jfs/jfs_metapage.c
770
mp->data = folio_address(folio) + page_offset;
fs/jfs/jfs_metapage.c
771
mp->index = lblock;
fs/jfs/jfs_metapage.c
772
if (unlikely(insert_metapage(folio, mp))) {
fs/jfs/jfs_metapage.c
773
free_metapage(mp);
fs/jfs/jfs_metapage.c
776
lock_metapage(mp);
fs/jfs/jfs_metapage.c
780
jfs_info("zeroing mp = 0x%p", mp);
fs/jfs/jfs_metapage.c
781
memset(mp->data, 0, PSIZE);
fs/jfs/jfs_metapage.c
785
jfs_info("__get_metapage: returning = 0x%p data = 0x%p", mp, mp->data);
fs/jfs/jfs_metapage.c
786
return mp;
fs/jfs/jfs_metapage.c
793
void grab_metapage(struct metapage * mp)
fs/jfs/jfs_metapage.c
795
jfs_info("grab_metapage: mp = 0x%p", mp);
fs/jfs/jfs_metapage.c
796
folio_get(mp->folio);
fs/jfs/jfs_metapage.c
797
folio_lock(mp->folio);
fs/jfs/jfs_metapage.c
798
mp->count++;
fs/jfs/jfs_metapage.c
799
lock_metapage(mp);
fs/jfs/jfs_metapage.c
800
folio_unlock(mp->folio);
fs/jfs/jfs_metapage.c
81
struct metapage *mp[MPS_PER_PAGE];
fs/jfs/jfs_metapage.c
831
void force_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
833
struct folio *folio = mp->folio;
fs/jfs/jfs_metapage.c
834
jfs_info("force_metapage: mp = 0x%p", mp);
fs/jfs/jfs_metapage.c
835
set_bit(META_forcewrite, &mp->flag);
fs/jfs/jfs_metapage.c
836
clear_bit(META_sync, &mp->flag);
fs/jfs/jfs_metapage.c
841
jfs_error(mp->sb, "metapage_write_one() failed\n");
fs/jfs/jfs_metapage.c
842
clear_bit(META_forcewrite, &mp->flag);
fs/jfs/jfs_metapage.c
846
void hold_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
848
folio_lock(mp->folio);
fs/jfs/jfs_metapage.c
851
void put_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.c
853
if (mp->count || mp->nohomeok) {
fs/jfs/jfs_metapage.c
855
folio_unlock(mp->folio);
fs/jfs/jfs_metapage.c
858
folio_get(mp->folio);
fs/jfs/jfs_metapage.c
859
mp->count++;
fs/jfs/jfs_metapage.c
860
lock_metapage(mp);
fs/jfs/jfs_metapage.c
861
folio_unlock(mp->folio);
fs/jfs/jfs_metapage.c
862
release_metapage(mp);
fs/jfs/jfs_metapage.c
865
void release_metapage(struct metapage * mp)
fs/jfs/jfs_metapage.c
867
struct folio *folio = mp->folio;
fs/jfs/jfs_metapage.c
868
jfs_info("release_metapage: mp = 0x%p, flag = 0x%lx", mp, mp->flag);
fs/jfs/jfs_metapage.c
871
unlock_metapage(mp);
fs/jfs/jfs_metapage.c
873
assert(mp->count);
fs/jfs/jfs_metapage.c
874
if (--mp->count || mp->nohomeok) {
fs/jfs/jfs_metapage.c
880
if (test_bit(META_dirty, &mp->flag)) {
fs/jfs/jfs_metapage.c
882
if (test_bit(META_sync, &mp->flag)) {
fs/jfs/jfs_metapage.c
883
clear_bit(META_sync, &mp->flag);
fs/jfs/jfs_metapage.c
885
jfs_error(mp->sb, "metapage_write_one() failed\n");
fs/jfs/jfs_metapage.c
888
} else if (mp->lsn) /* discard_metapage doesn't remove it */
fs/jfs/jfs_metapage.c
889
remove_from_logsync(mp);
fs/jfs/jfs_metapage.c
892
drop_metapage(folio, mp);
fs/jfs/jfs_metapage.c
90
return anchor->mp[offset >> L2PSIZE];
fs/jfs/jfs_metapage.c
906
struct metapage *mp;
fs/jfs/jfs_metapage.c
920
mp = folio_to_mp(folio, offset);
fs/jfs/jfs_metapage.c
921
if (!mp)
fs/jfs/jfs_metapage.c
923
if (mp->index < addr)
fs/jfs/jfs_metapage.c
925
if (mp->index >= addr + len)
fs/jfs/jfs_metapage.c
928
clear_bit(META_dirty, &mp->flag);
fs/jfs/jfs_metapage.c
929
set_bit(META_discard, &mp->flag);
fs/jfs/jfs_metapage.c
93
static inline int insert_metapage(struct folio *folio, struct metapage *mp)
fs/jfs/jfs_metapage.c
930
if (mp->lsn)
fs/jfs/jfs_metapage.c
931
remove_from_logsync(mp);
fs/jfs/jfs_metapage.h
107
static inline void metapage_wait_for_io(struct metapage *mp)
fs/jfs/jfs_metapage.h
109
if (test_bit(META_io, &mp->flag))
fs/jfs/jfs_metapage.h
110
folio_wait_writeback(mp->folio);
fs/jfs/jfs_metapage.h
116
static inline void _metapage_homeok(struct metapage *mp)
fs/jfs/jfs_metapage.h
118
if (!--mp->nohomeok)
fs/jfs/jfs_metapage.h
119
folio_put(mp->folio);
fs/jfs/jfs_metapage.h
122
static inline void metapage_homeok(struct metapage *mp)
fs/jfs/jfs_metapage.h
124
hold_metapage(mp);
fs/jfs/jfs_metapage.h
125
_metapage_homeok(mp);
fs/jfs/jfs_metapage.h
126
put_metapage(mp);
fs/jfs/jfs_metapage.h
45
#define mark_metapage_dirty(mp) set_bit(META_dirty, &(mp)->flag)
fs/jfs/jfs_metapage.h
72
static inline void write_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.h
74
set_bit(META_dirty, &mp->flag);
fs/jfs/jfs_metapage.h
75
release_metapage(mp);
fs/jfs/jfs_metapage.h
78
static inline void flush_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.h
80
set_bit(META_sync, &mp->flag);
fs/jfs/jfs_metapage.h
81
write_metapage(mp);
fs/jfs/jfs_metapage.h
84
static inline void discard_metapage(struct metapage *mp)
fs/jfs/jfs_metapage.h
86
clear_bit(META_dirty, &mp->flag);
fs/jfs/jfs_metapage.h
87
set_bit(META_discard, &mp->flag);
fs/jfs/jfs_metapage.h
88
release_metapage(mp);
fs/jfs/jfs_metapage.h
91
static inline void metapage_nohomeok(struct metapage *mp)
fs/jfs/jfs_metapage.h
93
struct folio *folio = mp->folio;
fs/jfs/jfs_metapage.h
95
if (!mp->nohomeok++) {
fs/jfs/jfs_metapage.h
96
mark_metapage_dirty(mp);
fs/jfs/jfs_txnmgr.c
1014
tlck->mp = NULL;
fs/jfs/jfs_txnmgr.c
1432
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
1436
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
1450
PXDaddress(pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1452
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1510
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
1513
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
1529
mp->lid = 0;
fs/jfs/jfs_txnmgr.c
1530
grab_metapage(mp);
fs/jfs/jfs_txnmgr.c
1531
metapage_homeok(mp);
fs/jfs/jfs_txnmgr.c
1532
discard_metapage(mp);
fs/jfs/jfs_txnmgr.c
1533
tlck->mp = NULL;
fs/jfs/jfs_txnmgr.c
1537
PXDaddress(pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1538
PXDlength(pxd, mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1556
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
1560
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
1587
PXDaddress(pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1589
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1616
PXDaddress(pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1618
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1660
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
1670
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
168
static void LogSyncRelease(struct metapage * mp);
fs/jfs/jfs_txnmgr.c
1685
p = (xtpage_t *) mp->data;
fs/jfs/jfs_txnmgr.c
1706
PXDaddress(page_pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1708
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1758
tlck->ip, mp, tlck, lwm, xadlock->count);
fs/jfs/jfs_txnmgr.c
1795
PXDaddress(page_pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1797
mp->logical_size >> tblk->sb->
fs/jfs/jfs_txnmgr.c
1861
tlck->ip, mp, xadlock->count);
fs/jfs/jfs_txnmgr.c
1916
PXDaddress(page_pxd, mp->index);
fs/jfs/jfs_txnmgr.c
1918
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1984
tlck->ip, mp, xadlock->count, lwm, next);
fs/jfs/jfs_txnmgr.c
2005
ip, mp, pxdlock->count, hwm);
fs/jfs/jfs_txnmgr.c
2024
tlck->ip, mp, xadlock->count, next, hwm);
fs/jfs/jfs_txnmgr.c
2182
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
2208
if ((mp = tlck->mp) != NULL &&
fs/jfs/jfs_txnmgr.c
2210
assert(mp->xflag & COMMIT_PAGE);
fs/jfs/jfs_txnmgr.c
2216
force_metapage(mp);
fs/jfs/jfs_txnmgr.c
2229
assert(mp->nohomeok);
fs/jfs/jfs_txnmgr.c
2230
set_bit(META_dirty, &mp->flag);
fs/jfs/jfs_txnmgr.c
2231
set_bit(META_sync, &mp->flag);
fs/jfs/jfs_txnmgr.c
2256
struct metapage *mp = NULL;
fs/jfs/jfs_txnmgr.c
2289
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
2290
ASSERT(mp->xflag & COMMIT_PAGE);
fs/jfs/jfs_txnmgr.c
2291
grab_metapage(mp);
fs/jfs/jfs_txnmgr.c
2338
ASSERT(mp->lid == lid);
fs/jfs/jfs_txnmgr.c
2339
tlck->mp->lid = 0;
fs/jfs/jfs_txnmgr.c
2341
assert(mp->nohomeok == 1);
fs/jfs/jfs_txnmgr.c
2342
metapage_homeok(mp);
fs/jfs/jfs_txnmgr.c
2343
discard_metapage(mp);
fs/jfs/jfs_txnmgr.c
2344
tlck->mp = NULL;
fs/jfs/jfs_txnmgr.c
2597
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
2607
mp = tlck->mp;
fs/jfs/jfs_txnmgr.c
2610
if (mp) {
fs/jfs/jfs_txnmgr.c
2611
mp->lid = 0;
fs/jfs/jfs_txnmgr.c
2621
if (mp->xflag & COMMIT_PAGE && mp->lsn)
fs/jfs/jfs_txnmgr.c
2622
LogSyncRelease(mp);
fs/jfs/jfs_txnmgr.c
2793
static void LogSyncRelease(struct metapage * mp)
fs/jfs/jfs_txnmgr.c
2795
struct jfs_log *log = mp->log;
fs/jfs/jfs_txnmgr.c
2797
assert(mp->nohomeok);
fs/jfs/jfs_txnmgr.c
2799
metapage_homeok(mp);
fs/jfs/jfs_txnmgr.c
583
struct tlock *txLock(tid_t tid, struct inode *ip, struct metapage * mp,
fs/jfs/jfs_txnmgr.c
599
!(mp->xflag & COMMIT_PAGE)) {
fs/jfs/jfs_txnmgr.c
607
lid = mp->lid;
fs/jfs/jfs_txnmgr.c
613
jfs_info("txLock: tid:%d ip:0x%p mp:0x%p lid:%d", tid, ip, mp, lid);
fs/jfs/jfs_txnmgr.c
698
if (mp->xflag & COMMIT_PAGE) {
fs/jfs/jfs_txnmgr.c
703
metapage_nohomeok(mp);
fs/jfs/jfs_txnmgr.c
706
mp, mp->nohomeok, tid, tlck);
fs/jfs/jfs_txnmgr.c
713
if ((tid == 0) && mp->lsn)
fs/jfs/jfs_txnmgr.c
727
tlck->mp = mp;
fs/jfs/jfs_txnmgr.c
731
mp->lid = lid;
fs/jfs/jfs_txnmgr.c
784
if (mp->xflag & COMMIT_PAGE)
fs/jfs/jfs_txnmgr.c
785
p = (xtpage_t *) mp->data;
fs/jfs/jfs_txnmgr.c
829
mp, sizeof(*mp), 0);
fs/jfs/jfs_txnmgr.c
839
release_metapage(mp);
fs/jfs/jfs_txnmgr.c
847
if (xtid && (tlck->mp == mp) && (mp->lid == lid))
fs/jfs/jfs_txnmgr.c
871
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
879
if ((mp = tlck->mp) != NULL &&
fs/jfs/jfs_txnmgr.c
881
assert(mp->xflag & COMMIT_PAGE);
fs/jfs/jfs_txnmgr.c
882
mp->lid = 0;
fs/jfs/jfs_txnmgr.c
906
struct metapage *mp;
fs/jfs/jfs_txnmgr.c
924
if ((mp = tlck->mp) != NULL &&
fs/jfs/jfs_txnmgr.c
926
assert(mp->xflag & COMMIT_PAGE);
fs/jfs/jfs_txnmgr.c
930
hold_metapage(mp);
fs/jfs/jfs_txnmgr.c
932
assert(mp->nohomeok > 0);
fs/jfs/jfs_txnmgr.c
933
_metapage_homeok(mp);
fs/jfs/jfs_txnmgr.c
937
if (mp->clsn) {
fs/jfs/jfs_txnmgr.c
939
logdiff(diffp, mp->clsn, log);
fs/jfs/jfs_txnmgr.c
941
mp->clsn = tblk->clsn;
fs/jfs/jfs_txnmgr.c
943
mp->clsn = tblk->clsn;
fs/jfs/jfs_txnmgr.c
948
put_metapage(mp);
fs/jfs/jfs_txnmgr.h
91
struct metapage *mp; /* 4/8: object page buffer locked */
fs/jfs/jfs_xtree.c
1080
p = xt_getpage(ip, nextbn, &mp);
fs/jfs/jfs_xtree.c
1086
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1093
tlck = txLock(tid, ip, mp, tlckXTREE | tlckRELINK);
fs/jfs/jfs_xtree.c
1101
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
111
static inline xtpage_t *xt_getpage(struct inode *ip, s64 bn, struct metapage **mp)
fs/jfs/jfs_xtree.c
116
BT_GETPAGE(ip, bn, *mp, xtpage_t, PSIZE, p, rc, i_xtroot);
fs/jfs/jfs_xtree.c
126
BT_PUTPAGE(*mp);
fs/jfs/jfs_xtree.c
127
*mp = NULL;
fs/jfs/jfs_xtree.c
1316
BT_MARK_DIRTY(split->mp, ip);
fs/jfs/jfs_xtree.c
1328
tlck = txLock(tid, ip, split->mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1358
struct metapage *mp; /* meta-page buffer */
fs/jfs/jfs_xtree.c
1376
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
1379
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1387
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1397
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1399
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1424
split.mp = mp;
fs/jfs/jfs_xtree.c
1435
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1448
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
145
struct metapage *mp;
fs/jfs/jfs_xtree.c
1451
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1455
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1457
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
1495
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1517
struct metapage *mp; /* meta-page buffer */
fs/jfs/jfs_xtree.c
1541
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index0);
fs/jfs/jfs_xtree.c
1544
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1549
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1554
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1567
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1697
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1718
split.mp = mp;
fs/jfs/jfs_xtree.c
1729
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1742
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1745
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1749
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1751
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
180
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
1803
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1806
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1810
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1812
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
1828
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1834
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index0);
fs/jfs/jfs_xtree.c
1837
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1843
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1871
split.mp = mp;
fs/jfs/jfs_xtree.c
1882
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1896
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1899
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
1903
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
1905
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
1933
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
1963
struct metapage *mp; /* meta-page buffer */
fs/jfs/jfs_xtree.c
1995
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
2051
split.mp = mp;
fs/jfs/jfs_xtree.c
2078
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
208
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2084
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
2104
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2206
struct metapage *mp;
fs/jfs/jfs_xtree.c
2285
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2304
tlck = txLock(tid, ip, mp, tlckXTREE|tlckGROW);
fs/jfs/jfs_xtree.c
2305
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2322
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2333
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2337
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2342
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
237
struct metapage *mp; /* page buffer */
fs/jfs/jfs_xtree.c
2457
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2489
XT_PUTPAGE(mp); /* debug */
fs/jfs/jfs_xtree.c
2495
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2498
if (mp->lid)
fs/jfs/jfs_xtree.c
2499
lid_to_tlock(mp->lid)->flag |= tlckFREELOCK;
fs/jfs/jfs_xtree.c
2502
discard_metapage(mp);
fs/jfs/jfs_xtree.c
2523
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2540
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2559
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2563
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2585
if (log && mp->lid && (tblk->last != mp->lid) &&
fs/jfs/jfs_xtree.c
2586
lid_to_tlock(mp->lid)->tid) {
fs/jfs/jfs_xtree.c
2587
lid_t lid = mp->lid;
fs/jfs/jfs_xtree.c
2616
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2631
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
2647
XT_PUTPAGE(mp); /* debug */
fs/jfs/jfs_xtree.c
2653
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2656
if (mp->lid)
fs/jfs/jfs_xtree.c
2657
lid_to_tlock(mp->lid)->flag |=
fs/jfs/jfs_xtree.c
2661
discard_metapage(mp);
fs/jfs/jfs_xtree.c
2693
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
270
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2706
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2765
struct metapage *mp;
fs/jfs/jfs_xtree.c
2789
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
2792
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2808
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2831
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2834
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2840
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2853
p = xt_getpage(ip, bn, &mp);
fs/jfs/jfs_xtree.c
2867
tlck = txLock(tid, ip, mp, tlckXTREE);
fs/jfs/jfs_xtree.c
2872
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2893
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
2906
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
353
btsp->mp = mp;
fs/jfs/jfs_xtree.c
400
btsp->mp = mp;
fs/jfs/jfs_xtree.c
458
btsp->mp = mp;
fs/jfs/jfs_xtree.c
495
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
504
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
534
struct metapage *mp; /* meta-page buffer */
fs/jfs/jfs_xtree.c
560
XT_GETSEARCH(ip, btstack.top, bn, mp, p, index);
fs/jfs/jfs_xtree.c
59
struct metapage *mp;
fs/jfs/jfs_xtree.c
601
split.mp = mp;
fs/jfs/jfs_xtree.c
629
BT_MARK_DIRTY(mp, ip);
fs/jfs/jfs_xtree.c
645
tlck = txLock(tid, ip, mp, tlckXTREE | tlckGROW);
fs/jfs/jfs_xtree.c
658
XT_PUTPAGE(mp);
fs/jfs/jfs_xtree.c
702
smp = split->mp;
fs/jfs/jfs_xtree.c
846
split->mp = smp;
fs/jfs/jfs_xtree.c
958
struct metapage *mp;
fs/jfs/jfs_xtree.c
969
smp = split->mp;
fs/jfs/namei.c
1001
if (mp == NULL) {
fs/jfs/namei.c
1007
memcpy(mp->data, name, copy_size);
fs/jfs/namei.c
1008
flush_metapage(mp);
fs/jfs/namei.c
892
struct metapage *mp;
fs/jfs/namei.c
999
mp = get_metapage(ip, xaddr, PSIZE, 1);
fs/jfs/xattr.c
213
struct metapage *mp;
fs/jfs/xattr.c
256
if (!(mp = get_metapage(ip, blkno + i, bytes_to_write, 1))) {
fs/jfs/xattr.c
261
memcpy(mp->data, cp, nb);
fs/jfs/xattr.c
270
if ((rc = flush_metapage(mp))) {
fs/jfs/xattr.c
280
flush_metapage(mp);
fs/jfs/xattr.c
362
struct metapage *mp;
fs/jfs/xattr.c
396
if (!(mp = read_metapage(ip, blkno + i, bytes_to_read, 1)))
fs/jfs/xattr.c
399
memcpy(cp, mp->data, nb);
fs/jfs/xattr.c
400
release_metapage(mp);
fs/jfs/xattr.c
528
ea_buf->mp = get_metapage(inode, blkno,
fs/jfs/xattr.c
531
if (ea_buf->mp == NULL) {
fs/jfs/xattr.c
536
ea_buf->xattr = ea_buf->mp->data;
fs/jfs/xattr.c
542
discard_metapage(ea_buf->mp);
fs/jfs/xattr.c
549
ea_buf->mp = read_metapage(inode, addressDXD(&ji->ea),
fs/jfs/xattr.c
552
if (ea_buf->mp == NULL) {
fs/jfs/xattr.c
556
ea_buf->xattr = ea_buf->mp->data;
fs/jfs/xattr.c
592
assert(ea_buf->mp);
fs/jfs/xattr.c
593
release_metapage(ea_buf->mp);
fs/jfs/xattr.c
623
flush_metapage(ea_buf->mp);
fs/jfs/xattr.c
627
discard_metapage(ea_buf->mp);
fs/jfs/xattr.c
63
struct metapage *mp; /* metapage containing ea list */
fs/namespace.c
1000
maybe_free_mountpoint(mp, shrink_list);
fs/namespace.c
1015
struct mountpoint *mp,
fs/namespace.c
1018
child_mnt->mnt_mountpoint = mp->m_dentry;
fs/namespace.c
1020
child_mnt->mnt_mp = mp;
fs/namespace.c
1021
hlist_add_head(&child_mnt->mnt_mp_list, &mp->m_list);
fs/namespace.c
1051
struct mountpoint *mp)
fs/namespace.c
1053
mnt_set_mountpoint(parent, mp, mnt);
fs/namespace.c
1057
void mnt_change_mountpoint(struct mount *parent, struct mountpoint *mp, struct mount *mnt)
fs/namespace.c
1065
attach_mnt(mnt, parent, mp);
fs/namespace.c
1978
struct pinned_mountpoint mp = {};
fs/namespace.c
1984
if (!lookup_mountpoint(dentry, &mp))
fs/namespace.c
1988
while (mp.node.next) {
fs/namespace.c
1989
mnt = hlist_entry(mp.node.next, struct mount, mnt_mp_list);
fs/namespace.c
1996
unpin_mountpoint(&mp);
fs/namespace.c
2564
struct mountpoint *dest_mp = dest->mp;
fs/namespace.c
2645
struct mountpoint *mp = root.mp;
fs/namespace.c
2648
mp = shorter;
fs/namespace.c
2649
mnt_change_mountpoint(r, mp, q);
fs/namespace.c
2790
inode_unlock(m->mp->m_dentry->d_inode);
fs/namespace.c
2804
struct pinned_mountpoint *mp, bool copy_mount,
fs/namespace.c
2807
#define LOCK_MOUNT_MAYBE_BENEATH(mp, path, beneath) \
fs/namespace.c
2808
struct pinned_mountpoint mp __cleanup(unlock_mount) = {}; \
fs/namespace.c
2809
do_lock_mount((path), &mp, (beneath))
fs/namespace.c
2810
#define LOCK_MOUNT(mp, path) LOCK_MOUNT_MAYBE_BENEATH(mp, (path), false)
fs/namespace.c
2811
#define LOCK_MOUNT_EXACT(mp, path) \
fs/namespace.c
2812
struct pinned_mountpoint mp __cleanup(unlock_mount) = {}; \
fs/namespace.c
2813
lock_mount_exact((path), &mp, false, 0)
fs/namespace.c
2814
#define LOCK_MOUNT_EXACT_COPY(mp, path, copy_flags) \
fs/namespace.c
2815
struct pinned_mountpoint mp __cleanup(unlock_mount) = {}; \
fs/namespace.c
2816
lock_mount_exact((path), &mp, true, (copy_flags))
fs/namespace.c
2818
static int graft_tree(struct mount *mnt, const struct pinned_mountpoint *mp)
fs/namespace.c
2823
if (d_is_dir(mp->mp->m_dentry) !=
fs/namespace.c
2827
return attach_recursive_mnt(mnt, mp);
fs/namespace.c
3013
LOCK_MOUNT(mp, path);
fs/namespace.c
3014
if (IS_ERR(mp.parent))
fs/namespace.c
3015
return PTR_ERR(mp.parent);
fs/namespace.c
3017
if (!check_mnt(mp.parent))
fs/namespace.c
3024
err = graft_tree(mnt, &mp);
fs/namespace.c
3114
LOCK_MOUNT_EXACT_COPY(mp, &to_path, copy_flags);
fs/namespace.c
3115
if (IS_ERR(mp.parent)) {
fs/namespace.c
3117
return ERR_CAST(mp.parent);
fs/namespace.c
3119
new_ns_root = mp.parent;
fs/namespace.c
3152
attach_mnt(mnt, new_ns_root, mp.mp);
fs/namespace.c
3531
const struct mountpoint *mp)
fs/namespace.c
3560
if (propagation_would_overmount(parent_mnt_to, mnt_to, mp))
fs/namespace.c
3576
propagation_would_overmount(parent_mnt_to, mnt_from, mp))
fs/namespace.c
3644
LOCK_MOUNT_MAYBE_BENEATH(mp, new_path, beneath);
fs/namespace.c
3645
if (IS_ERR(mp.parent))
fs/namespace.c
3646
return PTR_ERR(mp.parent);
fs/namespace.c
3657
if (!check_mnt(mp.parent))
fs/namespace.c
3670
if (old->mnt_ns == mp.parent->mnt_ns)
fs/namespace.c
3676
if (!may_use_mount(mp.parent))
fs/namespace.c
3683
if (mp.parent != over->mnt_parent)
fs/namespace.c
3684
over = mp.parent->overmount;
fs/namespace.c
3685
err = can_move_mount_beneath(old, over, mp.mp);
fs/namespace.c
3694
if (IS_MNT_SHARED(mp.parent) && tree_contains_unbindable(old))
fs/namespace.c
3698
if (mount_is_ancestor(old, mp.parent))
fs/namespace.c
3701
return attach_recursive_mnt(old, &mp);
fs/namespace.c
3722
static int do_add_mount(struct mount *newmnt, const struct pinned_mountpoint *mp,
fs/namespace.c
3725
struct mount *parent = mp->parent;
fs/namespace.c
3743
parent->mnt.mnt_root == mp->mp->m_dentry)
fs/namespace.c
3750
return graft_tree(newmnt, mp);
fs/namespace.c
3781
LOCK_MOUNT(mp, mountpoint);
fs/namespace.c
3782
error = do_add_mount(real_mount(mnt), &mp, mnt_flags);
fs/namespace.c
3846
struct pinned_mountpoint *mp, bool copy_mount,
fs/namespace.c
3862
err = get_mountpoint(dentry, mp);
fs/namespace.c
3866
mp->parent = ERR_PTR(err);
fs/namespace.c
3871
mp->parent = clone_mnt(real_mount(path->mnt), dentry, copy_flags);
fs/namespace.c
3873
mp->parent = real_mount(path->mnt);
fs/namespace.c
3874
if (unlikely(IS_ERR(mp->parent)))
fs/namespace.c
3875
__unlock_mount(mp);
fs/namespace.c
3899
LOCK_MOUNT_EXACT(mp, path);
fs/namespace.c
3900
if (mp.parent == ERR_PTR(-EBUSY))
fs/namespace.c
3903
err = do_add_mount(mnt, &mp, path->mnt->mnt_flags | MNT_SHRINKABLE);
fs/namespace.c
4675
if (!is_path_reachable(old_mnt, old_mp.mp->m_dentry, new))
fs/namespace.c
4690
attach_mnt(root_mnt, old_mnt, old_mp.mp);
fs/namespace.c
6172
LOCK_MOUNT_EXACT(mp, &root);
fs/namespace.c
6173
if (unlikely(IS_ERR(mp.parent)))
fs/namespace.c
6176
attach_mnt(real_mount(mnt), mp.parent, mp.mp);
fs/namespace.c
854
struct mountpoint *mp;
fs/namespace.c
861
struct mountpoint *mp;
fs/namespace.c
863
hlist_for_each_entry(mp, chain, m_hash) {
fs/namespace.c
864
if (mp->m_dentry == dentry) {
fs/namespace.c
865
hlist_add_head(&m->node, &mp->m_list);
fs/namespace.c
866
m->mp = mp;
fs/namespace.c
875
struct mountpoint *mp __free(kfree) = NULL;
fs/namespace.c
891
if (!mp)
fs/namespace.c
892
mp = kmalloc_obj(struct mountpoint);
fs/namespace.c
893
if (!mp)
fs/namespace.c
909
mp->m_dentry = dget(dentry);
fs/namespace.c
910
hlist_add_head(&mp->m_hash, mp_hash(dentry));
fs/namespace.c
911
INIT_HLIST_HEAD(&mp->m_list);
fs/namespace.c
912
hlist_add_head(&m->node, &mp->m_list);
fs/namespace.c
913
m->mp = no_free_ptr(mp);
fs/namespace.c
922
static void maybe_free_mountpoint(struct mountpoint *mp, struct list_head *list)
fs/namespace.c
924
if (hlist_empty(&mp->m_list)) {
fs/namespace.c
925
struct dentry *dentry = mp->m_dentry;
fs/namespace.c
930
hlist_del(&mp->m_hash);
fs/namespace.c
931
kfree(mp);
fs/namespace.c
940
if (m->mp) {
fs/namespace.c
942
maybe_free_mountpoint(m->mp, &ex_mountpoints);
fs/namespace.c
989
struct mountpoint *mp;
fs/namespace.c
998
mp = mnt->mnt_mp;
fs/pnode.c
397
const struct mountpoint *mp)
fs/pnode.c
402
if (to->mnt.mnt_root != mp->m_dentry)
fs/pnode.c
634
struct mountpoint *mp;
fs/pnode.c
637
mp = p->mnt_mp;
fs/pnode.c
641
mnt_change_mountpoint(p, mp, m);
fs/pnode.h
56
void mnt_change_mountpoint(struct mount *parent, struct mountpoint *mp,
fs/pnode.h
64
const struct mountpoint *mp);
fs/xfs/libxfs/xfs_ag.c
100
spin_unlock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_ag.c
102
xfs_reinit_percpu_counters(mp);
fs/xfs/libxfs/xfs_ag.c
104
xfs_fs_mark_healthy(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/libxfs/xfs_ag.c
125
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
133
xfs_group_free(mp, agno, XG_TYPE_AG, xfs_perag_uninit);
fs/xfs/libxfs/xfs_ag.c
139
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
147
return mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_ag.c
148
return dblocks - (agno * mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_ag.c
153
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
156
return __xfs_ag_block_count(mp, agno, mp->m_sb.sb_agcount,
fs/xfs/libxfs/xfs_ag.c
157
mp->m_sb.sb_dblocks);
fs/xfs/libxfs/xfs_ag.c
163
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
174
bno = round_up(XFS_AGFL_BLOCK(mp) + 1, M_IGEO(mp)->cluster_align);
fs/xfs/libxfs/xfs_ag.c
175
*first = XFS_AGB_TO_AGINO(mp, bno);
fs/xfs/libxfs/xfs_ag.c
181
bno = round_down(eoag, M_IGEO(mp)->cluster_align);
fs/xfs/libxfs/xfs_ag.c
182
*last = XFS_AGB_TO_AGINO(mp, bno) - 1;
fs/xfs/libxfs/xfs_ag.c
187
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
192
return __xfs_agino_range(mp, xfs_ag_block_count(mp, agno), first, last);
fs/xfs/libxfs/xfs_ag.c
201
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
204
struct xfs_perag *pag = xfs_perag_grab(mp, prev_agcount - 1);
fs/xfs/libxfs/xfs_ag.c
208
pag_group(pag)->xg_block_count = __xfs_ag_block_count(mp,
fs/xfs/libxfs/xfs_ag.c
209
prev_agcount - 1, mp->m_sb.sb_agcount,
fs/xfs/libxfs/xfs_ag.c
210
mp->m_sb.sb_dblocks);
fs/xfs/libxfs/xfs_ag.c
211
__xfs_agino_range(mp, pag_group(pag)->xg_block_count, &pag->agino_min,
fs/xfs/libxfs/xfs_ag.c
219
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
245
pag_group(pag)->xg_block_count = __xfs_ag_block_count(mp, index, agcount,
fs/xfs/libxfs/xfs_ag.c
247
pag_group(pag)->xg_min_gbno = XFS_AGFL_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_ag.c
248
__xfs_agino_range(mp, pag_group(pag)->xg_block_count, &pag->agino_min,
fs/xfs/libxfs/xfs_ag.c
251
error = xfs_group_insert(mp, pag_group(pag), index, XG_TYPE_AG);
fs/xfs/libxfs/xfs_ag.c
266
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
279
error = xfs_perag_alloc(mp, index, new_agcount, dblocks);
fs/xfs/libxfs/xfs_ag.c
284
*maxagi = xfs_set_inode_alloc(mp, new_agcount);
fs/xfs/libxfs/xfs_ag.c
285
mp->m_ag_prealloc_blocks = xfs_prealloc_blocks(mp);
fs/xfs/libxfs/xfs_ag.c
289
xfs_free_perag_range(mp, orig_agcount, index);
fs/xfs/libxfs/xfs_ag.c
295
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
304
error = xfs_buf_get_uncached(mp->m_ddev_targp, numblks, &bp);
fs/xfs/libxfs/xfs_ag.c
320
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
324
xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno);
fs/xfs/libxfs/xfs_ag.c
330
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
337
arec = XFS_ALLOC_REC_ADDR(mp, XFS_BUF_TO_BLOCK(bp), 1);
fs/xfs/libxfs/xfs_ag.c
338
arec->ar_startblock = cpu_to_be32(mp->m_ag_prealloc_blocks);
fs/xfs/libxfs/xfs_ag.c
340
if (xfs_ag_contains_log(mp, id->agno)) {
fs/xfs/libxfs/xfs_ag.c
342
xfs_agblock_t start = XFS_FSB_TO_AGBNO(mp,
fs/xfs/libxfs/xfs_ag.c
343
mp->m_sb.sb_logstart);
fs/xfs/libxfs/xfs_ag.c
345
ASSERT(start >= mp->m_ag_prealloc_blocks);
fs/xfs/libxfs/xfs_ag.c
346
if (start != mp->m_ag_prealloc_blocks) {
fs/xfs/libxfs/xfs_ag.c
352
mp->m_ag_prealloc_blocks);
fs/xfs/libxfs/xfs_ag.c
368
be32_add_cpu(&arec->ar_startblock, mp->m_sb.sb_logblocks);
fs/xfs/libxfs/xfs_ag.c
386
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
390
xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 0, id->agno);
fs/xfs/libxfs/xfs_ag.c
391
xfs_freesp_init_recs(mp, bp, id);
fs/xfs/libxfs/xfs_ag.c
399
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
406
xfs_btree_init_buf(mp, bp, id->bc_ops, 0, 4, id->agno);
fs/xfs/libxfs/xfs_ag.c
419
rrec->rm_blockcount = cpu_to_be32(XFS_BNO_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
425
rrec->rm_startblock = cpu_to_be32(XFS_BNO_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
432
rrec->rm_startblock = cpu_to_be32(XFS_IBT_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
433
rrec->rm_blockcount = cpu_to_be32(XFS_RMAP_BLOCK(mp) -
fs/xfs/libxfs/xfs_ag.c
434
XFS_IBT_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
440
rrec->rm_startblock = cpu_to_be32(XFS_RMAP_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
446
if (xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_ag.c
448
rrec->rm_startblock = cpu_to_be32(xfs_refc_block(mp));
fs/xfs/libxfs/xfs_ag.c
45
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
456
if (xfs_ag_contains_log(mp, id->agno)) {
fs/xfs/libxfs/xfs_ag.c
460
XFS_FSB_TO_AGBNO(mp, mp->m_sb.sb_logstart));
fs/xfs/libxfs/xfs_ag.c
461
rrec->rm_blockcount = cpu_to_be32(mp->m_sb.sb_logblocks);
fs/xfs/libxfs/xfs_ag.c
478
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
484
xfs_sb_to_disk(dsb, &mp->m_sb);
fs/xfs/libxfs/xfs_ag.c
490
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
50
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/libxfs/xfs_ag.c
501
agf->agf_bno_root = cpu_to_be32(XFS_BNO_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
502
agf->agf_cnt_root = cpu_to_be32(XFS_CNT_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
505
if (xfs_has_rmapbt(mp)) {
fs/xfs/libxfs/xfs_ag.c
506
agf->agf_rmap_root = cpu_to_be32(XFS_RMAP_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
514
tmpsize = id->agsize - mp->m_ag_prealloc_blocks;
fs/xfs/libxfs/xfs_ag.c
517
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_ag.c
518
uuid_copy(&agf->agf_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_ag.c
519
if (xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_ag.c
521
xfs_refc_block(mp));
fs/xfs/libxfs/xfs_ag.c
526
if (xfs_ag_contains_log(mp, id->agno)) {
fs/xfs/libxfs/xfs_ag.c
527
int64_t logblocks = mp->m_sb.sb_logblocks;
fs/xfs/libxfs/xfs_ag.c
531
XFS_FSB_TO_AGBNO(mp, mp->m_sb.sb_logstart) - logblocks);
fs/xfs/libxfs/xfs_ag.c
537
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
545
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_ag.c
548
uuid_copy(&agfl->agfl_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_ag.c
552
for (bucket = 0; bucket < xfs_agfl_size(mp); bucket++)
fs/xfs/libxfs/xfs_ag.c
558
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
570
agi->agi_root = cpu_to_be32(XFS_IBT_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
575
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_ag.c
576
uuid_copy(&agi->agi_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_ag.c
577
if (xfs_has_finobt(mp)) {
fs/xfs/libxfs/xfs_ag.c
578
agi->agi_free_root = cpu_to_be32(XFS_FIBT_BLOCK(mp));
fs/xfs/libxfs/xfs_ag.c
583
if (xfs_has_inobtcounts(mp)) {
fs/xfs/libxfs/xfs_ag.c
585
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_ag.c
590
typedef void (*aghdr_init_work_f)(struct xfs_mount *mp, struct xfs_buf *bp,
fs/xfs/libxfs/xfs_ag.c
594
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
602
error = xfs_get_aghdr_buf(mp, id->daddr, id->numblks, &bp, ops);
fs/xfs/libxfs/xfs_ag.c
606
(*work)(mp, bp, id);
fs/xfs/libxfs/xfs_ag.c
634
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
64
pag = xfs_perag_get(mp, index);
fs/xfs/libxfs/xfs_ag.c
640
.daddr = XFS_AG_DADDR(mp, id->agno, XFS_SB_DADDR),
fs/xfs/libxfs/xfs_ag.c
641
.numblks = XFS_FSS_TO_BB(mp, 1),
fs/xfs/libxfs/xfs_ag.c
647
.daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGF_DADDR(mp)),
fs/xfs/libxfs/xfs_ag.c
648
.numblks = XFS_FSS_TO_BB(mp, 1),
fs/xfs/libxfs/xfs_ag.c
654
.daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGFL_DADDR(mp)),
fs/xfs/libxfs/xfs_ag.c
655
.numblks = XFS_FSS_TO_BB(mp, 1),
fs/xfs/libxfs/xfs_ag.c
661
.daddr = XFS_AG_DADDR(mp, id->agno, XFS_AGI_DADDR(mp)),
fs/xfs/libxfs/xfs_ag.c
662
.numblks = XFS_FSS_TO_BB(mp, 1),
fs/xfs/libxfs/xfs_ag.c
668
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_BNO_BLOCK(mp)),
fs/xfs/libxfs/xfs_ag.c
669
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
676
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_CNT_BLOCK(mp)),
fs/xfs/libxfs/xfs_ag.c
677
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
684
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_IBT_BLOCK(mp)),
fs/xfs/libxfs/xfs_ag.c
685
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
692
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_FIBT_BLOCK(mp)),
fs/xfs/libxfs/xfs_ag.c
693
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
697
.need_init = xfs_has_finobt(mp)
fs/xfs/libxfs/xfs_ag.c
700
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, XFS_RMAP_BLOCK(mp)),
fs/xfs/libxfs/xfs_ag.c
701
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
705
.need_init = xfs_has_rmapbt(mp)
fs/xfs/libxfs/xfs_ag.c
708
.daddr = XFS_AGB_TO_DADDR(mp, id->agno, xfs_refc_block(mp)),
fs/xfs/libxfs/xfs_ag.c
709
.numblks = BTOBB(mp->m_sb.sb_blocksize),
fs/xfs/libxfs/xfs_ag.c
713
.need_init = xfs_has_reflink(mp)
fs/xfs/libxfs/xfs_ag.c
723
id->nfree += id->agsize - mp->m_ag_prealloc_blocks;
fs/xfs/libxfs/xfs_ag.c
731
error = xfs_ag_init_hdr(mp, id, dp->work, dp->ops);
fs/xfs/libxfs/xfs_ag.c
744
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag.c
747
.mp = mp,
fs/xfs/libxfs/xfs_ag.c
761
ASSERT(pag_agno(pag) == mp->m_sb.sb_agcount - 1);
fs/xfs/libxfs/xfs_ag.c
775
if (XFS_IS_CORRUPT(mp, agf->agf_length != agi->agi_length)) {
fs/xfs/libxfs/xfs_ag.c
858
__xfs_agino_range(mp, pag_group(pag)->xg_block_count, &pag->agino_min,
fs/xfs/libxfs/xfs_ag.c
870
xfs_warn(mp, "Error %d reserving per-AG metadata reserve pool.", err2);
fs/xfs/libxfs/xfs_ag.c
871
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/libxfs/xfs_ag.c
877
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.c
887
nb_mod = do_div(nb_div, mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_ag.c
89
xfs_alert(mp, "AGF corruption. Please run xfs_repair.");
fs/xfs/libxfs/xfs_ag.c
891
nb = nb_div * mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_ag.c
895
nb = nb_div * mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_ag.c
898
delta = nb - mp->m_sb.sb_dblocks;
fs/xfs/libxfs/xfs_ag.c
90
xfs_fs_mark_sick(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/libxfs/xfs_ag.c
912
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag.c
918
ASSERT(pag_agno(pag) == mp->m_sb.sb_agcount - 1);
fs/xfs/libxfs/xfs_ag.c
958
__xfs_agino_range(mp, pag_group(pag)->xg_block_count, &pag->agino_min,
fs/xfs/libxfs/xfs_ag.c
96
spin_lock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_ag.h
136
int xfs_initialize_perag(struct xfs_mount *mp, xfs_agnumber_t orig_agcount,
fs/xfs/libxfs/xfs_ag.h
139
void xfs_free_perag_range(struct xfs_mount *mp, xfs_agnumber_t first_agno,
fs/xfs/libxfs/xfs_ag.h
141
int xfs_initialize_perag_data(struct xfs_mount *mp, xfs_agnumber_t agno);
fs/xfs/libxfs/xfs_ag.h
142
int xfs_update_last_ag_size(struct xfs_mount *mp, xfs_agnumber_t prev_agcount);
fs/xfs/libxfs/xfs_ag.h
147
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.h
150
return to_perag(xfs_group_get(mp, agno, XG_TYPE_AG));
fs/xfs/libxfs/xfs_ag.h
170
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.h
173
return to_perag(xfs_group_grab(mp, agno, XG_TYPE_AG));
fs/xfs/libxfs/xfs_ag.h
185
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.h
190
return to_perag(xfs_group_next_range(mp, pag ? pag_group(pag) : NULL,
fs/xfs/libxfs/xfs_ag.h
196
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.h
200
return xfs_perag_next_range(mp, pag, start_agno, mp->m_sb.sb_agcount - 1);
fs/xfs/libxfs/xfs_ag.h
205
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ag.h
208
return xfs_perag_next_from(mp, pag, 0);
fs/xfs/libxfs/xfs_ag.h
214
xfs_agblock_t xfs_ag_block_count(struct xfs_mount *mp, xfs_agnumber_t agno);
fs/xfs/libxfs/xfs_ag.h
215
void xfs_agino_range(struct xfs_mount *mp, xfs_agnumber_t agno,
fs/xfs/libxfs/xfs_ag.h
260
xfs_ag_contains_log(struct xfs_mount *mp, xfs_agnumber_t agno)
fs/xfs/libxfs/xfs_ag.h
262
return mp->m_sb.sb_logstart > 0 &&
fs/xfs/libxfs/xfs_ag.h
263
agno == XFS_FSB_TO_AGNO(mp, mp->m_sb.sb_logstart);
fs/xfs/libxfs/xfs_ag.h
274
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag.h
285
pag = xfs_perag_grab(mp, *agno);
fs/xfs/libxfs/xfs_ag.h
297
#define for_each_perag_wrap_range(mp, start_agno, restart_agno, wrap_agno, agno, pag) \
fs/xfs/libxfs/xfs_ag.h
298
for ((agno) = (start_agno), (pag) = xfs_perag_grab((mp), (agno)); \
fs/xfs/libxfs/xfs_ag.h
306
#define for_each_perag_wrap_at(mp, start_agno, wrap_agno, agno, pag) \
fs/xfs/libxfs/xfs_ag.h
307
for_each_perag_wrap_range((mp), (start_agno), 0, (wrap_agno), (agno), (pag))
fs/xfs/libxfs/xfs_ag.h
313
#define for_each_perag_wrap(mp, start_agno, agno, pag) \
fs/xfs/libxfs/xfs_ag.h
314
for_each_perag_wrap_at((mp), (start_agno), (mp)->m_sb.sb_agcount, \
fs/xfs/libxfs/xfs_ag.h
331
int xfs_ag_init_headers(struct xfs_mount *mp, struct aghdr_init_data *id);
fs/xfs/libxfs/xfs_ag.h
335
xfs_growfs_compute_deltas(struct xfs_mount *mp, xfs_rfsblock_t nb,
fs/xfs/libxfs/xfs_ag_resv.c
174
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag_resv.c
205
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_AG_RESV_FAIL))
fs/xfs/libxfs/xfs_ag_resv.c
208
error = xfs_dec_fdblocks(mp, hidden_space, true);
fs/xfs/libxfs/xfs_ag_resv.c
211
xfs_warn(mp,
fs/xfs/libxfs/xfs_ag_resv.c
224
mp->m_ag_max_usable -= ask;
fs/xfs/libxfs/xfs_ag_resv.c
241
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag_resv.c
251
error = xfs_refcountbt_calc_reserves(mp, tp, pag, &ask, &used);
fs/xfs/libxfs/xfs_ag_resv.c
271
mp->m_finobt_nores = true;
fs/xfs/libxfs/xfs_ag_resv.c
273
error = xfs_refcountbt_calc_reserves(mp, tp, pag, &ask,
fs/xfs/libxfs/xfs_ag_resv.c
291
error = xfs_rmapbt_calc_reserves(mp, tp, pag, &ask, &used);
fs/xfs/libxfs/xfs_ag_resv.c
73
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ag_resv.c
95
return avail < orig / 10 || avail < mp->m_agbtree_maxlevels ||
fs/xfs/libxfs/xfs_ag_resv.c
96
XFS_TEST_ERROR(mp, XFS_ERRTAG_AG_RESV_CRITICAL);
fs/xfs/libxfs/xfs_alloc.c
1027
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
115
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_alloc.c
117
return mp->m_sb.sb_agcount * (XFS_ALLOCBT_AGFL_RESERVE + 4);
fs/xfs/libxfs/xfs_alloc.c
1233
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
1259
error = xfs_trans_get_buf(args->tp, args->mp->m_ddev_targp,
fs/xfs/libxfs/xfs_alloc.c
1261
args->mp->m_bsize, 0, &bp);
fs/xfs/libxfs/xfs_alloc.c
1268
if (XFS_IS_CORRUPT(args->mp, fbno >= be32_to_cpu(agf->agf_length))) {
fs/xfs/libxfs/xfs_alloc.c
1334
bno_cur = xfs_bnobt_init_cursor(args->mp, args->tp, args->agbp,
fs/xfs/libxfs/xfs_alloc.c
1354
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
136
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_alloc.c
1396
cnt_cur = xfs_cntbt_init_cursor(args->mp, args->tp, args->agbp,
fs/xfs/libxfs/xfs_alloc.c
140
blocks = XFS_BB_TO_FSB(mp, XFS_FSS_TO_BB(mp, 4)); /* ag headers */
fs/xfs/libxfs/xfs_alloc.c
143
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_alloc.c
145
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_alloc.c
147
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_alloc.c
150
return mp->m_sb.sb_agblocks - blocks;
fs/xfs/libxfs/xfs_alloc.c
1636
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
1686
args->max_agbno = args->mp->m_sb.sb_agblocks - 1;
fs/xfs/libxfs/xfs_alloc.c
1812
cnt_cur = xfs_cntbt_init_cursor(args->mp, args->tp, args->agbp,
fs/xfs/libxfs/xfs_alloc.c
1851
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
1896
if (XFS_IS_CORRUPT(args->mp,
fs/xfs/libxfs/xfs_alloc.c
1922
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
1932
if (XFS_IS_CORRUPT(args->mp,
fs/xfs/libxfs/xfs_alloc.c
1952
if (XFS_IS_CORRUPT(args->mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
1992
if (XFS_IS_CORRUPT(args->mp, rlen > flen)) {
fs/xfs/libxfs/xfs_alloc.c
2000
bno_cur = xfs_bnobt_init_cursor(args->mp, args->tp, args->agbp,
fs/xfs/libxfs/xfs_alloc.c
2010
if (XFS_IS_CORRUPT(args->mp,
fs/xfs/libxfs/xfs_alloc.c
2047
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_alloc.c
2064
mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
2075
bno_cur = xfs_bnobt_init_cursor(mp, tp, agbp, pag);
fs/xfs/libxfs/xfs_alloc.c
2088
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2104
if (XFS_IS_CORRUPT(mp, ltbno + ltlen > bno)) {
fs/xfs/libxfs/xfs_alloc.c
2123
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2139
if (XFS_IS_CORRUPT(mp, bno + len > gtbno)) {
fs/xfs/libxfs/xfs_alloc.c
2149
cnt_cur = xfs_cntbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/libxfs/xfs_alloc.c
2160
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2167
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2177
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2184
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2194
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2204
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2221
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_alloc.c
2249
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2256
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2267
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2287
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2294
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2317
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2335
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_alloc.c
2344
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
2366
XFS_STATS_INC(mp, xs_freex);
fs/xfs/libxfs/xfs_alloc.c
2367
XFS_STATS_ADD(mp, xs_freeb, len);
fs/xfs/libxfs/xfs_alloc.c
2392
xfs_mount_t *mp) /* file system mount structure */
fs/xfs/libxfs/xfs_alloc.c
2394
mp->m_alloc_maxlevels = xfs_btree_compute_maxlevels(mp->m_alloc_mnr,
fs/xfs/libxfs/xfs_alloc.c
2395
(mp->m_sb.sb_agblocks + 1) / 2);
fs/xfs/libxfs/xfs_alloc.c
2396
ASSERT(mp->m_alloc_maxlevels <= xfs_allocbt_maxlevels_ondisk());
fs/xfs/libxfs/xfs_alloc.c
2446
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc.c
2455
ASSERT(mp->m_alloc_maxlevels > 0);
fs/xfs/libxfs/xfs_alloc.c
2479
min_free = min(bno_level + 1, mp->m_alloc_maxlevels) * 2 - 2;
fs/xfs/libxfs/xfs_alloc.c
2481
min_free += min(cnt_level + 1, mp->m_alloc_maxlevels) * 2 - 2;
fs/xfs/libxfs/xfs_alloc.c
2483
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_alloc.c
2484
min_free += min(rmap_level + 1, mp->m_rmap_maxlevels) * 2 - 2;
fs/xfs/libxfs/xfs_alloc.c
2558
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc.c
2564
int agfl_size = xfs_agfl_size(mp);
fs/xfs/libxfs/xfs_alloc.c
2607
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
2611
trace_xfs_agfl_reset(mp, agf, 0, _RET_IP_);
fs/xfs/libxfs/xfs_alloc.c
2613
xfs_warn(mp,
fs/xfs/libxfs/xfs_alloc.c
2619
agf->agf_fllast = cpu_to_be32(xfs_agfl_size(mp) - 1);
fs/xfs/libxfs/xfs_alloc.c
2643
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
2654
if (XFS_IS_CORRUPT(mp, !xfs_verify_rtbext(mp, bno, len)))
fs/xfs/libxfs/xfs_alloc.c
2657
if (XFS_IS_CORRUPT(mp, !xfs_verify_fsbext(mp, bno, len)))
fs/xfs/libxfs/xfs_alloc.c
272
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_alloc.c
274
xfs_warn(mp,
fs/xfs/libxfs/xfs_alloc.c
277
xfs_warn(mp,
fs/xfs/libxfs/xfs_alloc.c
2796
cnt_cur = xfs_cntbt_init_cursor(args->mp, args->tp, agbp,
fs/xfs/libxfs/xfs_alloc.c
2830
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
2865
need = xfs_alloc_min_freelist(mp, pag);
fs/xfs/libxfs/xfs_alloc.c
2889
need = xfs_alloc_min_freelist(mp, pag);
fs/xfs/libxfs/xfs_alloc.c
2956
targs.mp = mp;
fs/xfs/libxfs/xfs_alloc.c
3040
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
3067
if (be32_to_cpu(agf->agf_flfirst) == xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
3141
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
3156
if (be32_to_cpu(agf->agf_fllast) == xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
3170
ASSERT(be32_to_cpu(agf->agf_flcount) <= xfs_agfl_size(mp));
fs/xfs/libxfs/xfs_alloc.c
3195
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
3209
if (length != mp->m_sb.sb_agblocks) {
fs/xfs/libxfs/xfs_alloc.c
3215
if (bp->b_pag && seqno != mp->m_sb.sb_agcount - 1)
fs/xfs/libxfs/xfs_alloc.c
3219
if (length > mp->m_sb.sb_agblocks)
fs/xfs/libxfs/xfs_alloc.c
3243
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
3249
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_alloc.c
3250
if (!uuid_equal(&agf->agf_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_alloc.c
3252
if (!xfs_log_check_lsn(mp, be64_to_cpu(agf->agf_lsn)))
fs/xfs/libxfs/xfs_alloc.c
3270
if (be32_to_cpu(agf->agf_flfirst) >= xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
3272
if (be32_to_cpu(agf->agf_fllast) >= xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
3274
if (be32_to_cpu(agf->agf_flcount) > xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
3283
be32_to_cpu(agf->agf_bno_level) > mp->m_alloc_maxlevels ||
fs/xfs/libxfs/xfs_alloc.c
3284
be32_to_cpu(agf->agf_cnt_level) > mp->m_alloc_maxlevels)
fs/xfs/libxfs/xfs_alloc.c
3287
if (xfs_has_lazysbcount(mp) &&
fs/xfs/libxfs/xfs_alloc.c
3291
if (xfs_has_rmapbt(mp)) {
fs/xfs/libxfs/xfs_alloc.c
3296
be32_to_cpu(agf->agf_rmap_level) > mp->m_rmap_maxlevels)
fs/xfs/libxfs/xfs_alloc.c
3300
if (xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_alloc.c
3305
be32_to_cpu(agf->agf_refcount_level) > mp->m_refc_maxlevels)
fs/xfs/libxfs/xfs_alloc.c
3316
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
3319
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_alloc.c
3324
if (fa || XFS_TEST_ERROR(mp, XFS_ERRTAG_ALLOC_READ_AGF))
fs/xfs/libxfs/xfs_alloc.c
3333
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
3344
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_alloc.c
3371
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_alloc.c
3376
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_alloc.c
3377
XFS_AG_DADDR(mp, pag_agno(pag), XFS_AGF_DADDR(mp)),
fs/xfs/libxfs/xfs_alloc.c
3378
XFS_FSS_TO_BB(mp, 1), flags, agfbpp, &xfs_agf_buf_ops);
fs/xfs/libxfs/xfs_alloc.c
3400
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_alloc.c
3427
if (xfs_agfl_needs_reset(mp, agf))
fs/xfs/libxfs/xfs_alloc.c
3440
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_alloc.c
3443
atomic64_add(allocbt_blks, &mp->m_allocbt_blks);
fs/xfs/libxfs/xfs_alloc.c
3500
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3514
agsize = mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_alloc.c
3525
ASSERT(XFS_FSB_TO_AGNO(mp, target) < mp->m_sb.sb_agcount);
fs/xfs/libxfs/xfs_alloc.c
3526
ASSERT(XFS_FSB_TO_AGBNO(mp, target) < agsize);
fs/xfs/libxfs/xfs_alloc.c
3531
if (XFS_FSB_TO_AGNO(mp, target) >= mp->m_sb.sb_agcount ||
fs/xfs/libxfs/xfs_alloc.c
3532
XFS_FSB_TO_AGBNO(mp, target) >= agsize ||
fs/xfs/libxfs/xfs_alloc.c
3563
args->pag = xfs_perag_get(args->mp, args->agno);
fs/xfs/libxfs/xfs_alloc.c
3598
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3639
XFS_AG_CHECK_DADDR(mp, XFS_FSB_TO_DADDR(mp, args->fsbno), args->len);
fs/xfs/libxfs/xfs_alloc.c
3661
XFS_STATS_INC(mp, xs_allocx);
fs/xfs/libxfs/xfs_alloc.c
3662
XFS_STATS_ADD(mp, xs_allocb, args->len);
fs/xfs/libxfs/xfs_alloc.c
3739
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3747
for_each_perag_wrap_range(mp, start_agno, restart_agno,
fs/xfs/libxfs/xfs_alloc.c
3748
mp->m_sb.sb_agcount, agno, args->pag) {
fs/xfs/libxfs/xfs_alloc.c
3808
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3831
xfs_is_inode32(mp)) {
fs/xfs/libxfs/xfs_alloc.c
3832
target = XFS_AGB_TO_FSB(mp,
fs/xfs/libxfs/xfs_alloc.c
3833
((mp->m_agfrotor / rotorstep) %
fs/xfs/libxfs/xfs_alloc.c
3834
mp->m_sb.sb_agcount), 0);
fs/xfs/libxfs/xfs_alloc.c
3838
start_agno = max(minimum_agno, XFS_FSB_TO_AGNO(mp, target));
fs/xfs/libxfs/xfs_alloc.c
3840
XFS_FSB_TO_AGBNO(mp, target), alloc_flags);
fs/xfs/libxfs/xfs_alloc.c
3844
mp->m_agfrotor = (mp->m_agfrotor + 1) %
fs/xfs/libxfs/xfs_alloc.c
3845
(mp->m_sb.sb_agcount * rotorstep);
fs/xfs/libxfs/xfs_alloc.c
3847
mp->m_agfrotor = (args->agno * rotorstep + 1) %
fs/xfs/libxfs/xfs_alloc.c
3848
(mp->m_sb.sb_agcount * rotorstep);
fs/xfs/libxfs/xfs_alloc.c
3864
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3884
start_agno = max(minimum_agno, XFS_FSB_TO_AGNO(mp, target));
fs/xfs/libxfs/xfs_alloc.c
3886
XFS_FSB_TO_AGBNO(mp, target), alloc_flags);
fs/xfs/libxfs/xfs_alloc.c
3899
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3904
ASSERT(pag_agno(args->pag) == XFS_FSB_TO_AGNO(mp, target));
fs/xfs/libxfs/xfs_alloc.c
3906
args->agno = XFS_FSB_TO_AGNO(mp, target);
fs/xfs/libxfs/xfs_alloc.c
3907
args->agbno = XFS_FSB_TO_AGBNO(mp, target);
fs/xfs/libxfs/xfs_alloc.c
3936
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_alloc.c
3943
ASSERT(pag_agno(args->pag) == XFS_FSB_TO_AGNO(mp, target));
fs/xfs/libxfs/xfs_alloc.c
3945
args->agno = XFS_FSB_TO_AGNO(mp, target);
fs/xfs/libxfs/xfs_alloc.c
3946
args->agbno = XFS_FSB_TO_AGBNO(mp, target);
fs/xfs/libxfs/xfs_alloc.c
3958
args->pag = xfs_perag_grab(mp, args->agno);
fs/xfs/libxfs/xfs_alloc.c
3979
args.mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
3987
if (args.agno >= args.mp->m_sb.sb_agcount)
fs/xfs/libxfs/xfs_alloc.c
4013
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_alloc.c
4022
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_FREE_EXTENT))
fs/xfs/libxfs/xfs_alloc.c
4034
if (XFS_IS_CORRUPT(mp, agbno >= mp->m_sb.sb_agblocks)) {
fs/xfs/libxfs/xfs_alloc.c
4041
if (XFS_IS_CORRUPT(mp, agbno + len > be32_to_cpu(agf->agf_length))) {
fs/xfs/libxfs/xfs_alloc.c
4146
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc.c
4165
error = walk_fn(mp, be32_to_cpu(agfl_bno[i]), priv);
fs/xfs/libxfs/xfs_alloc.c
4170
if (++i == xfs_agfl_size(mp))
fs/xfs/libxfs/xfs_alloc.c
46
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_alloc.c
48
unsigned int size = mp->m_sb.sb_sectsize;
fs/xfs/libxfs/xfs_alloc.c
50
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_alloc.c
58
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_alloc.c
582
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_alloc.c
585
mp = cnt_cur->bc_mp;
fs/xfs/libxfs/xfs_alloc.c
594
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_alloc.c
60
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_alloc.c
605
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
61
return XFS_RMAP_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
617
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_alloc.c
62
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_alloc.c
628
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
63
return XFS_FIBT_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
64
return XFS_IBT_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
642
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_alloc.c
681
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
69
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_alloc.c
691
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_alloc.c
697
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
705
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_alloc.c
71
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_alloc.c
711
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
72
return xfs_refc_block(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
725
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
73
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_alloc.c
74
return XFS_RMAP_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
742
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_alloc.c
748
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_alloc.c
75
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_alloc.c
76
return XFS_FIBT_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
77
return XFS_IBT_BLOCK(mp) + 1;
fs/xfs/libxfs/xfs_alloc.c
783
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
788
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_alloc.c
793
if (!uuid_equal(&agfl->agfl_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_alloc.c
804
for (i = 0; i < xfs_agfl_size(mp); i++) {
fs/xfs/libxfs/xfs_alloc.c
806
be32_to_cpu(agfl_bno[i]) >= mp->m_sb.sb_agblocks)
fs/xfs/libxfs/xfs_alloc.c
810
if (!xfs_log_check_lsn(mp, be64_to_cpu(XFS_BUF_TO_AGFL(bp)->agfl_lsn)))
fs/xfs/libxfs/xfs_alloc.c
819
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
828
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_alloc.c
844
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc.c
849
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_alloc.c
881
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_alloc.c
885
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_alloc.c
886
XFS_AG_DADDR(mp, pag_agno(pag), XFS_AGFL_DADDR(mp)),
fs/xfs/libxfs/xfs_alloc.c
887
XFS_FSS_TO_BB(mp, 1), 0, &bp, &xfs_agfl_buf_ops);
fs/xfs/libxfs/xfs_alloc.c
964
acur->cnt = xfs_cntbt_init_cursor(args->mp, args->tp,
fs/xfs/libxfs/xfs_alloc.c
974
acur->bnolt = xfs_bnobt_init_cursor(args->mp, args->tp,
fs/xfs/libxfs/xfs_alloc.c
977
acur->bnogt = xfs_bnobt_init_cursor(args->mp, args->tp,
fs/xfs/libxfs/xfs_alloc.h
17
unsigned int xfs_agfl_size(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_alloc.h
202
xfs_extlen_t xfs_prealloc_blocks(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_alloc.h
219
typedef int (*xfs_agfl_walk_fn)(struct xfs_mount *mp, xfs_agblock_t bno,
fs/xfs/libxfs/xfs_alloc.h
221
int xfs_agfl_walk(struct xfs_mount *mp, struct xfs_agf *agf,
fs/xfs/libxfs/xfs_alloc.h
36
struct xfs_mount *mp; /* file system mount point */
fs/xfs/libxfs/xfs_alloc.h
69
unsigned int xfs_alloc_set_aside(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_alloc.h
70
unsigned int xfs_alloc_ag_max_usable(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_alloc.h
74
unsigned int xfs_alloc_min_freelist(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc.h
90
struct xfs_mount *mp); /* file system mount structure */
fs/xfs/libxfs/xfs_alloc_btree.c
246
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_alloc_btree.c
255
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_alloc_btree.c
293
} else if (level >= mp->m_alloc_maxlevels)
fs/xfs/libxfs/xfs_alloc_btree.c
296
return xfs_btree_agblock_verify(bp, mp->m_alloc_mxr[level != 0]);
fs/xfs/libxfs/xfs_alloc_btree.c
478
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.c
485
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_bnobt_ops,
fs/xfs/libxfs/xfs_alloc_btree.c
486
mp->m_alloc_maxlevels, xfs_allocbt_cur_cache);
fs/xfs/libxfs/xfs_alloc_btree.c
504
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.c
511
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_cntbt_ops,
fs/xfs/libxfs/xfs_alloc_btree.c
512
mp->m_alloc_maxlevels, xfs_allocbt_cur_cache);
fs/xfs/libxfs/xfs_alloc_btree.c
566
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.c
570
blocklen -= XFS_ALLOC_BLOCK_LEN(mp);
fs/xfs/libxfs/xfs_alloc_btree.c
596
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.c
599
return xfs_btree_calc_size(mp->m_alloc_mnr, len);
fs/xfs/libxfs/xfs_alloc_btree.h
22
#define XFS_ALLOC_BLOCK_LEN(mp) \
fs/xfs/libxfs/xfs_alloc_btree.h
23
(xfs_has_crc(((mp))) ? \
fs/xfs/libxfs/xfs_alloc_btree.h
31
#define XFS_ALLOC_REC_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_alloc_btree.h
34
XFS_ALLOC_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_alloc_btree.h
37
#define XFS_ALLOC_KEY_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_alloc_btree.h
40
XFS_ALLOC_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_alloc_btree.h
43
#define XFS_ALLOC_PTR_ADDR(mp, block, index, maxrecs) \
fs/xfs/libxfs/xfs_alloc_btree.h
46
XFS_ALLOC_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_alloc_btree.h
50
struct xfs_btree_cur *xfs_bnobt_init_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.h
53
struct xfs_btree_cur *xfs_cntbt_init_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_alloc_btree.h
56
unsigned int xfs_allocbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_alloc_btree.h
58
extern xfs_extlen_t xfs_allocbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr.c
1005
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/libxfs/xfs_attr.c
1007
blks = XFS_ADDAFORK_SPACE_RES(mp);
fs/xfs/libxfs/xfs_attr.c
1009
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_addafork, blks, 0,
fs/xfs/libxfs/xfs_attr.c
1137
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr.c
1149
XFS_STATS_INC(mp, xs_attr_set);
fs/xfs/libxfs/xfs_attr.c
1167
rmt_blks = xfs_attr3_rmt_blocks(mp, args->valuelen);
fs/xfs/libxfs/xfs_attr.c
1173
XFS_STATS_INC(mp, xs_attr_remove);
fs/xfs/libxfs/xfs_attr.c
1174
rmt_blks = xfs_attr3_max_rmt_blocks(mp);
fs/xfs/libxfs/xfs_attr.c
1175
tres = M_RES(mp)->tr_attrrm;
fs/xfs/libxfs/xfs_attr.c
1176
total = XFS_ATTRRM_SPACE_RES(mp);
fs/xfs/libxfs/xfs_attr.c
1235
if (xfs_has_wsync(mp))
fs/xfs/libxfs/xfs_attr.c
302
struct xfs_mount *mp = args->dp->i_mount;
fs/xfs/libxfs/xfs_attr.c
311
nblks = XFS_DAENTER_SPACE_RES(mp, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_attr.c
322
uint dblocks = xfs_attr3_rmt_blocks(mp, args->valuelen);
fs/xfs/libxfs/xfs_attr.c
324
nblks += XFS_NEXTENTADD_SPACE_RES(mp, dblocks, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_attr.c
335
struct xfs_mount *mp = args->dp->i_mount;
fs/xfs/libxfs/xfs_attr.c
337
.tr_logres = M_RES(mp)->tr_attrsetm.tr_logres +
fs/xfs/libxfs/xfs_attr.c
338
M_RES(mp)->tr_attrsetrt.tr_logres * args->total,
fs/xfs/libxfs/xfs_attr.c
420
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr.c
430
return xfs_parent_hashattr(mp, name, namelen, value, valuelen);
fs/xfs/libxfs/xfs_attr.c
999
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_attr.h
634
xfs_dahash_t xfs_attr_hashval(struct xfs_mount *mp, unsigned int attr_flags,
fs/xfs/libxfs/xfs_attr_leaf.c
1111
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
1114
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &leafhdr, leaf);
fs/xfs/libxfs/xfs_attr_leaf.c
1306
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
1314
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_ATTR_LEAF_TO_NODE)) {
fs/xfs/libxfs/xfs_attr_leaf.c
1343
xfs_da3_node_hdr_from_disk(mp, &icnodehdr, node);
fs/xfs/libxfs/xfs_attr_leaf.c
1377
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
1395
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_attr_leaf.c
1402
uuid_copy(&hdr3->uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_attr_leaf.c
1598
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_attr_leaf.c
1625
mp = args->trans->t_mountp;
fs/xfs/libxfs/xfs_attr_leaf.c
1679
args->rmtblkcnt = xfs_attr3_rmt_blocks(mp, args->valuelen);
fs/xfs/libxfs/xfs_attr_leaf.c
1834
struct xfs_mount *mp = leaf1_bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
1836
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr1, leaf1_bp->b_addr);
fs/xfs/libxfs/xfs_attr_leaf.c
1837
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr2, leaf2_bp->b_addr);
fs/xfs/libxfs/xfs_attr_leaf.c
2853
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
2855
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr, bp->b_addr);
fs/xfs/libxfs/xfs_attr_leaf.c
297
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_leaf.c
319
if (nameidx < leafhdr->firstused || nameidx >= mp->m_attr_geo->blksize)
fs/xfs/libxfs/xfs_attr_leaf.c
372
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
382
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr, leaf);
fs/xfs/libxfs/xfs_attr_leaf.c
392
if (ichdr.firstused > mp->m_attr_geo->blksize)
fs/xfs/libxfs/xfs_attr_leaf.c
410
buf_end = (char *)bp->b_addr + mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_leaf.c
412
fa = xfs_attr3_leaf_verify_entry(mp, buf_end, leaf, &ichdr,
fs/xfs/libxfs/xfs_attr_leaf.c
428
if (ichdr.freemap[i].base > mp->m_attr_geo->blksize)
fs/xfs/libxfs/xfs_attr_leaf.c
432
if (ichdr.freemap[i].size > mp->m_attr_geo->blksize)
fs/xfs/libxfs/xfs_attr_leaf.c
441
if (end > mp->m_attr_geo->blksize)
fs/xfs/libxfs/xfs_attr_leaf.c
457
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
459
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_attr_leaf.c
477
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
488
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_leaf.c
507
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
510
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_attr_leaf.c
696
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
705
if (bytes > XFS_LITINO(mp))
fs/xfs/libxfs/xfs_attr_leaf.c
709
offset = (XFS_LITINO(mp) - bytes) >> 3;
fs/xfs/libxfs/xfs_attr_leaf.c
772
maxforkoff = XFS_LITINO(mp) - xfs_bmdr_space_calc(MINABTPTRS);
fs/xfs/libxfs/xfs_attr_leaf.c
788
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_leaf.c
791
if (mp->m_sb.sb_features2 & XFS_SB_VERSION2_ATTR2BIT)
fs/xfs/libxfs/xfs_attr_leaf.c
794
spin_lock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_attr_leaf.c
795
xfs_add_attr2(mp);
fs/xfs/libxfs/xfs_attr_leaf.c
796
spin_unlock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_attr_leaf.c
893
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
919
xfs_sbversion_add_attr2(mp, args->trans);
fs/xfs/libxfs/xfs_attr_leaf.c
946
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_leaf.c
984
!xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_attr_leaf.c
993
xfs_has_parent(mp));
fs/xfs/libxfs/xfs_attr_leaf.c
998
xfs_sbversion_add_attr2(mp, args->trans);
fs/xfs/libxfs/xfs_attr_remote.c
105
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_remote.c
114
if (!uuid_equal(&rmt->rm_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_attr_remote.c
118
if (be32_to_cpu(rmt->rm_bytes) > mp->m_attr_geo->blksize - sizeof(*rmt))
fs/xfs/libxfs/xfs_attr_remote.c
135
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_remote.c
139
unsigned int blksize = mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_remote.c
142
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_remote.c
156
*failaddr = xfs_attr3_rmt_verify(mp, bp, ptr, bno);
fs/xfs/libxfs/xfs_attr_remote.c
199
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_attr_remote.c
201
unsigned int blksize = mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_remote.c
207
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_remote.c
218
fa = xfs_attr3_rmt_verify(mp, bp, ptr, bno);
fs/xfs/libxfs/xfs_attr_remote.c
253
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_remote.c
262
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_remote.c
268
uuid_copy(&rmt->rm_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_attr_remote.c
292
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_remote.c
303
unsigned int blksize = mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_remote.c
309
unsigned int byte_cnt = xfs_attr3_rmt_buf_space(mp);
fs/xfs/libxfs/xfs_attr_remote.c
313
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_attr_remote.c
316
xfs_alert(mp,
fs/xfs/libxfs/xfs_attr_remote.c
342
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_remote.c
352
unsigned int blksize = mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_remote.c
358
unsigned int byte_cnt = xfs_attr3_rmt_buf_space(mp);
fs/xfs/libxfs/xfs_attr_remote.c
361
hdr_size = xfs_attr3_rmt_hdr_set(mp, dst, ino, *offset,
fs/xfs/libxfs/xfs_attr_remote.c
400
struct xfs_mount *mp = args->dp->i_mount;
fs/xfs/libxfs/xfs_attr_remote.c
432
dblkno = XFS_FSB_TO_DADDR(mp, map[i].br_startblock);
fs/xfs/libxfs/xfs_attr_remote.c
433
dblkcnt = XFS_FSB_TO_BB(mp, map[i].br_blockcount);
fs/xfs/libxfs/xfs_attr_remote.c
434
error = xfs_buf_read(mp->m_ddev_targp, dblkno, dblkcnt,
fs/xfs/libxfs/xfs_attr_remote.c
448
error = xfs_attr_rmtval_copyout(mp, bp, args->dp,
fs/xfs/libxfs/xfs_attr_remote.c
472
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_remote.c
482
blkcnt = xfs_attr3_rmt_blocks(mp, args->rmtvaluelen);
fs/xfs/libxfs/xfs_attr_remote.c
49
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_attr_remote.c
499
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_attr_remote.c
51
unsigned int blocksize = mp->m_attr_geo->blksize;
fs/xfs/libxfs/xfs_attr_remote.c
53
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_remote.c
535
dblkno = XFS_FSB_TO_DADDR(mp, map.br_startblock),
fs/xfs/libxfs/xfs_attr_remote.c
536
dblkcnt = XFS_FSB_TO_BB(mp, map.br_blockcount);
fs/xfs/libxfs/xfs_attr_remote.c
538
error = xfs_buf_get(mp->m_ddev_targp, dblkno, dblkcnt, &bp);
fs/xfs/libxfs/xfs_attr_remote.c
543
xfs_attr_rmtval_copyin(mp, bp, args->owner, &offset, &valuelen,
fs/xfs/libxfs/xfs_attr_remote.c
567
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_attr_remote.c
573
if (XFS_IS_CORRUPT(mp, map->br_startblock == DELAYSTARTBLOCK) ||
fs/xfs/libxfs/xfs_attr_remote.c
574
XFS_IS_CORRUPT(mp, map->br_startblock == HOLESTARTBLOCK)) {
fs/xfs/libxfs/xfs_attr_remote.c
579
error = xfs_buf_incore(mp->m_ddev_targp,
fs/xfs/libxfs/xfs_attr_remote.c
580
XFS_FSB_TO_DADDR(mp, map->br_startblock),
fs/xfs/libxfs/xfs_attr_remote.c
581
XFS_FSB_TO_BB(mp, map->br_blockcount),
fs/xfs/libxfs/xfs_attr_remote.c
62
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_attr_remote.c
69
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_attr_remote.c
70
return howmany(attrlen, xfs_attr3_rmt_buf_space(mp));
fs/xfs/libxfs/xfs_attr_remote.c
72
return XFS_B_TO_FSB(mp, attrlen);
fs/xfs/libxfs/xfs_attr_remote.h
12
static inline unsigned int xfs_attr3_max_rmt_blocks(struct xfs_mount *mp)
fs/xfs/libxfs/xfs_attr_remote.h
14
return xfs_attr3_rmt_blocks(mp, XFS_XATTR_SIZE_MAX);
fs/xfs/libxfs/xfs_attr_remote.h
9
unsigned int xfs_attr3_rmt_blocks(struct xfs_mount *mp, unsigned int attrlen);
fs/xfs/libxfs/xfs_bmap.c
103
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_bmap.c
1034
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_bmap.c
1040
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/libxfs/xfs_bmap.c
105
if (mp->m_sb.sb_inodesize == 256)
fs/xfs/libxfs/xfs_bmap.c
106
return XFS_LITINO(mp) - xfs_bmdr_space_calc(MINABTPTRS);
fs/xfs/libxfs/xfs_bmap.c
1068
if (!xfs_has_attr(mp)) {
fs/xfs/libxfs/xfs_bmap.c
1071
spin_lock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_bmap.c
1072
if (!xfs_has_attr(mp)) {
fs/xfs/libxfs/xfs_bmap.c
1073
xfs_add_attr(mp);
fs/xfs/libxfs/xfs_bmap.c
1074
xfs_add_attr2(mp);
fs/xfs/libxfs/xfs_bmap.c
1077
spin_unlock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_bmap.c
1101
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1111
xfs_warn(mp,
fs/xfs/libxfs/xfs_bmap.c
1114
xfs_warn(mp,
fs/xfs/libxfs/xfs_bmap.c
1130
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_bmap.c
1154
frp = xfs_bmbt_rec_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
1190
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1201
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
1208
if (XFS_IS_CORRUPT(mp, ir.loaded != ifp->if_nextents)) {
fs/xfs/libxfs/xfs_bmap.c
1424
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1426
if (xfs_ifork_is_realtime(ip, whichfork) && xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_bmap.c
1427
if (xfs_rtb_to_rgno(mp, left->br_startblock) !=
fs/xfs/libxfs/xfs_bmap.c
1428
xfs_rtb_to_rgno(mp, right->br_startblock))
fs/xfs/libxfs/xfs_bmap.c
1443
struct xfs_mount *mp = bma->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1463
XFS_STATS_INC(mp, xs_add_exlist);
fs/xfs/libxfs/xfs_bmap.c
1560
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1568
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1576
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1607
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1642
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1672
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
1680
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1716
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
1743
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
1751
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
179
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
1796
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
180
int maxrecs = mp->m_bmap_dmxr[0];
fs/xfs/libxfs/xfs_bmap.c
1834
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
1842
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
185
level < XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_bmap.c
191
return rval + XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) -
fs/xfs/libxfs/xfs_bmap.c
1923
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
1931
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
194
maxrecs = mp->m_bmap_dmxr[1];
fs/xfs/libxfs/xfs_bmap.c
1990
xfs_add_fdblocks(mp, da_old - da_new);
fs/xfs/libxfs/xfs_bmap.c
1992
error = xfs_dec_fdblocks(mp, da_new - da_old, true);
fs/xfs/libxfs/xfs_bmap.c
2026
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
2036
XFS_STATS_INC(mp, xs_add_exlist);
fs/xfs/libxfs/xfs_bmap.c
2129
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2136
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2143
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2150
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2157
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2186
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2193
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2200
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2232
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2239
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2246
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2273
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2307
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
233
struct xfs_mount *mp, /* file system mount point */
fs/xfs/libxfs/xfs_bmap.c
2345
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2356
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2387
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
241
if (!xfs_verify_fsbno(mp, fsbno))
fs/xfs/libxfs/xfs_bmap.c
2424
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
243
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_bmap.c
2435
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
244
XFS_FSB_TO_DADDR(mp, fsbno), mp->m_bsize, 0, &bp,
fs/xfs/libxfs/xfs_bmap.c
2442
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2479
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2492
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2505
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
2513
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2535
xfs_rmap_convert_extent(mp, tp, ip, whichfork, new);
fs/xfs/libxfs/xfs_bmap.c
2579
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
2592
XFS_STATS_INC(mp, xs_add_exlist);
fs/xfs/libxfs/xfs_bmap.c
2661
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2669
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2677
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2707
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2738
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2765
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
2773
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
2818
xfs_mount_t *mp,
fs/xfs/libxfs/xfs_bmap.c
287
xfs_mount_t *mp,
fs/xfs/libxfs/xfs_bmap.c
2937
if (rt && (temp = xfs_extlen_to_rtxmod(mp, align_alen))) {
fs/xfs/libxfs/xfs_bmap.c
2964
align_alen -= xfs_extlen_to_rtxmod(mp, align_alen);
fs/xfs/libxfs/xfs_bmap.c
299
dmxr = mp->m_bmap_dmxr[0];
fs/xfs/libxfs/xfs_bmap.c
2996
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
300
keyp = xfs_bmbt_key_addr(mp, block, i);
fs/xfs/libxfs/xfs_bmap.c
3000
if (!xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_bmap.c
3001
return x < mp->m_sb.sb_rblocks;
fs/xfs/libxfs/xfs_bmap.c
3003
return xfs_rtb_to_rgno(mp, x) == xfs_rtb_to_rgno(mp, y) &&
fs/xfs/libxfs/xfs_bmap.c
3004
xfs_rtb_to_rgno(mp, x) < mp->m_sb.sb_rgcount &&
fs/xfs/libxfs/xfs_bmap.c
3005
xfs_rtb_to_rtx(mp, x) < mp->m_sb.sb_rgextents;
fs/xfs/libxfs/xfs_bmap.c
3009
return XFS_FSB_TO_AGNO(mp, x) == XFS_FSB_TO_AGNO(mp, y) &&
fs/xfs/libxfs/xfs_bmap.c
3010
XFS_FSB_TO_AGNO(mp, x) < mp->m_sb.sb_agcount &&
fs/xfs/libxfs/xfs_bmap.c
3011
XFS_FSB_TO_AGBNO(mp, x) < mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_bmap.c
312
pp = xfs_bmap_broot_ptr_addr(mp, block, i, sz);
fs/xfs/libxfs/xfs_bmap.c
314
pp = xfs_bmbt_ptr_addr(mp, block, i, dmxr);
fs/xfs/libxfs/xfs_bmap.c
318
thispa = xfs_bmap_broot_ptr_addr(mp, block, j, sz);
fs/xfs/libxfs/xfs_bmap.c
320
thispa = xfs_bmbt_ptr_addr(mp, block, j, dmxr);
fs/xfs/libxfs/xfs_bmap.c
3203
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_bmap.c
3215
startag = XFS_FSB_TO_AGNO(mp, ap->blkno);
fs/xfs/libxfs/xfs_bmap.c
322
xfs_warn(mp, "%s: thispa(%d) == pp(%d) %lld",
fs/xfs/libxfs/xfs_bmap.c
3220
for_each_perag_wrap(mp, startag, agno, pag) {
fs/xfs/libxfs/xfs_bmap.c
325
xfs_err(mp, "%s: ptrs are equal in node\n",
fs/xfs/libxfs/xfs_bmap.c
327
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/libxfs/xfs_bmap.c
3293
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_bmap.c
3298
if (mp->m_swidth && xfs_has_swalloc(mp))
fs/xfs/libxfs/xfs_bmap.c
3299
stripe_align = mp->m_swidth;
fs/xfs/libxfs/xfs_bmap.c
3300
else if (mp->m_dalign)
fs/xfs/libxfs/xfs_bmap.c
3301
stripe_align = mp->m_dalign;
fs/xfs/libxfs/xfs_bmap.c
3313
if (xfs_bmap_extsize_align(mp, &ap->got, &ap->prev, align, 0,
fs/xfs/libxfs/xfs_bmap.c
3326
} else if (mp->m_sb.sb_blocksize >= PAGE_SIZE) {
fs/xfs/libxfs/xfs_bmap.c
3330
args->prod = PAGE_SIZE >> mp->m_sb.sb_blocklog;
fs/xfs/libxfs/xfs_bmap.c
3416
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_bmap.c
3446
args->pag = xfs_perag_get(mp, XFS_FSB_TO_AGNO(mp, ap->blkno));
fs/xfs/libxfs/xfs_bmap.c
346
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
3593
ap->blkno = XFS_INO_TO_FSB(args->mp, ap->ip->i_ino);
fs/xfs/libxfs/xfs_bmap.c
3630
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
3633
.mp = mp,
fs/xfs/libxfs/xfs_bmap.c
3655
args.maxlen = min(ap->length, mp->m_ag_max_usable);
fs/xfs/libxfs/xfs_bmap.c
3657
if (unlikely(XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAP_ALLOC_MINLEN_EXTENT)))
fs/xfs/libxfs/xfs_bmap.c
374
xfs_check_block(block, mp, 1, ifp->if_broot_bytes);
fs/xfs/libxfs/xfs_bmap.c
375
pp = xfs_bmap_broot_ptr_addr(mp, block, 1, ifp->if_broot_bytes);
fs/xfs/libxfs/xfs_bmap.c
379
ASSERT(XFS_FSB_TO_AGNO(mp, bno) < mp->m_sb.sb_agcount);
fs/xfs/libxfs/xfs_bmap.c
380
ASSERT(XFS_FSB_TO_AGBNO(mp, bno) < mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_bmap.c
3822
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
3842
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
3843
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
3848
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
3851
XFS_STATS_INC(mp, xs_blk_mapr);
fs/xfs/libxfs/xfs_bmap.c
389
bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
fs/xfs/libxfs/xfs_bmap.c
3900
struct xfs_mount *mp = bma->ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
392
error = xfs_bmap_read_buf(mp, NULL, bno, &bp);
fs/xfs/libxfs/xfs_bmap.c
3926
if (mp->m_dalign && bma->length >= mp->m_dalign) {
fs/xfs/libxfs/xfs_bmap.c
3956
bma->cur = xfs_bmbt_init_cursor(mp, bma->tp, bma->ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
407
xfs_check_block(block, mp, 0, 0);
fs/xfs/libxfs/xfs_bmap.c
408
pp = xfs_bmbt_ptr_addr(mp, block, 1, mp->m_bmap_dmxr[1]);
fs/xfs/libxfs/xfs_bmap.c
410
if (XFS_IS_CORRUPT(mp, !xfs_verify_fsbno(mp, bno))) {
fs/xfs/libxfs/xfs_bmap.c
4150
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4193
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
4194
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
4199
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
4202
XFS_STATS_INC(mp, xs_blk_mapw);
fs/xfs/libxfs/xfs_bmap.c
4363
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4364
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/libxfs/xfs_bmap.c
4377
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, 0, 0,
fs/xfs/libxfs/xfs_bmap.c
4451
XFS_STATS_ADD(mp, xs_xstrat_bytes, XFS_FSB_TO_B(mp, bma.length));
fs/xfs/libxfs/xfs_bmap.c
4452
XFS_STATS_INC(mp, xs_xstrat_quick);
fs/xfs/libxfs/xfs_bmap.c
448
ep = xfs_bmbt_rec_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
4521
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4538
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
4539
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
4544
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
455
nextp = xfs_bmbt_rec_addr(mp, block, j + 1);
fs/xfs/libxfs/xfs_bmap.c
4562
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
4670
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4680
XFS_STATS_INC(mp, xs_del_exlist);
fs/xfs/libxfs/xfs_bmap.c
476
bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
fs/xfs/libxfs/xfs_bmap.c
479
error = xfs_bmap_read_buf(mp, NULL, bno, &bp);
fs/xfs/libxfs/xfs_bmap.c
4793
rtxlen = xfs_blen_to_rtbxlen(mp, del->br_blockcount);
fs/xfs/libxfs/xfs_bmap.c
4795
xfs_zoned_add_available(mp, rtxlen);
fs/xfs/libxfs/xfs_bmap.c
4796
xfs_add_frextents(mp, rtxlen);
fs/xfs/libxfs/xfs_bmap.c
4801
xfs_add_fdblocks(mp, fdblocks);
fs/xfs/libxfs/xfs_bmap.c
4812
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4818
XFS_STATS_INC(mp, xs_del_exlist);
fs/xfs/libxfs/xfs_bmap.c
491
xfs_warn(mp, "%s: at error0", __func__);
fs/xfs/libxfs/xfs_bmap.c
4926
xfs_mount_t *mp; /* mount structure */
fs/xfs/libxfs/xfs_bmap.c
4936
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
4937
XFS_STATS_INC(mp, xs_del_exlist);
fs/xfs/libxfs/xfs_bmap.c
495
xfs_warn(mp, "%s: BAD after btree leaves for %llu extents",
fs/xfs/libxfs/xfs_bmap.c
497
xfs_err(mp, "%s: CORRUPTED BTREE OR SOMETHING", __func__);
fs/xfs/libxfs/xfs_bmap.c
4975
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
498
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/libxfs/xfs_bmap.c
5002
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5077
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5096
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5120
} else if (isrt && !xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_bmap.c
5188
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5202
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp))) {
fs/xfs/libxfs/xfs_bmap.c
5206
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
5221
XFS_STATS_INC(mp, xs_blk_unmap);
fs/xfs/libxfs/xfs_bmap.c
5234
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5277
mod = xfs_rtb_to_rtxoff(mp,
fs/xfs/libxfs/xfs_bmap.c
5326
mod = xfs_rtb_to_rtxoff(mp, del.br_startblock);
fs/xfs/libxfs/xfs_bmap.c
5328
xfs_extlen_t off = mp->m_sb.sb_rextsize - mod;
fs/xfs/libxfs/xfs_bmap.c
5540
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5565
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5573
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5582
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5621
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5633
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5663
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5672
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
5673
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
5678
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
568
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5688
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5694
if (XFS_IS_CORRUPT(mp, isnullstartblock(got.br_startblock))) {
fs/xfs/libxfs/xfs_bmap.c
57
xfs_mount_t *mp, /* file system mount structure */
fs/xfs/libxfs/xfs_bmap.c
5779
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5788
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
5789
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
5794
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
5804
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5819
if (XFS_IS_CORRUPT(mp, isnullstartblock(got.br_startblock))) {
fs/xfs/libxfs/xfs_bmap.c
5825
if (XFS_IS_CORRUPT(mp, stop_fsb > got.br_startoff)) {
fs/xfs/libxfs/xfs_bmap.c
586
ASSERT(xfs_bmbt_maxrecs(mp, ifp->if_broot_bytes, false) == 1);
fs/xfs/libxfs/xfs_bmap.c
588
pp = xfs_bmap_broot_ptr_addr(mp, rblock, 1, ifp->if_broot_bytes);
fs/xfs/libxfs/xfs_bmap.c
5886
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
5893
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(ifp)) ||
fs/xfs/libxfs/xfs_bmap.c
5894
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/libxfs/xfs_bmap.c
5899
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_bmap.c
591
if (XFS_IS_CORRUPT(cur->bc_mp, !xfs_verify_fsbno(mp, cbno))) {
fs/xfs/libxfs/xfs_bmap.c
5921
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
5925
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
5953
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_bmap.c
596
error = xfs_bmap_read_buf(mp, tp, cbno, &cbp);
fs/xfs/libxfs/xfs_bmap.c
5961
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_bmap.c
6088
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap.c
6093
if (!xfs_verify_fileext(mp, irec->br_startoff, irec->br_blockcount))
fs/xfs/libxfs/xfs_bmap.c
6097
if (!xfs_verify_rtbext(mp, irec->br_startblock,
fs/xfs/libxfs/xfs_bmap.c
6101
if (!xfs_verify_fsbext(mp, irec->br_startblock,
fs/xfs/libxfs/xfs_bmap.c
645
struct xfs_mount *mp; /* mount structure */
fs/xfs/libxfs/xfs_bmap.c
651
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
669
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/libxfs/xfs_bmap.c
678
args.mp = mp;
fs/xfs/libxfs/xfs_bmap.c
685
XFS_INO_TO_FSB(mp, ip->i_ino));
fs/xfs/libxfs/xfs_bmap.c
700
error = xfs_trans_get_buf(tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_bmap.c
701
XFS_FSB_TO_DADDR(mp, args.fsbno),
fs/xfs/libxfs/xfs_bmap.c
702
mp->m_bsize, 0, &abp);
fs/xfs/libxfs/xfs_bmap.c
715
arp = xfs_bmbt_rec_addr(mp, ablock, 1 + cnt);
fs/xfs/libxfs/xfs_bmap.c
725
kp = xfs_bmbt_key_addr(mp, block, 1);
fs/xfs/libxfs/xfs_bmap.c
726
arp = xfs_bmbt_rec_addr(mp, ablock, 1);
fs/xfs/libxfs/xfs_bmap.c
728
pp = xfs_bmbt_ptr_addr(mp, block, 1, xfs_bmbt_get_maxrecs(cur,
fs/xfs/libxfs/xfs_bmap.c
80
maxleafents = xfs_iext_max_nextents(xfs_has_large_extent_counts(mp),
fs/xfs/libxfs/xfs_bmap.c
820
args.mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
832
XFS_INO_TO_FSB(args.mp, ip->i_ino));
fs/xfs/libxfs/xfs_bmap.c
839
error = xfs_trans_get_buf(tp, args.mp->m_ddev_targp,
fs/xfs/libxfs/xfs_bmap.c
840
XFS_FSB_TO_DADDR(args.mp, args.fsbno),
fs/xfs/libxfs/xfs_bmap.c
841
args.mp->m_bsize, 0, &bp);
fs/xfs/libxfs/xfs_bmap.c
88
minleafrecs = mp->m_bmap_dmnr[0];
fs/xfs/libxfs/xfs_bmap.c
89
minnoderecs = mp->m_bmap_dmnr[1];
fs/xfs/libxfs/xfs_bmap.c
892
xfs_mount_t *mp; /* file system mount struct */
fs/xfs/libxfs/xfs_bmap.c
895
mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap.c
900
cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_bmap.c
905
if (XFS_IS_CORRUPT(mp, stat != 1)) {
fs/xfs/libxfs/xfs_bmap.c
97
mp->m_bm_maxlevels[whichfork] = level;
fs/xfs/libxfs/xfs_bmap.c
98
ASSERT(mp->m_bm_maxlevels[whichfork] <= xfs_bmbt_maxlevels_ondisk());
fs/xfs/libxfs/xfs_bmap.h
182
unsigned int xfs_bmap_compute_attr_offset(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_bmap.h
193
void xfs_bmap_compute_maxlevels(struct xfs_mount *mp, int whichfork);
fs/xfs/libxfs/xfs_bmap.h
274
xfs_failaddr_t xfs_bmap_validate_extent_raw(struct xfs_mount *mp, bool rtfile,
fs/xfs/libxfs/xfs_bmap_btree.c
145
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.c
157
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_bmap_btree.c
160
&mp->m_sb.sb_meta_uuid));
fs/xfs/libxfs/xfs_bmap_btree.c
171
fkp = xfs_bmbt_key_addr(mp, rblock, 1);
fs/xfs/libxfs/xfs_bmap_btree.c
173
fpp = xfs_bmap_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_bmap_btree.c
219
args.mp = cur->bc_mp;
fs/xfs/libxfs/xfs_bmap_btree.c
276
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_bmap_btree.c
279
xfs_fsblock_t fsbno = XFS_DADDR_TO_FSB(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_bmap_btree.c
398
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_bmap_btree.c
406
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_bmap_btree.c
424
if (level > max(mp->m_bm_maxlevels[0], mp->m_bm_maxlevels[1]))
fs/xfs/libxfs/xfs_bmap_btree.c
427
return xfs_btree_fsblock_verify(bp, mp->m_bmap_dmxr[level != 0]);
fs/xfs/libxfs/xfs_bmap_btree.c
509
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.c
518
sptr = xfs_bmap_broot_ptr_addr(mp, broot, 1, old_size);
fs/xfs/libxfs/xfs_bmap_btree.c
519
dptr = xfs_bmap_broot_ptr_addr(mp, broot, 1, new_size);
fs/xfs/libxfs/xfs_bmap_btree.c
546
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap_btree.c
56
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_bmap_btree.c
560
new_size = xfs_bmap_broot_space_calc(mp, new_numrecs);
fs/xfs/libxfs/xfs_bmap_btree.c
582
old_numrecs = xfs_bmbt_maxrecs(mp, old_size, false);
fs/xfs/libxfs/xfs_bmap_btree.c
586
xfs_bmbt_move_ptrs(mp, broot, old_size, new_size, old_numrecs);
fs/xfs/libxfs/xfs_bmap_btree.c
601
xfs_bmbt_move_ptrs(mp, ifp->if_broot, old_size, new_size, new_numrecs);
fs/xfs/libxfs/xfs_bmap_btree.c
654
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.c
670
maxlevels = mp->m_bm_maxlevels[XFS_DATA_FORK];
fs/xfs/libxfs/xfs_bmap_btree.c
673
maxlevels = mp->m_bm_maxlevels[whichfork];
fs/xfs/libxfs/xfs_bmap_btree.c
676
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_bmbt_ops, maxlevels,
fs/xfs/libxfs/xfs_bmap_btree.c
69
tkp = xfs_bmbt_key_addr(mp, rblock, 1);
fs/xfs/libxfs/xfs_bmap_btree.c
71
tpp = xfs_bmap_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_bmap_btree.c
749
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.c
753
blocklen -= xfs_bmbt_block_len(mp);
fs/xfs/libxfs/xfs_bmap_btree.c
837
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.c
840
return xfs_btree_calc_size(mp->m_bmap_dmnr, len);
fs/xfs/libxfs/xfs_bmap_btree.h
103
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
150
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
155
return xfs_bmbt_ptr_addr(mp, bb, i, xfs_bmbt_maxrecs(mp, sz, false));
fs/xfs/libxfs/xfs_bmap_btree.h
164
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
167
return xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
177
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
180
return xfs_bmap_broot_space_calc(mp, be16_to_cpu(bb->bb_numrecs));
fs/xfs/libxfs/xfs_bmap_btree.h
19
#define XFS_BM_MAXLEVELS(mp,w) ((mp)->m_bm_maxlevels[(w)])
fs/xfs/libxfs/xfs_bmap_btree.h
38
unsigned int xfs_bmbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_bmap_btree.h
50
extern unsigned long long xfs_bmbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
65
xfs_bmbt_block_len(struct xfs_mount *mp)
fs/xfs/libxfs/xfs_bmap_btree.h
67
return xfs_has_crc(mp) ?
fs/xfs/libxfs/xfs_bmap_btree.h
75
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
80
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
86
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_bmap_btree.h
91
((char *)block + xfs_bmbt_block_len(mp) +
fs/xfs/libxfs/xfs_bmap_btree.h
97
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
1013
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
1020
xfs_buf_readahead(mp->m_ddev_targp,
fs/xfs/libxfs/xfs_btree.c
1021
xfs_agbno_to_daddr(pag, left), mp->m_bsize,
fs/xfs/libxfs/xfs_btree.c
1027
xfs_buf_readahead(mp->m_ddev_targp,
fs/xfs/libxfs/xfs_btree.c
1028
xfs_agbno_to_daddr(pag, right), mp->m_bsize,
fs/xfs/libxfs/xfs_btree.c
1235
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
1243
bool crc = xfs_has_crc(mp);
fs/xfs/libxfs/xfs_btree.c
1244
__u32 magic = xfs_btree_magic(mp, ops);
fs/xfs/libxfs/xfs_btree.c
1256
uuid_copy(&buf->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_btree.c
1267
uuid_copy(&buf->bb_u.s.bb_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_btree.c
1275
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
1282
__xfs_btree_init_block(mp, block, ops, XFS_BUF_DADDR_NULL, level,
fs/xfs/libxfs/xfs_btree.c
1288
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
1295
__xfs_btree_init_block(mp, XFS_BUF_TO_BLOCK(bp), ops,
fs/xfs/libxfs/xfs_btree.c
131
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
133
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_btree.c
134
if (!uuid_equal(&block->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
1392
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
1402
error = xfs_trans_read_buf(mp, cur->bc_tp, xfs_btree_buftarg(cur), d,
fs/xfs/libxfs/xfs_btree.c
143
if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops))
fs/xfs/libxfs/xfs_btree.c
165
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
185
fsb = XFS_DADDR_TO_FSB(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_btree.c
186
fa = xfs_btree_check_fsblock_siblings(mp, fsb,
fs/xfs/libxfs/xfs_btree.c
189
fa = xfs_btree_check_fsblock_siblings(mp, fsb,
fs/xfs/libxfs/xfs_btree.c
233
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
238
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_btree.c
239
if (!uuid_equal(&block->bb_u.s.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
245
if (be32_to_cpu(block->bb_magic) != xfs_btree_magic(mp, cur->bc_ops))
fs/xfs/libxfs/xfs_btree.c
253
agbno = xfs_daddr_to_agbno(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_btree.c
304
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
308
if (XFS_IS_CORRUPT(mp, fa != NULL) ||
fs/xfs/libxfs/xfs_btree.c
309
XFS_TEST_ERROR(mp, xfs_btree_block_errtag(cur))) {
fs/xfs/libxfs/xfs_btree.c
424
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
426
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_btree.c
427
if (!xfs_log_check_lsn(mp, be64_to_cpu(block->bb_u.l.bb_lsn)))
fs/xfs/libxfs/xfs_btree.c
45
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
462
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
464
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_btree.c
465
if (!xfs_log_check_lsn(mp, be64_to_cpu(block->bb_u.s.bb_lsn)))
fs/xfs/libxfs/xfs_btree.c
48
int idx = xfs_has_crc(mp) ? 1 : 0;
fs/xfs/libxfs/xfs_btree.c
4835
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
4838
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_btree.c
4840
if (!uuid_equal(&block->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
4856
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
4868
fsb = XFS_DADDR_TO_FSB(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_btree.c
4869
fa = xfs_btree_check_fsblock_siblings(mp, fsb,
fs/xfs/libxfs/xfs_btree.c
4872
fa = xfs_btree_check_fsblock_siblings(mp, fsb,
fs/xfs/libxfs/xfs_btree.c
4917
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
4921
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_btree.c
4923
if (!uuid_equal(&block->bb_u.s.bb_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_btree.c
4943
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_btree.c
4955
agbno = xfs_daddr_to_agbno(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_btree.c
5599
.mp = cur->bc_mp,
fs/xfs/libxfs/xfs_btree.c
5636
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
5639
xfs_fsblock_t fsbno = XFS_DADDR_TO_FSB(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_btree.c
566
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
600
error = xfs_trans_read_buf(mp, tp,
fs/xfs/libxfs/xfs_btree.c
69
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.c
81
if (!xfs_verify_fsbno(mp, sibling))
fs/xfs/libxfs/xfs_btree.c
961
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_btree.c
967
xfs_buf_readahead(mp->m_ddev_targp, XFS_FSB_TO_DADDR(mp, left),
fs/xfs/libxfs/xfs_btree.c
968
mp->m_bsize, cur->bc_ops->buf_ops);
fs/xfs/libxfs/xfs_btree.c
973
xfs_buf_readahead(mp->m_ddev_targp, XFS_FSB_TO_DADDR(mp, right),
fs/xfs/libxfs/xfs_btree.c
974
mp->m_bsize, cur->bc_ops->buf_ops);
fs/xfs/libxfs/xfs_btree.h
393
void xfs_btree_init_buf(struct xfs_mount *mp, struct xfs_buf *bp,
fs/xfs/libxfs/xfs_btree.h
396
void xfs_btree_init_block(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.h
59
uint32_t xfs_btree_magic(struct xfs_mount *mp, const struct xfs_btree_ops *ops);
fs/xfs/libxfs/xfs_btree.h
671
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree.h
687
cur->bc_mp = mp;
fs/xfs/libxfs/xfs_btree_mem.c
102
xfs_btree_init_buf(mp, bp, ops, 0, 0, xfbt->owner);
fs/xfs/libxfs/xfs_btree_mem.c
115
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree_mem.c
120
unsigned int blocklen = xfbtree_rec_bytes(mp, ops);
fs/xfs/libxfs/xfs_btree_mem.c
125
if (!xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_btree_mem.c
126
ASSERT(xfs_has_crc(mp));
fs/xfs/libxfs/xfs_btree_mem.c
147
error = xfbtree_init_leaf_block(mp, xfbt, ops);
fs/xfs/libxfs/xfs_btree_mem.c
151
trace_xfbtree_init(mp, xfbt, ops);
fs/xfs/libxfs/xfs_btree_mem.c
77
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree_mem.c
86
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_btree_mem.h
65
int xfbtree_init(struct xfs_mount *mp, struct xfbtree *xfbt,
fs/xfs/libxfs/xfs_da_btree.c
130
state->mp = state->args->dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
133
static inline int xfs_dabuf_nfsb(struct xfs_mount *mp, int whichfork)
fs/xfs/libxfs/xfs_da_btree.c
136
return mp->m_dir_geo->fsbcount;
fs/xfs/libxfs/xfs_da_btree.c
137
return mp->m_attr_geo->fsbcount;
fs/xfs/libxfs/xfs_da_btree.c
142
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_da_btree.c
146
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_da_btree.c
169
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_da_btree.c
173
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_da_btree.c
202
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_da_btree.c
208
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_da_btree.c
209
if (!uuid_equal(&hdr3->uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_da_btree.c
213
if (!xfs_log_check_lsn(mp, be64_to_cpu(hdr3->lsn)))
fs/xfs/libxfs/xfs_da_btree.c
224
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_da_btree.c
229
xfs_da3_node_hdr_from_disk(mp, &ichdr, hdr);
fs/xfs/libxfs/xfs_da_btree.c
2440
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_da_btree.c
246
if (ichdr.count > mp->m_dir_geo->node_ents &&
fs/xfs/libxfs/xfs_da_btree.c
2466
mp = dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
247
ichdr.count > mp->m_attr_geo->node_ents)
fs/xfs/libxfs/xfs_da_btree.c
2471
if (XFS_IS_CORRUPT(mp, lastoff == 0)) {
fs/xfs/libxfs/xfs_da_btree.c
2536
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_da_btree.c
2564
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_da_btree.c
2595
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_da_btree.c
260
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_da_btree.c
2608
if (XFS_IS_CORRUPT(mp, entno == par_hdr.count)) {
fs/xfs/libxfs/xfs_da_btree.c
262
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_da_btree.c
2634
if (XFS_IS_CORRUPT(mp, par_blkno == 0)) {
fs/xfs/libxfs/xfs_da_btree.c
2651
if (XFS_IS_CORRUPT(mp, par_hdr.level != level)) {
fs/xfs/libxfs/xfs_da_btree.c
2728
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
2729
int nfsb = xfs_dabuf_nfsb(mp, whichfork);
fs/xfs/libxfs/xfs_da_btree.c
2762
map[i].bm_bn = XFS_FSB_TO_DADDR(mp, irecs[i].br_startblock);
fs/xfs/libxfs/xfs_da_btree.c
2763
map[i].bm_len = XFS_FSB_TO_BB(mp, irecs[i].br_blockcount);
fs/xfs/libxfs/xfs_da_btree.c
2778
if (XFS_IS_CORRUPT(mp, !(flags & XFS_DABUF_MAP_HOLE_OK))) {
fs/xfs/libxfs/xfs_da_btree.c
2782
xfs_alert(mp, "%s: bno %u inode %llu",
fs/xfs/libxfs/xfs_da_btree.c
2786
xfs_alert(mp,
fs/xfs/libxfs/xfs_da_btree.c
280
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_da_btree.c
2811
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
2822
error = xfs_trans_get_buf_map(tp, mp->m_ddev_targp, mapp, nmap, 0, &bp);
fs/xfs/libxfs/xfs_da_btree.c
283
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_da_btree.c
2848
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
2859
error = xfs_trans_read_buf_map(mp, tp, mp->m_ddev_targp, mapp, nmap, 0,
fs/xfs/libxfs/xfs_da_btree.c
304
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_da_btree.c
315
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_da_btree.c
455
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.c
458
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, mappedbno,
fs/xfs/libxfs/xfs_da_btree.c
459
XFS_FSB_TO_BB(mp, xfs_dabuf_nfsb(mp, whichfork)), 0,
fs/xfs/libxfs/xfs_da_btree.c
512
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_da_btree.c
528
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_da_btree.c
535
uuid_copy(&hdr3->info.uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_da_btree.c
568
if (XFS_TEST_ERROR(state->mp, XFS_ERRTAG_DA_LEAF_SPLIT))
fs/xfs/libxfs/xfs_da_btree.c
92
state->mp = args->dp->i_mount;
fs/xfs/libxfs/xfs_da_btree.h
133
struct xfs_mount *mp; /* filesystem mount point */
fs/xfs/libxfs/xfs_da_btree.h
238
void xfs_da3_node_hdr_from_disk(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_da_btree.h
240
void xfs_da3_node_hdr_to_disk(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_da_format.h
883
unsigned int xfs_attr3_rmt_buf_space(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_defer.c
1030
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
1035
xfs_defer_pending_abort_list(mp, &dfc->dfc_dfops);
fs/xfs/libxfs/xfs_defer.c
1036
xfs_defer_cancel_list(mp, &dfc->dfc_dfops);
fs/xfs/libxfs/xfs_defer.c
1060
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_defer.c
1076
xfs_defer_ops_capture_abort(mp, dfc);
fs/xfs/libxfs/xfs_defer.c
329
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
332
trace_xfs_defer_pending_abort(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
342
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
348
trace_xfs_defer_cancel_list(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
354
trace_xfs_defer_cancel_item(mp, dfp, pwi);
fs/xfs/libxfs/xfs_defer.c
363
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
370
xfs_defer_pending_abort(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
493
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
504
xfs_defer_pending_cancel_work(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
766
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_defer.c
770
xfs_defer_cancel_list(mp, &tp->t_dfops);
fs/xfs/libxfs/xfs_defer.c
907
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
910
xfs_defer_pending_abort(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
911
xfs_defer_pending_cancel_work(mp, dfp);
fs/xfs/libxfs/xfs_defer.c
917
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.c
927
trace_xlog_intent_recovery_failed(mp, ops, error);
fs/xfs/libxfs/xfs_defer.h
140
void xfs_defer_ops_capture_abort(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.h
146
void xfs_defer_cancel_recovery(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_defer.h
148
int xfs_defer_finish_recovery(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2.c
111
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_dir2.c
116
ASSERT(mp->m_sb.sb_versionnum & XFS_SB_VERSION_DIRV2BIT);
fs/xfs/libxfs/xfs_dir2.c
117
ASSERT(xfs_dir2_dirblock_bytes(&mp->m_sb) <= XFS_MAX_BLOCKSIZE);
fs/xfs/libxfs/xfs_dir2.c
119
mp->m_dir_geo = kzalloc_obj(struct xfs_da_geometry,
fs/xfs/libxfs/xfs_dir2.c
1194
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_dir2.c
121
mp->m_attr_geo = kzalloc_obj(struct xfs_da_geometry,
fs/xfs/libxfs/xfs_dir2.c
123
if (!mp->m_dir_geo || !mp->m_attr_geo) {
fs/xfs/libxfs/xfs_dir2.c
124
kfree(mp->m_dir_geo);
fs/xfs/libxfs/xfs_dir2.c
1249
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, du_wip->ip->i_ino));
fs/xfs/libxfs/xfs_dir2.c
125
kfree(mp->m_attr_geo);
fs/xfs/libxfs/xfs_dir2.c
130
dageo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2.c
131
dageo->blklog = mp->m_sb.sb_blocklog + mp->m_sb.sb_dirblklog;
fs/xfs/libxfs/xfs_dir2.c
132
dageo->fsblog = mp->m_sb.sb_blocklog;
fs/xfs/libxfs/xfs_dir2.c
133
dageo->blksize = xfs_dir2_dirblock_bytes(&mp->m_sb);
fs/xfs/libxfs/xfs_dir2.c
134
dageo->fsbcount = 1 << mp->m_sb.sb_dirblklog;
fs/xfs/libxfs/xfs_dir2.c
135
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2.c
154
xfs_dir2_data_entsize(mp, 1) +
fs/xfs/libxfs/xfs_dir2.c
155
xfs_dir2_data_entsize(mp, 2);
fs/xfs/libxfs/xfs_dir2.c
167
mp->m_sb.sb_blocklog;
fs/xfs/libxfs/xfs_dir2.c
171
dageo = mp->m_attr_geo;
fs/xfs/libxfs/xfs_dir2.c
172
dageo->blklog = mp->m_sb.sb_blocklog;
fs/xfs/libxfs/xfs_dir2.c
173
dageo->fsblog = mp->m_sb.sb_blocklog;
fs/xfs/libxfs/xfs_dir2.c
176
dageo->node_hdr_size = mp->m_dir_geo->node_hdr_size;
fs/xfs/libxfs/xfs_dir2.c
180
if (xfs_has_large_extent_counts(mp))
fs/xfs/libxfs/xfs_dir2.c
191
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_dir2.c
193
kfree(mp->m_dir_geo);
fs/xfs/libxfs/xfs_dir2.c
194
kfree(mp->m_attr_geo);
fs/xfs/libxfs/xfs_dir2.c
220
xfs_mount_t *mp,
fs/xfs/libxfs/xfs_dir2.c
223
bool ino_ok = xfs_verify_dir_ino(mp, ino);
fs/xfs/libxfs/xfs_dir2.c
225
if (XFS_IS_CORRUPT(mp, !ino_ok) ||
fs/xfs/libxfs/xfs_dir2.c
226
XFS_TEST_ERROR(mp, XFS_ERRTAG_DIR_INO_VALIDATE)) {
fs/xfs/libxfs/xfs_dir2.c
227
xfs_warn(mp, "Invalid inode number 0x%Lx",
fs/xfs/libxfs/xfs_dir2.c
270
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2.c
271
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2.c
284
if (eof == XFS_B_TO_FSB(mp, geo->blksize)) {
fs/xfs/libxfs/xfs_dir2.c
285
if (XFS_IS_CORRUPT(mp, dp->i_disk_size != geo->blksize)) {
fs/xfs/libxfs/xfs_dir2.c
614
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2.c
624
bno = XFS_B_TO_FSBT(mp, space * XFS_DIR2_SPACE_SIZE);
fs/xfs/libxfs/xfs_dir2.c
639
size = XFS_FSB_TO_B(mp, bno + count);
fs/xfs/libxfs/xfs_dir2.c
664
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_dir2.c
670
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2.c
720
dp->i_disk_size = XFS_FSB_TO_B(mp, bno);
fs/xfs/libxfs/xfs_dir2.c
744
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2.c
747
if (unlikely(xfs_has_asciici(mp)))
fs/xfs/libxfs/xfs_dir2.c
804
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_dir2.c
806
xfs_hooks_call(&mp->m_dir_update_hooks, 0, &p);
fs/xfs/libxfs/xfs_dir2.c
813
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2.c
816
return xfs_hooks_add(&mp->m_dir_update_hooks, &hook->dirent_hook);
fs/xfs/libxfs/xfs_dir2.c
822
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2.c
825
xfs_hooks_del(&mp->m_dir_update_hooks, &hook->dirent_hook);
fs/xfs/libxfs/xfs_dir2.c
903
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_dir2.c
922
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/libxfs/xfs_dir2.h
114
extern int xfs_dir_ino_validate(struct xfs_mount *mp, xfs_ino_t ino);
fs/xfs/libxfs/xfs_dir2.h
275
unsigned char xfs_dir3_get_dtype(struct xfs_mount *mp, uint8_t filetype);
fs/xfs/libxfs/xfs_dir2.h
329
int xfs_dir_hook_add(struct xfs_mount *mp, struct xfs_dir_hook *hook);
fs/xfs/libxfs/xfs_dir2.h
330
void xfs_dir_hook_del(struct xfs_mount *mp, struct xfs_dir_hook *hook);
fs/xfs/libxfs/xfs_dir2.h
58
extern int xfs_da_mount(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_dir2.h
59
extern void xfs_da_unmount(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_dir2.h
94
extern void xfs_dir2_data_freescan(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_block.c
100
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dir2_block.c
1081
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_block.c
1181
xfs_dir2_data_put_ftype(mp, dep, XFS_DIR3_FT_DIR);
fs/xfs/libxfs/xfs_dir2_block.c
1182
tagp = xfs_dir2_data_entry_tag_p(mp, dep);
fs/xfs/libxfs/xfs_dir2_block.c
1187
offset += xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/libxfs/xfs_dir2_block.c
1196
xfs_dir2_data_put_ftype(mp, dep, XFS_DIR3_FT_DIR);
fs/xfs/libxfs/xfs_dir2_block.c
1197
tagp = xfs_dir2_data_entry_tag_p(mp, dep);
fs/xfs/libxfs/xfs_dir2_block.c
1202
offset += xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/libxfs/xfs_dir2_block.c
123
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_block.c
1235
xfs_dir2_data_bestfree_p(mp, hdr),
fs/xfs/libxfs/xfs_dir2_block.c
1244
dep->inumber = cpu_to_be64(xfs_dir2_sf_get_ino(mp, sfp, sfep));
fs/xfs/libxfs/xfs_dir2_block.c
1246
xfs_dir2_data_put_ftype(mp, dep,
fs/xfs/libxfs/xfs_dir2_block.c
1247
xfs_dir2_sf_get_ftype(mp, sfep));
fs/xfs/libxfs/xfs_dir2_block.c
1249
tagp = xfs_dir2_data_entry_tag_p(mp, dep);
fs/xfs/libxfs/xfs_dir2_block.c
125
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_block.c
1254
blp[2 + i].hashval = cpu_to_be32(xfs_dir2_hashname(mp, &name));
fs/xfs/libxfs/xfs_dir2_block.c
1261
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_block.c
145
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_block.c
149
err = xfs_da_read_buf(tp, dp, mp->m_dir_geo->datablk, 0, bpp,
fs/xfs/libxfs/xfs_dir2_block.c
175
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_block.c
181
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_block.c
186
uuid_copy(&hdr3->uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_dir2_block.c
51
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_block.c
57
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_block.c
58
if (!uuid_equal(&hdr3->uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_dir2_block.c
62
if (!xfs_log_check_lsn(mp, be64_to_cpu(hdr3->lsn)))
fs/xfs/libxfs/xfs_dir2_block.c
72
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_block.c
75
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_dir2_block.c
89
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_block.c
944
xfs_mount_t *mp; /* file system mount point */
fs/xfs/libxfs/xfs_dir2_block.c
958
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_block.c
960
xfs_dir2_leaf_hdr_from_disk(mp, &leafhdr, leaf);
fs/xfs/libxfs/xfs_dir2_data.c
114
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
119
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2_data.c
153
bf = xfs_dir2_data_bestfree_p(mp, hdr);
fs/xfs/libxfs/xfs_dir2_data.c
231
if (offset > end - xfs_dir2_data_entsize(mp, 1))
fs/xfs/libxfs/xfs_dir2_data.c
242
reclen = xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/libxfs/xfs_dir2_data.c
245
if (!xfs_verify_dir_ino(mp, be64_to_cpu(dep->inumber)))
fs/xfs/libxfs/xfs_dir2_data.c
247
if (be16_to_cpu(*xfs_dir2_data_entry_tag_p(mp, dep)) != offset)
fs/xfs/libxfs/xfs_dir2_data.c
249
if (xfs_dir2_data_get_ftype(mp, dep) >= XFS_DIR3_FT_MAX)
fs/xfs/libxfs/xfs_dir2_data.c
260
hash = xfs_dir2_hashname(mp, &name);
fs/xfs/libxfs/xfs_dir2_data.c
30
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_data.c
316
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
322
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_data.c
323
if (!uuid_equal(&hdr3->uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_dir2_data.c
327
if (!xfs_log_check_lsn(mp, be64_to_cpu(hdr3->lsn)))
fs/xfs/libxfs/xfs_dir2_data.c
33
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dir2_data.c
365
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
368
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_dir2_data.c
382
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
393
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dir2_data.c
424
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
426
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_data.c
43
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_data.c
47
xfs_dir2_data_entsize(mp, dep->namelen) - sizeof(__be16));
fs/xfs/libxfs/xfs_dir2_data.c
52
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_data.c
55
if (xfs_has_ftype(mp)) {
fs/xfs/libxfs/xfs_dir2_data.c
666
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_data.c
67
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_data.c
670
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2_data.c
671
struct xfs_dir2_data_free *bf = xfs_dir2_data_bestfree_p(mp, hdr);
fs/xfs/libxfs/xfs_dir2_data.c
707
be16_to_cpu(*xfs_dir2_data_entry_tag_p(mp, dep)));
fs/xfs/libxfs/xfs_dir2_data.c
708
offset += xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/libxfs/xfs_dir2_data.c
724
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_data.c
74
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dir2_data.c
747
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_data.c
754
uuid_copy(&hdr3->uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_dir2_data.c
759
bf = xfs_dir2_data_bestfree_p(mp, hdr);
fs/xfs/libxfs/xfs_dir2_data.c
793
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_data.c
802
(uint)((char *)(xfs_dir2_data_entry_tag_p(mp, dep) + 1) -
fs/xfs/libxfs/xfs_dir2_leaf.c
1282
xfs_mount_t *mp; /* filesystem mount point */
fs/xfs/libxfs/xfs_dir2_leaf.c
1290
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
1300
xfs_dir2_leaf_hdr_from_disk(mp, leafhdr, leaf);
fs/xfs/libxfs/xfs_dir2_leaf.c
141
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_leaf.c
146
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2_leaf.c
1747
xfs_mount_t *mp; /* filesystem mount point */
fs/xfs/libxfs/xfs_dir2_leaf.c
1763
mp = state->mp;
fs/xfs/libxfs/xfs_dir2_leaf.c
1797
if (XFS_FSB_TO_B(mp, fo) > XFS_DIR2_LEAF_OFFSET + args->geo->blksize)
fs/xfs/libxfs/xfs_dir2_leaf.c
1801
xfs_dir2_leaf_hdr_from_disk(mp, &leafhdr, leaf);
fs/xfs/libxfs/xfs_dir2_leaf.c
1813
xfs_dir2_free_hdr_from_disk(mp, &freehdr, fbp->b_addr);
fs/xfs/libxfs/xfs_dir2_leaf.c
1850
xfs_dir2_leaf_hdr_to_disk(mp, leaf, &leafhdr);
fs/xfs/libxfs/xfs_dir2_leaf.c
199
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
207
xfs_dir2_leaf_hdr_from_disk(mp, &leafhdr, bp->b_addr);
fs/xfs/libxfs/xfs_dir2_leaf.c
208
return xfs_dir3_leaf_check_int(mp, &leafhdr, bp->b_addr, true);
fs/xfs/libxfs/xfs_dir2_leaf.c
216
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
218
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_leaf.c
238
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
241
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_dir2_leaf.c
255
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
266
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dir2_leaf.c
362
struct xfs_mount *mp = args->dp->i_mount;
fs/xfs/libxfs/xfs_dir2_leaf.c
368
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_leaf.c
37
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_leaf.c
378
uuid_copy(&leaf3->info.uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_dir2_leaf.c
391
ltp = xfs_dir2_leaf_tail_p(mp->m_dir_geo, leaf);
fs/xfs/libxfs/xfs_dir2_leaf.c
41
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_leaf.c
68
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_leaf.c
72
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
103
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_node.c
109
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
112
if (!uuid_equal(&hdr3->uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_dir2_node.c
116
if (!xfs_log_check_lsn(mp, be64_to_cpu(hdr3->lsn)))
fs/xfs/libxfs/xfs_dir2_node.c
129
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_node.c
132
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_dir2_node.c
146
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_node.c
157
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dir2_node.c
1699
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_node.c
1737
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_dir2_node.c
1740
xfs_alert(mp,
fs/xfs/libxfs/xfs_dir2_node.c
1746
xfs_alert(mp,
fs/xfs/libxfs/xfs_dir2_node.c
1751
xfs_alert(mp, " ... fblk is NULL");
fs/xfs/libxfs/xfs_dir2_node.c
1761
xfs_dir2_free_hdr_from_disk(mp, hdr, fbp->b_addr);
fs/xfs/libxfs/xfs_dir2_node.c
1768
xfs_dir2_free_hdr_from_disk(mp, hdr, fbp->b_addr);
fs/xfs/libxfs/xfs_dir2_node.c
1787
xfs_dir2_free_hdr_to_disk(mp, fbp->b_addr, hdr);
fs/xfs/libxfs/xfs_dir2_node.c
1792
bf = xfs_dir2_data_bestfree_p(mp, dbp->b_addr);
fs/xfs/libxfs/xfs_dir2_node.c
182
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dir2_node.c
183
int maxbests = mp->m_dir_geo->free_max_bests;
fs/xfs/libxfs/xfs_dir2_node.c
186
firstdb = (xfs_dir2_da_to_db(mp->m_dir_geo, fbno) -
fs/xfs/libxfs/xfs_dir2_node.c
187
xfs_dir2_byte_to_db(mp->m_dir_geo, XFS_DIR2_FREE_OFFSET)) *
fs/xfs/libxfs/xfs_dir2_node.c
189
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
2245
xfs_dir2_leaf_hdr_from_disk(state->mp, &leafhdr,
fs/xfs/libxfs/xfs_dir2_node.c
2262
xfs_dir2_data_put_ftype(state->mp, dep, ftype);
fs/xfs/libxfs/xfs_dir2_node.c
249
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_node.c
253
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
276
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_node.c
280
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
330
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_node.c
350
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dir2_node.c
357
uuid_copy(&hdr3->hdr.uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_dir2_node.c
360
xfs_dir2_free_hdr_to_disk(mp, bp->b_addr, &hdr);
fs/xfs/libxfs/xfs_dir2_node.c
650
xfs_mount_t *mp; /* filesystem mount point */
fs/xfs/libxfs/xfs_dir2_node.c
658
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_node.c
660
xfs_dir2_leaf_hdr_from_disk(mp, &leafhdr, leaf);
fs/xfs/libxfs/xfs_dir2_node.c
680
length = xfs_dir2_data_entsize(mp, args->namelen);
fs/xfs/libxfs/xfs_dir2_node.c
741
xfs_dir2_free_hdr_from_disk(mp, &freehdr, free);
fs/xfs/libxfs/xfs_dir2_node.c
742
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_dir2_node.c
801
xfs_mount_t *mp; /* filesystem mount point */
fs/xfs/libxfs/xfs_dir2_node.c
809
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_node.c
811
xfs_dir2_leaf_hdr_from_disk(mp, &leafhdr, leaf);
fs/xfs/libxfs/xfs_dir2_node.c
899
args->filetype = xfs_dir2_data_get_ftype(mp, dep);
fs/xfs/libxfs/xfs_dir2_priv.h
130
extern xfs_failaddr_t xfs_dir3_leaf_check_int(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
135
void xfs_dir2_free_hdr_from_disk(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
162
xfs_ino_t xfs_dir2_sf_get_ino(struct xfs_mount *mp, struct xfs_dir2_sf_hdr *hdr,
fs/xfs/libxfs/xfs_dir2_priv.h
166
uint8_t xfs_dir2_sf_get_ftype(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
168
struct xfs_dir2_sf_entry *xfs_dir2_sf_nextentry(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
179
xfs_failaddr_t xfs_dir2_sf_verify(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
181
int xfs_dir2_sf_entsize(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
183
void xfs_dir2_sf_put_ino(struct xfs_mount *mp, struct xfs_dir2_sf_hdr *hdr,
fs/xfs/libxfs/xfs_dir2_priv.h
185
void xfs_dir2_sf_put_ftype(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
201
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
208
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dir2_priv.h
213
xfs_dahash_t xfs_dir2_hashname(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
63
struct xfs_dir2_data_free *xfs_dir2_data_bestfree_p(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
65
__be16 *xfs_dir2_data_entry_tag_p(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
67
uint8_t xfs_dir2_data_get_ftype(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
69
void xfs_dir2_data_put_ftype(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
94
void xfs_dir2_leaf_hdr_from_disk(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_priv.h
96
void xfs_dir2_leaf_hdr_to_disk(struct xfs_mount *mp, struct xfs_dir2_leaf *to,
fs/xfs/libxfs/xfs_dir2_sf.c
1021
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
1075
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
1078
ino = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
1080
xfs_dir2_sf_put_ino(mp, sfp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
1082
xfs_dir2_sf_put_ftype(mp, sfep, args->filetype);
fs/xfs/libxfs/xfs_dir2_sf.c
1136
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
1180
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep),
fs/xfs/libxfs/xfs_dir2_sf.c
1181
oldsfep = xfs_dir2_sf_nextentry(mp, oldsfp, oldsfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
1185
xfs_dir2_sf_put_ino(mp, sfp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
1186
xfs_dir2_sf_get_ino(mp, oldsfp, oldsfep));
fs/xfs/libxfs/xfs_dir2_sf.c
1187
xfs_dir2_sf_put_ftype(mp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
1188
xfs_dir2_sf_get_ftype(mp, oldsfep));
fs/xfs/libxfs/xfs_dir2_sf.c
1208
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
1252
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep),
fs/xfs/libxfs/xfs_dir2_sf.c
1253
oldsfep = xfs_dir2_sf_nextentry(mp, oldsfp, oldsfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
1257
xfs_dir2_sf_put_ino(mp, sfp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
1258
xfs_dir2_sf_get_ino(mp, oldsfp, oldsfep));
fs/xfs/libxfs/xfs_dir2_sf.c
1259
xfs_dir2_sf_put_ftype(mp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
1260
xfs_dir2_sf_get_ftype(mp, oldsfep));
fs/xfs/libxfs/xfs_dir2_sf.c
135
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
138
if (xfs_has_ftype(mp)) {
fs/xfs/libxfs/xfs_dir2_sf.c
150
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
156
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dir2_sf.c
181
xfs_mount_t *mp; /* mount structure pointer */
fs/xfs/libxfs/xfs_dir2_sf.c
188
mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
189
geo = mp->m_dir_geo;
fs/xfs/libxfs/xfs_dir2_sf.c
195
has_ftype = xfs_has_ftype(mp) ? 1 : 0;
fs/xfs/libxfs/xfs_dir2_sf.c
264
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
279
sfp = kmalloc(mp->m_sb.sb_inodesize, GFP_KERNEL | __GFP_NOFAIL);
fs/xfs/libxfs/xfs_dir2_sf.c
319
xfs_dir2_sf_put_ino(mp, sfp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
321
xfs_dir2_sf_put_ftype(mp, sfep,
fs/xfs/libxfs/xfs_dir2_sf.c
322
xfs_dir2_data_get_ftype(mp, dep));
fs/xfs/libxfs/xfs_dir2_sf.c
324
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
326
offset += xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
42
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
462
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
469
sfp = xfs_idata_realloc(dp, xfs_dir2_sf_entsize(mp, sfp, args->namelen),
fs/xfs/libxfs/xfs_dir2_sf.c
481
xfs_dir2_sf_put_ino(mp, sfp, sfep, args->inumber);
fs/xfs/libxfs/xfs_dir2_sf.c
482
xfs_dir2_sf_put_ftype(mp, sfep, args->filetype);
fs/xfs/libxfs/xfs_dir2_sf.c
51
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dir2_sf.c
510
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
537
add_datasize = xfs_dir2_data_entsize(mp, args->namelen),
fs/xfs/libxfs/xfs_dir2_sf.c
540
offset = new_offset + xfs_dir2_data_entsize(mp, oldsfep->namelen),
fs/xfs/libxfs/xfs_dir2_sf.c
541
oldsfep = xfs_dir2_sf_nextentry(mp, oldsfp, oldsfep),
fs/xfs/libxfs/xfs_dir2_sf.c
567
xfs_dir2_sf_put_ino(mp, sfp, sfep, args->inumber);
fs/xfs/libxfs/xfs_dir2_sf.c
568
xfs_dir2_sf_put_ftype(mp, sfep, args->filetype);
fs/xfs/libxfs/xfs_dir2_sf.c
576
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
58
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
599
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
608
size = xfs_dir2_data_entsize(mp, args->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
62
return (void *)sfep + xfs_dir2_sf_entsize(mp, hdr, sfep->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
621
xfs_dir2_data_entsize(mp, sfep->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
622
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
665
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
679
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
681
ino = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
685
xfs_dir2_data_entsize(mp, sfep->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
686
ASSERT(xfs_dir2_sf_get_ftype(mp, sfep) < XFS_DIR3_FT_MAX);
fs/xfs/libxfs/xfs_dir2_sf.c
699
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
725
error = xfs_dir_ino_validate(mp, ino);
fs/xfs/libxfs/xfs_dir2_sf.c
728
offset = mp->m_dir_geo->data_first_offset;
fs/xfs/libxfs/xfs_dir2_sf.c
73
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
750
next_sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
759
ino = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
761
error = xfs_dir_ino_validate(mp, ino);
fs/xfs/libxfs/xfs_dir2_sf.c
766
filetype = xfs_dir2_sf_get_ftype(mp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
771
xfs_dir2_data_entsize(mp, sfep->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
782
(uint)sizeof(xfs_dir2_block_tail_t) > mp->m_dir_geo->blksize)
fs/xfs/libxfs/xfs_dir2_sf.c
79
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dir2_sf.c
846
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
886
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
89
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dir2_sf.c
895
args->inumber = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
896
args->filetype = xfs_dir2_sf_get_ftype(mp, sfep);
fs/xfs/libxfs/xfs_dir2_sf.c
921
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_dir2_sf.c
943
i++, sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep)) {
fs/xfs/libxfs/xfs_dir2_sf.c
946
ASSERT(xfs_dir2_sf_get_ino(mp, sfp, sfep) ==
fs/xfs/libxfs/xfs_dir2_sf.c
960
entsize = xfs_dir2_sf_entsize(mp, sfp, args->namelen);
fs/xfs/libxfs/xfs_dir2_sf.c
98
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_dquot_buf.c
108
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
112
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_dquot_buf.c
113
!uuid_equal(&dqb->dd_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_dquot_buf.c
116
return xfs_dquot_verify(mp, &dqb->dd_diskdq, id);
fs/xfs/libxfs/xfs_dquot_buf.c
124
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
140
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_dquot_buf.c
141
uuid_copy(&dqb->dd_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_dquot_buf.c
149
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
157
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_dquot_buf.c
165
if (mp->m_quotainfo)
fs/xfs/libxfs/xfs_dquot_buf.c
166
ndquots = mp->m_quotainfo->qi_dqperchunk;
fs/xfs/libxfs/xfs_dquot_buf.c
184
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
199
if (mp->m_quotainfo)
fs/xfs/libxfs/xfs_dquot_buf.c
200
ndquots = mp->m_quotainfo->qi_dqperchunk;
fs/xfs/libxfs/xfs_dquot_buf.c
219
fa = xfs_dqblk_verify(mp, &dqb[i], id + i);
fs/xfs/libxfs/xfs_dquot_buf.c
236
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dquot_buf.c
238
return xfs_dquot_buf_verify(mp, bp, false);
fs/xfs/libxfs/xfs_dquot_buf.c
245
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dquot_buf.c
247
if (!xfs_dquot_buf_verify_crc(mp, bp, false))
fs/xfs/libxfs/xfs_dquot_buf.c
249
xfs_dquot_buf_verify(mp, bp, false);
fs/xfs/libxfs/xfs_dquot_buf.c
262
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dquot_buf.c
264
if (!xfs_dquot_buf_verify_crc(mp, bp, true) ||
fs/xfs/libxfs/xfs_dquot_buf.c
265
xfs_dquot_buf_verify(mp, bp, true) != NULL) {
fs/xfs/libxfs/xfs_dquot_buf.c
280
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_dquot_buf.c
282
xfs_dquot_buf_verify(mp, bp, false);
fs/xfs/libxfs/xfs_dquot_buf.c
362
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_dquot_buf.c
367
if (!xfs_has_metadir(mp)) {
fs/xfs/libxfs/xfs_dquot_buf.c
372
ino = mp->m_sb.sb_uquotino;
fs/xfs/libxfs/xfs_dquot_buf.c
375
ino = mp->m_sb.sb_gquotino;
fs/xfs/libxfs/xfs_dquot_buf.c
378
ino = mp->m_sb.sb_pquotino;
fs/xfs/libxfs/xfs_dquot_buf.c
398
xfs_fs_mark_sick(mp, xfs_dqinode_sick_mask(type));
fs/xfs/libxfs/xfs_dquot_buf.c
402
if (XFS_IS_CORRUPT(mp, ip->i_df.if_format != XFS_DINODE_FMT_EXTENTS &&
fs/xfs/libxfs/xfs_dquot_buf.c
405
xfs_fs_mark_sick(mp, xfs_dqinode_sick_mask(type));
fs/xfs/libxfs/xfs_dquot_buf.c
409
if (XFS_IS_CORRUPT(mp, ip->i_projid != 0)) {
fs/xfs/libxfs/xfs_dquot_buf.c
41
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
411
xfs_fs_mark_sick(mp, xfs_dqinode_sick_mask(type));
fs/xfs/libxfs/xfs_dquot_buf.c
485
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_dquot_buf.c
488
if (!mp->m_metadirip) {
fs/xfs/libxfs/xfs_dquot_buf.c
489
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_dquot_buf.c
493
return xfs_metadir_mkdir(mp->m_metadirip, "quota", dpp);
fs/xfs/libxfs/xfs_dquot_buf.c
506
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_dquot_buf.c
508
if (!mp->m_metadirip) {
fs/xfs/libxfs/xfs_dquot_buf.c
509
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_dquot_buf.c
513
return xfs_metadir_load(tp, mp->m_metadirip, "quota", XFS_METAFILE_DIR,
fs/xfs/libxfs/xfs_dquot_buf.c
76
!xfs_has_bigtime(mp))
fs/xfs/libxfs/xfs_exchmaps.c
138
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_exchmaps.c
196
struct xfs_mount *mp = xmi->xmi_ip1->i_mount;
fs/xfs/libxfs/xfs_exchmaps.c
237
if (!isaligned_64(irec->br_startoff, mp->m_sb.sb_rextsize)) {
fs/xfs/libxfs/xfs_exchmaps.c
240
new_end = roundup_64(irec->br_startoff, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_exchmaps.c
245
if (isaligned_64(irec->br_blockcount, mp->m_sb.sb_rextsize))
fs/xfs/libxfs/xfs_exchmaps.c
257
if (irec->br_blockcount > mp->m_sb.sb_rextsize) {
fs/xfs/libxfs/xfs_exchmaps.c
261
mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_exchmaps.c
648
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_exchmaps.c
652
XFS_MAX_CONTIG_BMAPS_PER_BLOCK(mp)) *
fs/xfs/libxfs/xfs_exchmaps.c
653
XFS_EXTENTADD_SPACE_RES(mp, xfs_exchmaps_reqfork(req));
fs/xfs/libxfs/xfs_exchmaps.c
659
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_exchmaps.c
662
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_exchmaps.c
666
XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp)) *
fs/xfs/libxfs/xfs_exchmaps.c
667
XFS_RTRMAPADD_SPACE_RES(mp);
fs/xfs/libxfs/xfs_exchmaps.c
670
XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp)) *
fs/xfs/libxfs/xfs_exchmaps.c
671
XFS_RMAPADD_SPACE_RES(mp);
fs/xfs/libxfs/xfs_exchmaps.c
679
struct xfs_mount *mp = req->ip1->i_mount;
fs/xfs/libxfs/xfs_exchmaps.c
688
bmbt_blocks = xfs_exchmaps_bmbt_blocks(mp, req);
fs/xfs/libxfs/xfs_exchmaps.c
689
rmapbt_blocks = xfs_exchmaps_rmapbt_blocks(mp, req);
fs/xfs/libxfs/xfs_exchmaps.c
691
trace_xfs_exchmaps_overhead(mp, bmbt_blocks, rmapbt_blocks);
fs/xfs/libxfs/xfs_exchmaps.c
775
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_exchmaps.c
855
trace_xfs_exchmaps_delta_nextents_step(mp, left, curr, new, right, ret,
fs/xfs/libxfs/xfs_exchmaps.c
867
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_exchmaps.c
885
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_REDUCE_MAX_IEXTENTS) &&
fs/xfs/libxfs/xfs_exchmaps.c
894
max_nextents = xfs_iext_max_nextents(xfs_has_large_extent_counts(mp),
fs/xfs/libxfs/xfs_exchmaps.c
90
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_exchmaps.c
96
len = min(XFS_FSB_TO_B(mp, imap->br_startoff + imap->br_blockcount),
fs/xfs/libxfs/xfs_exchmaps.c
960
struct xfs_mount *mp = req->ip1->i_mount;
fs/xfs/libxfs/xfs_exchmaps.c
966
if (req->blockcount != XFS_B_TO_FSB(mp, req->ip1->i_disk_size))
fs/xfs/libxfs/xfs_exchmaps.c
968
if (req->blockcount != XFS_B_TO_FSB(mp, req->ip2->i_disk_size))
fs/xfs/libxfs/xfs_exchmaps.h
118
int xfs_exchmaps_check_forks(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_format.h
1089
#define XFS_DINODE_SIZE(mp) \
fs/xfs/libxfs/xfs_format.h
1090
(xfs_has_v3inodes(mp) ? \
fs/xfs/libxfs/xfs_format.h
1093
#define XFS_LITINO(mp) \
fs/xfs/libxfs/xfs_format.h
1094
((mp)->m_sb.sb_inodesize - XFS_DINODE_SIZE(mp))
fs/xfs/libxfs/xfs_format.h
1101
#define XFS_DFORK_DSIZE(dip,mp) \
fs/xfs/libxfs/xfs_format.h
1102
((dip)->di_forkoff ? XFS_DFORK_BOFF(dip) : XFS_LITINO(mp))
fs/xfs/libxfs/xfs_format.h
1103
#define XFS_DFORK_ASIZE(dip,mp) \
fs/xfs/libxfs/xfs_format.h
1104
((dip)->di_forkoff ? XFS_LITINO(mp) - XFS_DFORK_BOFF(dip) : 0)
fs/xfs/libxfs/xfs_format.h
1105
#define XFS_DFORK_SIZE(dip,mp,w) \
fs/xfs/libxfs/xfs_format.h
1107
XFS_DFORK_DSIZE(dip, mp) : \
fs/xfs/libxfs/xfs_format.h
1108
XFS_DFORK_ASIZE(dip, mp))
fs/xfs/libxfs/xfs_format.h
1110
#define XFS_DFORK_MAXEXT(dip, mp, w) \
fs/xfs/libxfs/xfs_format.h
1111
(XFS_DFORK_SIZE(dip, mp, w) / sizeof(struct xfs_bmbt_rec))
fs/xfs/libxfs/xfs_format.h
1271
#define XFS_INO_OFFSET_BITS(mp) (mp)->m_sb.sb_inopblog
fs/xfs/libxfs/xfs_format.h
1272
#define XFS_INO_AGBNO_BITS(mp) (mp)->m_sb.sb_agblklog
fs/xfs/libxfs/xfs_format.h
1273
#define XFS_INO_AGINO_BITS(mp) ((mp)->m_ino_geo.agino_log)
fs/xfs/libxfs/xfs_format.h
1274
#define XFS_INO_AGNO_BITS(mp) (mp)->m_agno_log
fs/xfs/libxfs/xfs_format.h
1275
#define XFS_INO_BITS(mp) \
fs/xfs/libxfs/xfs_format.h
1276
XFS_INO_AGNO_BITS(mp) + XFS_INO_AGINO_BITS(mp)
fs/xfs/libxfs/xfs_format.h
1277
#define XFS_INO_TO_AGNO(mp,i) \
fs/xfs/libxfs/xfs_format.h
1278
((xfs_agnumber_t)((i) >> XFS_INO_AGINO_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1279
#define XFS_INO_TO_AGINO(mp,i) \
fs/xfs/libxfs/xfs_format.h
1280
((xfs_agino_t)(i) & XFS_INO_MASK(XFS_INO_AGINO_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1281
#define XFS_INO_TO_AGBNO(mp,i) \
fs/xfs/libxfs/xfs_format.h
1282
(((xfs_agblock_t)(i) >> XFS_INO_OFFSET_BITS(mp)) & \
fs/xfs/libxfs/xfs_format.h
1283
XFS_INO_MASK(XFS_INO_AGBNO_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1284
#define XFS_INO_TO_OFFSET(mp,i) \
fs/xfs/libxfs/xfs_format.h
1285
((int)(i) & XFS_INO_MASK(XFS_INO_OFFSET_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1286
#define XFS_INO_TO_FSB(mp,i) \
fs/xfs/libxfs/xfs_format.h
1287
XFS_AGB_TO_FSB(mp, XFS_INO_TO_AGNO(mp,i), XFS_INO_TO_AGBNO(mp,i))
fs/xfs/libxfs/xfs_format.h
1288
#define XFS_AGINO_TO_INO(mp,a,i) \
fs/xfs/libxfs/xfs_format.h
1289
(((xfs_ino_t)(a) << XFS_INO_AGINO_BITS(mp)) | (i))
fs/xfs/libxfs/xfs_format.h
1290
#define XFS_AGINO_TO_AGBNO(mp,i) ((i) >> XFS_INO_OFFSET_BITS(mp))
fs/xfs/libxfs/xfs_format.h
1291
#define XFS_AGINO_TO_OFFSET(mp,i) \
fs/xfs/libxfs/xfs_format.h
1292
((i) & XFS_INO_MASK(XFS_INO_OFFSET_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1293
#define XFS_OFFBNO_TO_AGINO(mp,b,o) \
fs/xfs/libxfs/xfs_format.h
1294
((xfs_agino_t)(((b) << XFS_INO_OFFSET_BITS(mp)) | (o)))
fs/xfs/libxfs/xfs_format.h
1295
#define XFS_FSB_TO_INO(mp, b) ((xfs_ino_t)((b) << XFS_INO_OFFSET_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1296
#define XFS_AGB_TO_AGINO(mp, b) ((xfs_agino_t)((b) << XFS_INO_OFFSET_BITS(mp)))
fs/xfs/libxfs/xfs_format.h
1523
#define XFS_SYMLINK_BUF_SPACE(mp, bufsize) \
fs/xfs/libxfs/xfs_format.h
1524
((bufsize) - (xfs_has_crc((mp)) ? \
fs/xfs/libxfs/xfs_format.h
1560
#define XFS_BNO_BLOCK(mp) ((xfs_agblock_t)(XFS_AGFL_BLOCK(mp) + 1))
fs/xfs/libxfs/xfs_format.h
1561
#define XFS_CNT_BLOCK(mp) ((xfs_agblock_t)(XFS_BNO_BLOCK(mp) + 1))
fs/xfs/libxfs/xfs_format.h
1642
#define XFS_IBT_BLOCK(mp) ((xfs_agblock_t)(XFS_CNT_BLOCK(mp) + 1))
fs/xfs/libxfs/xfs_format.h
1643
#define XFS_FIBT_BLOCK(mp) ((xfs_agblock_t)(XFS_IBT_BLOCK(mp) + 1))
fs/xfs/libxfs/xfs_format.h
1741
#define XFS_RMAP_BLOCK(mp) \
fs/xfs/libxfs/xfs_format.h
1742
(xfs_has_finobt(((mp))) ? \
fs/xfs/libxfs/xfs_format.h
1743
XFS_FIBT_BLOCK(mp) + 1 : \
fs/xfs/libxfs/xfs_format.h
1744
XFS_IBT_BLOCK(mp) + 1)
fs/xfs/libxfs/xfs_format.h
1770
unsigned int xfs_refc_block(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_format.h
2004
#define XFS_ACL_MAX_ENTRIES(mp) \
fs/xfs/libxfs/xfs_format.h
2005
(xfs_has_crc(mp) \
fs/xfs/libxfs/xfs_format.h
2014
#define XFS_ACL_MAX_SIZE(mp) \
fs/xfs/libxfs/xfs_format.h
2015
XFS_ACL_SIZE(XFS_ACL_MAX_ENTRIES((mp)))
fs/xfs/libxfs/xfs_format.h
467
#define XFS_SB_BLOCK(mp) XFS_HDR_BLOCK(mp, XFS_SB_DADDR)
fs/xfs/libxfs/xfs_format.h
469
#define XFS_HDR_BLOCK(mp,d) ((xfs_agblock_t)XFS_BB_TO_FSBT(mp,d))
fs/xfs/libxfs/xfs_format.h
470
#define XFS_DADDR_TO_FSB(mp,d) XFS_AGB_TO_FSB(mp, \
fs/xfs/libxfs/xfs_format.h
471
xfs_daddr_to_agno(mp,d), xfs_daddr_to_agbno(mp,d))
fs/xfs/libxfs/xfs_format.h
472
#define XFS_FSB_TO_DADDR(mp,fsbno) XFS_AGB_TO_DADDR(mp, \
fs/xfs/libxfs/xfs_format.h
473
XFS_FSB_TO_AGNO(mp,fsbno), XFS_FSB_TO_AGBNO(mp,fsbno))
fs/xfs/libxfs/xfs_format.h
478
#define XFS_FSS_TO_BB(mp,sec) ((sec) << (mp)->m_sectbb_log)
fs/xfs/libxfs/xfs_format.h
483
#define XFS_FSB_TO_BB(mp,fsbno) ((fsbno) << (mp)->m_blkbb_log)
fs/xfs/libxfs/xfs_format.h
484
#define XFS_BB_TO_FSB(mp,bb) \
fs/xfs/libxfs/xfs_format.h
485
(((bb) + (XFS_FSB_TO_BB(mp,1) - 1)) >> (mp)->m_blkbb_log)
fs/xfs/libxfs/xfs_format.h
486
#define XFS_BB_TO_FSBT(mp,bb) ((bb) >> (mp)->m_blkbb_log)
fs/xfs/libxfs/xfs_format.h
491
#define XFS_FSB_TO_B(mp,fsbno) ((xfs_fsize_t)(fsbno) << (mp)->m_sb.sb_blocklog)
fs/xfs/libxfs/xfs_format.h
492
#define XFS_B_TO_FSB(mp,b) \
fs/xfs/libxfs/xfs_format.h
493
((((uint64_t)(b)) + (mp)->m_blockmask) >> (mp)->m_sb.sb_blocklog)
fs/xfs/libxfs/xfs_format.h
494
#define XFS_B_TO_FSBT(mp,b) (((uint64_t)(b)) >> (mp)->m_sb.sb_blocklog)
fs/xfs/libxfs/xfs_format.h
610
#define XFS_AGF_DADDR(mp) ((xfs_daddr_t)(1 << (mp)->m_sectbb_log))
fs/xfs/libxfs/xfs_format.h
611
#define XFS_AGF_BLOCK(mp) XFS_HDR_BLOCK(mp, XFS_AGF_DADDR(mp))
fs/xfs/libxfs/xfs_format.h
681
#define XFS_AGI_DADDR(mp) ((xfs_daddr_t)(2 << (mp)->m_sectbb_log))
fs/xfs/libxfs/xfs_format.h
682
#define XFS_AGI_BLOCK(mp) XFS_HDR_BLOCK(mp, XFS_AGI_DADDR(mp))
fs/xfs/libxfs/xfs_format.h
688
#define XFS_AGFL_DADDR(mp) ((xfs_daddr_t)(3 << (mp)->m_sectbb_log))
fs/xfs/libxfs/xfs_format.h
689
#define XFS_AGFL_BLOCK(mp) XFS_HDR_BLOCK(mp, XFS_AGFL_DADDR(mp))
fs/xfs/libxfs/xfs_format.h
702
#define XFS_AGB_TO_FSB(mp,agno,agbno) \
fs/xfs/libxfs/xfs_format.h
703
(((xfs_fsblock_t)(agno) << (mp)->m_sb.sb_agblklog) | (agbno))
fs/xfs/libxfs/xfs_format.h
704
#define XFS_FSB_TO_AGNO(mp,fsbno) \
fs/xfs/libxfs/xfs_format.h
705
((xfs_agnumber_t)((fsbno) >> (mp)->m_sb.sb_agblklog))
fs/xfs/libxfs/xfs_format.h
706
#define XFS_FSB_TO_AGBNO(mp,fsbno) \
fs/xfs/libxfs/xfs_format.h
707
((xfs_agblock_t)((fsbno) & xfs_mask32lo((mp)->m_sb.sb_agblklog)))
fs/xfs/libxfs/xfs_format.h
708
#define XFS_AGB_TO_DADDR(mp,agno,agbno) \
fs/xfs/libxfs/xfs_format.h
709
((xfs_daddr_t)XFS_FSB_TO_BB(mp, \
fs/xfs/libxfs/xfs_format.h
710
(xfs_fsblock_t)(agno) * (mp)->m_sb.sb_agblocks + (agbno)))
fs/xfs/libxfs/xfs_format.h
711
#define XFS_AG_DADDR(mp,agno,d) (XFS_AGB_TO_DADDR(mp, agno, 0) + (d))
fs/xfs/libxfs/xfs_format.h
717
#define XFS_AG_CHECK_DADDR(mp,d,len) \
fs/xfs/libxfs/xfs_format.h
720
xfs_daddr_to_agbno(mp, d) != XFS_SB_DADDR) : \
fs/xfs/libxfs/xfs_format.h
721
ASSERT(xfs_daddr_to_agno(mp, d) == \
fs/xfs/libxfs/xfs_format.h
722
xfs_daddr_to_agno(mp, (d) + (len) - 1)))
fs/xfs/libxfs/xfs_group.c
100
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
114
return xfs_group_grab(mp, index, type);
fs/xfs/libxfs/xfs_group.c
122
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
135
xg = xa_find(&mp->m_groups[type].xa, &index, ULONG_MAX, mark);
fs/xfs/libxfs/xfs_group.c
155
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
160
struct xfs_group *xg = xa_erase(&mp->m_groups[type].xa, index);
fs/xfs/libxfs/xfs_group.c
162
XFS_IS_CORRUPT(mp, atomic_read(&xg->xg_ref) != 0);
fs/xfs/libxfs/xfs_group.c
175
XFS_IS_CORRUPT(mp, atomic_read(&xg->xg_active_ref) > 0);
fs/xfs/libxfs/xfs_group.c
176
XFS_IS_CORRUPT(mp, atomic_read(&xg->xg_active_ref) < 0);
fs/xfs/libxfs/xfs_group.c
182
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
189
xg->xg_mount = mp;
fs/xfs/libxfs/xfs_group.c
194
if (xfs_group_has_extent_busy(mp, type)) {
fs/xfs/libxfs/xfs_group.c
207
error = xa_insert(&mp->m_groups[type].xa, index, xg, GFP_KERNEL);
fs/xfs/libxfs/xfs_group.c
225
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
229
return xfs_group_get(mp, xfs_fsb_to_gno(mp, fsbno, type), type);
fs/xfs/libxfs/xfs_group.c
33
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
40
xg = xa_load(&mp->m_groups[type].xa, index);
fs/xfs/libxfs/xfs_group.c
74
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.c
81
xg = xa_load(&mp->m_groups[type].xa, index);
fs/xfs/libxfs/xfs_group.h
103
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
107
return (xfs_rfsblock_t)mp->m_groups[type].blocks * nr_groups;
fs/xfs/libxfs/xfs_group.h
131
struct xfs_mount *mp = xg->xg_mount;
fs/xfs/libxfs/xfs_group.h
132
struct xfs_groups *g = &mp->m_groups[xg->xg_type];
fs/xfs/libxfs/xfs_group.h
140
return XFS_FSB_TO_BB(mp, g->start_fsb + fsbno);
fs/xfs/libxfs/xfs_group.h
145
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
149
if (!mp->m_groups[type].blklog)
fs/xfs/libxfs/xfs_group.h
151
return fsbno >> mp->m_groups[type].blklog;
fs/xfs/libxfs/xfs_group.h
156
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
160
return fsbno & mp->m_groups[type].blkmask;
fs/xfs/libxfs/xfs_group.h
64
struct xfs_group *xfs_group_get(struct xfs_mount *mp, uint32_t index,
fs/xfs/libxfs/xfs_group.h
66
struct xfs_group *xfs_group_get_by_fsb(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
71
struct xfs_group *xfs_group_grab(struct xfs_mount *mp, uint32_t index,
fs/xfs/libxfs/xfs_group.h
73
struct xfs_group *xfs_group_next_range(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
76
struct xfs_group *xfs_group_grab_next_mark(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_group.h
80
void xfs_group_free(struct xfs_mount *mp, uint32_t index,
fs/xfs/libxfs/xfs_group.h
82
int xfs_group_insert(struct xfs_mount *mp, struct xfs_group *xg,
fs/xfs/libxfs/xfs_health.h
199
void xfs_fs_mark_sick(struct xfs_mount *mp, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
200
void xfs_fs_mark_corrupt(struct xfs_mount *mp, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
201
void xfs_fs_mark_healthy(struct xfs_mount *mp, unsigned int mask);
fs/xfs/libxfs/xfs_health.h
202
void xfs_fs_measure_sickness(struct xfs_mount *mp, unsigned int *sick,
fs/xfs/libxfs/xfs_health.h
205
void xfs_rgno_mark_sick(struct xfs_mount *mp, xfs_rgnumber_t rgno,
fs/xfs/libxfs/xfs_health.h
208
void xfs_agno_mark_sick(struct xfs_mount *mp, xfs_agnumber_t agno,
fs/xfs/libxfs/xfs_health.h
224
void xfs_health_unmount(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_health.h
233
xfs_fs_has_sickness(struct xfs_mount *mp, unsigned int mask)
fs/xfs/libxfs/xfs_health.h
237
xfs_fs_measure_sickness(mp, &sick, &checked);
fs/xfs/libxfs/xfs_health.h
272
xfs_fs_is_healthy(struct xfs_mount *mp)
fs/xfs/libxfs/xfs_health.h
274
return !xfs_fs_has_sickness(mp, -1U);
fs/xfs/libxfs/xfs_health.h
283
void xfs_fsop_geom_health(struct xfs_mount *mp, struct xfs_fsop_geom *geo);
fs/xfs/libxfs/xfs_ialloc.c
1097
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
1099
xfs_agnumber_t pagno = XFS_INO_TO_AGNO(mp, parent);
fs/xfs/libxfs/xfs_ialloc.c
1100
xfs_agino_t pagino = XFS_INO_TO_AGINO(mp, parent);
fs/xfs/libxfs/xfs_ialloc.c
1136
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
1145
if (XFS_IS_CORRUPT(mp, j != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
1304
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
1314
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
1324
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
1335
ASSERT((XFS_AGINO_TO_OFFSET(mp, rec.ir_startino) %
fs/xfs/libxfs/xfs_ialloc.c
141
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_ialloc.c
143
xfs_warn(mp,
fs/xfs/libxfs/xfs_ialloc.c
146
xfs_warn(mp,
fs/xfs/libxfs/xfs_ialloc.c
1573
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
1575
xfs_agnumber_t pagno = XFS_INO_TO_AGNO(mp, parent);
fs/xfs/libxfs/xfs_ialloc.c
1576
xfs_agino_t pagino = XFS_INO_TO_AGINO(mp, parent);
fs/xfs/libxfs/xfs_ialloc.c
1585
if (!xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_ialloc.c
1616
ASSERT((XFS_AGINO_TO_OFFSET(mp, rec.ir_startino) %
fs/xfs/libxfs/xfs_ialloc.c
163
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_ialloc.c
172
xfs_inobt_btrec_to_irec(mp, rec, irec);
fs/xfs/libxfs/xfs_ialloc.c
1729
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
1781
ineed = M_IGEO(mp)->ialloc_min_blks;
fs/xfs/libxfs/xfs_ialloc.c
1783
ineed += M_IGEO(mp)->cluster_align;
fs/xfs/libxfs/xfs_ialloc.c
1856
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc.c
1865
if (mp->m_sb.sb_logstart)
fs/xfs/libxfs/xfs_ialloc.c
1866
return XFS_FSB_TO_AGNO(mp, mp->m_sb.sb_logstart);
fs/xfs/libxfs/xfs_ialloc.c
1871
return (atomic_inc_return(&mp->m_agirotor) - 1) % mp->m_maxagi;
fs/xfs/libxfs/xfs_ialloc.c
1873
start_agno = XFS_INO_TO_AGNO(mp, dp->i_ino);
fs/xfs/libxfs/xfs_ialloc.c
1874
if (start_agno >= mp->m_maxagi)
fs/xfs/libxfs/xfs_ialloc.c
1894
struct xfs_mount *mp = (*tpp)->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
1896
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/libxfs/xfs_ialloc.c
1907
start_agno = xfs_dialloc_pick_ag(mp, args->pip, mode);
fs/xfs/libxfs/xfs_ialloc.c
1918
percpu_counter_read_positive(&mp->m_icount) + igeo->ialloc_inos
fs/xfs/libxfs/xfs_ialloc.c
1931
if (xfs_estimate_freecounter(mp, XC_FREE_BLOCKS) <
fs/xfs/libxfs/xfs_ialloc.c
1932
mp->m_low_space[XFS_LOWSP_1_PCNT]) {
fs/xfs/libxfs/xfs_ialloc.c
1944
for_each_perag_wrap_at(mp, start_agno, mp->m_maxagi, agno, pag) {
fs/xfs/libxfs/xfs_ialloc.c
1953
if (xfs_is_shutdown(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
1979
if (ino == parent || !xfs_verify_dir_ino(mp, ino)) {
fs/xfs/libxfs/xfs_ialloc.c
1980
xfs_alert(mp, "Allocated a known in-use inode 0x%llx!", ino);
fs/xfs/libxfs/xfs_ialloc.c
1981
xfs_agno_mark_sick(mp, XFS_INO_TO_AGNO(mp, ino),
fs/xfs/libxfs/xfs_ialloc.c
2001
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
2002
xfs_agblock_t sagbno = XFS_AGINO_TO_AGBNO(mp,
fs/xfs/libxfs/xfs_ialloc.c
2013
M_IGEO(mp)->ialloc_blks, &XFS_RMAP_OINFO_INODES,
fs/xfs/libxfs/xfs_ialloc.c
2051
mp->m_sb.sb_inopblock;
fs/xfs/libxfs/xfs_ialloc.c
2054
mp->m_sb.sb_inopblock;
fs/xfs/libxfs/xfs_ialloc.c
2056
ASSERT(agbno % mp->m_sb.sb_spino_align == 0);
fs/xfs/libxfs/xfs_ialloc.c
2057
ASSERT(contigblk % mp->m_sb.sb_spino_align == 0);
fs/xfs/libxfs/xfs_ialloc.c
2082
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
2092
ASSERT(XFS_AGINO_TO_AGBNO(mp, agino) < be32_to_cpu(agi->agi_length));
fs/xfs/libxfs/xfs_ialloc.c
2107
xfs_warn(mp, "%s: xfs_inobt_lookup() returned error %d.",
fs/xfs/libxfs/xfs_ialloc.c
2111
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
2118
xfs_warn(mp, "%s: xfs_inobt_get_rec() returned error %d.",
fs/xfs/libxfs/xfs_ialloc.c
2122
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
2145
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
fs/xfs/libxfs/xfs_ialloc.c
2165
xfs_warn(mp, "%s: xfs_btree_delete returned error %d.",
fs/xfs/libxfs/xfs_ialloc.c
2178
xfs_warn(mp, "%s: xfs_inobt_update returned error %d.",
fs/xfs/libxfs/xfs_ialloc.c
2216
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
2234
if (XFS_IS_CORRUPT(mp, ibtrec->ir_freecount != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
2261
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
2270
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_ialloc.c
2291
mp->m_sb.sb_inopblock <= XFS_INODES_PER_CHUNK) {
fs/xfs/libxfs/xfs_ialloc.c
2333
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
2339
if (pag_agno(pag) != XFS_INO_TO_AGNO(mp, inode)) {
fs/xfs/libxfs/xfs_ialloc.c
2340
xfs_warn(mp, "%s: agno != pag_agno(pag) (%d != %d).",
fs/xfs/libxfs/xfs_ialloc.c
2341
__func__, XFS_INO_TO_AGNO(mp, inode), pag_agno(pag));
fs/xfs/libxfs/xfs_ialloc.c
2345
agino = XFS_INO_TO_AGINO(mp, inode);
fs/xfs/libxfs/xfs_ialloc.c
2347
xfs_warn(mp, "%s: inode != xfs_agino_to_ino() (%llu != %llu).",
fs/xfs/libxfs/xfs_ialloc.c
2353
agbno = XFS_AGINO_TO_AGBNO(mp, agino);
fs/xfs/libxfs/xfs_ialloc.c
2354
if (agbno >= xfs_ag_block_count(mp, pag_agno(pag))) {
fs/xfs/libxfs/xfs_ialloc.c
2355
xfs_warn(mp, "%s: agbno >= xfs_ag_block_count (%d >= %d).",
fs/xfs/libxfs/xfs_ialloc.c
2356
__func__, agbno, xfs_ag_block_count(mp, pag_agno(pag)));
fs/xfs/libxfs/xfs_ialloc.c
2365
xfs_warn(mp, "%s: xfs_ialloc_read_agi() returned error %d.",
fs/xfs/libxfs/xfs_ialloc.c
2380
if (xfs_has_finobt(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
2402
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
2411
xfs_alert(mp,
fs/xfs/libxfs/xfs_ialloc.c
2439
rec.ir_startino + M_IGEO(mp)->ialloc_inos <= agino)
fs/xfs/libxfs/xfs_ialloc.c
2447
*chunk_agbno = XFS_AGINO_TO_AGBNO(mp, rec.ir_startino);
fs/xfs/libxfs/xfs_ialloc.c
2463
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
2477
agino = XFS_INO_TO_AGINO(mp, ino);
fs/xfs/libxfs/xfs_ialloc.c
2478
agbno = XFS_AGINO_TO_AGBNO(mp, agino);
fs/xfs/libxfs/xfs_ialloc.c
2479
if (agbno >= xfs_ag_block_count(mp, pag_agno(pag)) ||
fs/xfs/libxfs/xfs_ialloc.c
2489
if (agbno >= xfs_ag_block_count(mp, pag_agno(pag))) {
fs/xfs/libxfs/xfs_ialloc.c
2490
xfs_alert(mp,
fs/xfs/libxfs/xfs_ialloc.c
2493
(unsigned long)xfs_ag_block_count(mp,
fs/xfs/libxfs/xfs_ialloc.c
2497
xfs_alert(mp,
fs/xfs/libxfs/xfs_ialloc.c
2526
if (M_IGEO(mp)->blocks_per_cluster == 1) {
fs/xfs/libxfs/xfs_ialloc.c
2527
offset = XFS_INO_TO_OFFSET(mp, ino);
fs/xfs/libxfs/xfs_ialloc.c
2528
ASSERT(offset < mp->m_sb.sb_inopblock);
fs/xfs/libxfs/xfs_ialloc.c
2531
imap->im_len = XFS_FSB_TO_BB(mp, 1);
fs/xfs/libxfs/xfs_ialloc.c
2533
mp->m_sb.sb_inodelog);
fs/xfs/libxfs/xfs_ialloc.c
2542
if (M_IGEO(mp)->inoalign_mask) {
fs/xfs/libxfs/xfs_ialloc.c
2543
offset_agbno = agbno & M_IGEO(mp)->inoalign_mask;
fs/xfs/libxfs/xfs_ialloc.c
2555
((offset_agbno / M_IGEO(mp)->blocks_per_cluster) *
fs/xfs/libxfs/xfs_ialloc.c
2556
M_IGEO(mp)->blocks_per_cluster);
fs/xfs/libxfs/xfs_ialloc.c
2557
offset = ((agbno - cluster_agbno) * mp->m_sb.sb_inopblock) +
fs/xfs/libxfs/xfs_ialloc.c
2558
XFS_INO_TO_OFFSET(mp, ino);
fs/xfs/libxfs/xfs_ialloc.c
2561
imap->im_len = XFS_FSB_TO_BB(mp, M_IGEO(mp)->blocks_per_cluster);
fs/xfs/libxfs/xfs_ialloc.c
2562
imap->im_boffset = (unsigned short)(offset << mp->m_sb.sb_inodelog);
fs/xfs/libxfs/xfs_ialloc.c
2571
XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks)) {
fs/xfs/libxfs/xfs_ialloc.c
2572
xfs_alert(mp,
fs/xfs/libxfs/xfs_ialloc.c
2576
XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks));
fs/xfs/libxfs/xfs_ialloc.c
2653
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_ialloc.c
2660
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
2661
if (!uuid_equal(&agi->agi_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_ialloc.c
2663
if (!xfs_log_check_lsn(mp, be64_to_cpu(agi->agi_lsn)))
fs/xfs/libxfs/xfs_ialloc.c
2680
be32_to_cpu(agi->agi_level) > M_IGEO(mp)->inobt_maxlevels)
fs/xfs/libxfs/xfs_ialloc.c
2683
if (xfs_has_finobt(mp) &&
fs/xfs/libxfs/xfs_ialloc.c
2685
be32_to_cpu(agi->agi_free_level) > M_IGEO(mp)->inobt_maxlevels))
fs/xfs/libxfs/xfs_ialloc.c
2691
if (!xfs_verify_ino(mp, be32_to_cpu(agi->agi_unlinked[i])))
fs/xfs/libxfs/xfs_ialloc.c
2702
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_ialloc.c
2705
if (xfs_has_crc(mp) &&
fs/xfs/libxfs/xfs_ialloc.c
2710
if (fa || XFS_TEST_ERROR(mp, XFS_ERRTAG_IALLOC_READ_AGI))
fs/xfs/libxfs/xfs_ialloc.c
2719
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_ialloc.c
2730
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_ialloc.c
2756
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
2761
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_ialloc.c
2762
XFS_AG_DADDR(mp, pag_agno(pag), XFS_AGI_DADDR(mp)),
fs/xfs/libxfs/xfs_ialloc.c
2763
XFS_FSS_TO_BB(mp, 1), flags, agibpp, &xfs_agi_buf_ops);
fs/xfs/libxfs/xfs_ialloc.c
297
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc.c
2979
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_ialloc.c
2981
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/libxfs/xfs_ialloc.c
2982
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/libxfs/xfs_ialloc.c
2987
if (xfs_has_bigtime(mp))
fs/xfs/libxfs/xfs_ialloc.c
2989
if (xfs_has_large_extent_counts(mp))
fs/xfs/libxfs/xfs_ialloc.c
2994
igeo->inobt_mxr[0] = xfs_inobt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_ialloc.c
2995
igeo->inobt_mxr[1] = xfs_inobt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_ialloc.c
3009
inodes = (1LL << XFS_INO_AGINO_BITS(mp)) >> XFS_INODES_PER_CHUNK_LOG;
fs/xfs/libxfs/xfs_ialloc.c
3028
igeo->maxicount = XFS_FSB_TO_INO(mp,
fs/xfs/libxfs/xfs_ialloc.c
3045
if (xfs_has_v3inodes(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
3048
new_size *= mp->m_sb.sb_inodesize / XFS_DINODE_MIN_SIZE;
fs/xfs/libxfs/xfs_ialloc.c
3049
if (mp->m_sb.sb_inoalignmt >= XFS_B_TO_FSBT(mp, new_size))
fs/xfs/libxfs/xfs_ialloc.c
3054
if (igeo->inode_cluster_size_raw > mp->m_sb.sb_blocksize)
fs/xfs/libxfs/xfs_ialloc.c
3055
igeo->blocks_per_cluster = XFS_B_TO_FSBT(mp,
fs/xfs/libxfs/xfs_ialloc.c
3059
igeo->inode_cluster_size = XFS_FSB_TO_B(mp, igeo->blocks_per_cluster);
fs/xfs/libxfs/xfs_ialloc.c
3060
igeo->inodes_per_cluster = XFS_FSB_TO_INO(mp, igeo->blocks_per_cluster);
fs/xfs/libxfs/xfs_ialloc.c
3063
if (xfs_has_align(mp) &&
fs/xfs/libxfs/xfs_ialloc.c
3064
mp->m_sb.sb_inoalignmt >= igeo->blocks_per_cluster)
fs/xfs/libxfs/xfs_ialloc.c
3065
igeo->cluster_align = mp->m_sb.sb_inoalignmt;
fs/xfs/libxfs/xfs_ialloc.c
3069
igeo->cluster_align_inodes = XFS_FSB_TO_INO(mp, igeo->cluster_align);
fs/xfs/libxfs/xfs_ialloc.c
3075
if (mp->m_dalign && igeo->inoalign_mask &&
fs/xfs/libxfs/xfs_ialloc.c
3076
!(mp->m_dalign & igeo->inoalign_mask))
fs/xfs/libxfs/xfs_ialloc.c
3077
igeo->ialloc_align = mp->m_dalign;
fs/xfs/libxfs/xfs_ialloc.c
3081
if (mp->m_sb.sb_blocksize > PAGE_SIZE)
fs/xfs/libxfs/xfs_ialloc.c
3082
igeo->min_folio_order = mp->m_sb.sb_blocklog - PAGE_SHIFT;
fs/xfs/libxfs/xfs_ialloc.c
3090
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc.c
3093
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/libxfs/xfs_ialloc.c
3104
first_bno = howmany(4 * mp->m_sb.sb_sectsize, mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_ialloc.c
3113
first_bno += xfs_alloc_min_freelist(mp, NULL);
fs/xfs/libxfs/xfs_ialloc.c
3116
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_ialloc.c
3120
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_ialloc.c
3124
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_ialloc.c
3134
if (xfs_ag_contains_log(mp, 0))
fs/xfs/libxfs/xfs_ialloc.c
3135
first_bno += mp->m_sb.sb_logblocks;
fs/xfs/libxfs/xfs_ialloc.c
3141
if (xfs_has_dalign(mp) && igeo->ialloc_align > 0)
fs/xfs/libxfs/xfs_ialloc.c
3143
else if (xfs_has_align(mp) &&
fs/xfs/libxfs/xfs_ialloc.c
3144
mp->m_sb.sb_inoalignmt > 1)
fs/xfs/libxfs/xfs_ialloc.c
3145
first_bno = roundup(first_bno, mp->m_sb.sb_inoalignmt);
fs/xfs/libxfs/xfs_ialloc.c
3147
return XFS_AGINO_TO_INO(mp, 0, XFS_AGB_TO_AGINO(mp, first_bno));
fs/xfs/libxfs/xfs_ialloc.c
320
nbufs = length / M_IGEO(mp)->blocks_per_cluster;
fs/xfs/libxfs/xfs_ialloc.c
341
if (xfs_has_v3inodes(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
343
ino = XFS_AGINO_TO_INO(mp, agno, XFS_AGB_TO_AGINO(mp, agbno));
fs/xfs/libxfs/xfs_ialloc.c
355
mp->m_sb.sb_inodesize, length, gen);
fs/xfs/libxfs/xfs_ialloc.c
363
d = XFS_AGB_TO_DADDR(mp, agno, agbno +
fs/xfs/libxfs/xfs_ialloc.c
364
(j * M_IGEO(mp)->blocks_per_cluster));
fs/xfs/libxfs/xfs_ialloc.c
365
error = xfs_trans_get_buf(tp, mp->m_ddev_targp, d,
fs/xfs/libxfs/xfs_ialloc.c
366
mp->m_bsize * M_IGEO(mp)->blocks_per_cluster,
fs/xfs/libxfs/xfs_ialloc.c
374
for (i = 0; i < M_IGEO(mp)->inodes_per_cluster; i++) {
fs/xfs/libxfs/xfs_ialloc.c
375
int ioffset = i << mp->m_sb.sb_inodelog;
fs/xfs/libxfs/xfs_ialloc.c
377
free = xfs_make_iptr(mp, fbuf, i);
fs/xfs/libxfs/xfs_ialloc.c
387
&mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_ialloc.c
388
xfs_dinode_calc_crc(mp, free);
fs/xfs/libxfs/xfs_ialloc.c
392
ioffset + XFS_DINODE_SIZE(mp) - 1);
fs/xfs/libxfs/xfs_ialloc.c
449
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc.c
457
agbno = XFS_AGINO_TO_AGBNO(mp, *startino);
fs/xfs/libxfs/xfs_ialloc.c
458
mod = agbno % mp->m_sb.sb_inoalignmt;
fs/xfs/libxfs/xfs_ialloc.c
463
offset = XFS_AGB_TO_AGINO(mp, mod);
fs/xfs/libxfs/xfs_ialloc.c
556
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
575
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
590
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
595
if (XFS_IS_CORRUPT(mp, rec.ir_startino != nrec->ir_startino)) {
fs/xfs/libxfs/xfs_ialloc.c
605
if (XFS_IS_CORRUPT(mp, !__xfs_inobt_can_merge(nrec, &rec))) {
fs/xfs/libxfs/xfs_ialloc.c
618
error = xfs_inobt_rec_check_count(mp, nrec);
fs/xfs/libxfs/xfs_ialloc.c
650
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc.c
668
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_ialloc.c
714
args.mp = tp->t_mountp;
fs/xfs/libxfs/xfs_ialloc.c
732
percpu_counter_read_positive(&args.mp->m_icount) + newlen >
fs/xfs/libxfs/xfs_ialloc.c
743
args.agbno = XFS_AGINO_TO_AGBNO(args.mp, newino) +
fs/xfs/libxfs/xfs_ialloc.c
77
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc.c
798
ASSERT(!xfs_has_noalign(args.mp));
fs/xfs/libxfs/xfs_ialloc.c
799
args.alignment = args.mp->m_dalign;
fs/xfs/libxfs/xfs_ialloc.c
82
if (xfs_has_sparseinodes(mp)) {
fs/xfs/libxfs/xfs_ialloc.c
835
if (xfs_has_sparseinodes(args.mp) &&
fs/xfs/libxfs/xfs_ialloc.c
839
args.alignment = args.mp->m_sb.sb_spino_align;
fs/xfs/libxfs/xfs_ialloc.c
857
args.min_agbno = args.mp->m_sb.sb_inoalignmt;
fs/xfs/libxfs/xfs_ialloc.c
858
args.max_agbno = round_down(xfs_ag_block_count(args.mp,
fs/xfs/libxfs/xfs_ialloc.c
860
args.mp->m_sb.sb_inoalignmt) - 1;
fs/xfs/libxfs/xfs_ialloc.c
868
newlen = XFS_AGB_TO_AGINO(args.mp, args.len);
fs/xfs/libxfs/xfs_ialloc.c
887
error = xfs_ialloc_inode_init(args.mp, tp, NULL, newlen, pag_agno(pag),
fs/xfs/libxfs/xfs_ialloc.c
895
newino = XFS_AGB_TO_AGINO(args.mp, args.agbno);
fs/xfs/libxfs/xfs_ialloc.c
901
xfs_align_sparse_ino(args.mp, &newino, &allocmask);
fs/xfs/libxfs/xfs_ialloc.c
916
xfs_alert(args.mp,
fs/xfs/libxfs/xfs_ialloc.c
920
xfs_force_shutdown(args.mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/libxfs/xfs_ialloc.c
936
if (xfs_has_finobt(args.mp)) {
fs/xfs/libxfs/xfs_ialloc.c
947
if (xfs_has_finobt(args.mp)) {
fs/xfs/libxfs/xfs_ialloc.h
111
int xfs_ialloc_cluster_alignment(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_ialloc.h
112
void xfs_ialloc_setup_geometry(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_ialloc.h
113
xfs_ino_t xfs_ialloc_calc_rootino(struct xfs_mount *mp, int sunit);
fs/xfs/libxfs/xfs_ialloc.h
31
xfs_make_iptr(struct xfs_mount *mp, struct xfs_buf *b, int o)
fs/xfs/libxfs/xfs_ialloc.h
33
return xfs_buf_offset(b, o << (mp)->m_sb.sb_inodelog);
fs/xfs/libxfs/xfs_ialloc.h
90
int xfs_ialloc_inode_init(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/libxfs/xfs_ialloc.h
97
void xfs_inobt_btrec_to_irec(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc_btree.c
114
args.mp = cur->bc_mp;
fs/xfs/libxfs/xfs_ialloc_btree.c
133
new->s = cpu_to_be32(XFS_FSB_TO_AGBNO(args.mp, args.fsbno));
fs/xfs/libxfs/xfs_ialloc_btree.c
294
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_ialloc_btree.c
312
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_ialloc_btree.c
320
if (level >= M_IGEO(mp)->inobt_maxlevels)
fs/xfs/libxfs/xfs_ialloc_btree.c
324
M_IGEO(mp)->inobt_mxr[level != 0]);
fs/xfs/libxfs/xfs_ialloc_btree.c
482
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc_btree.c
485
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_inobt_ops,
fs/xfs/libxfs/xfs_ialloc_btree.c
486
M_IGEO(mp)->inobt_maxlevels, xfs_inobt_cur_cache);
fs/xfs/libxfs/xfs_ialloc_btree.c
508
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc_btree.c
511
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_finobt_ops,
fs/xfs/libxfs/xfs_ialloc_btree.c
512
M_IGEO(mp)->inobt_maxlevels, xfs_inobt_cur_cache);
fs/xfs/libxfs/xfs_ialloc_btree.c
578
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc_btree.c
582
blocklen -= XFS_INOBT_BLOCK_LEN(mp);
fs/xfs/libxfs/xfs_ialloc_btree.c
690
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc_btree.c
719
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_ialloc_btree.c
723
if (M_IGEO(mp)->inobt_mxr[0] == 0)
fs/xfs/libxfs/xfs_ialloc_btree.c
731
if (xfs_ag_contains_log(mp, pag_agno(pag)))
fs/xfs/libxfs/xfs_ialloc_btree.c
732
agblocks -= mp->m_sb.sb_logblocks;
fs/xfs/libxfs/xfs_ialloc_btree.c
734
return xfs_btree_calc_size(M_IGEO(mp)->inobt_mnr,
fs/xfs/libxfs/xfs_ialloc_btree.c
735
(uint64_t)agblocks * mp->m_sb.sb_inopblock /
fs/xfs/libxfs/xfs_ialloc_btree.c
815
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_ialloc_btree.c
818
return xfs_btree_calc_size(M_IGEO(mp)->inobt_mnr, len);
fs/xfs/libxfs/xfs_ialloc_btree.h
21
#define XFS_INOBT_BLOCK_LEN(mp) \
fs/xfs/libxfs/xfs_ialloc_btree.h
22
(xfs_has_crc(((mp))) ? \
fs/xfs/libxfs/xfs_ialloc_btree.h
30
#define XFS_INOBT_REC_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_ialloc_btree.h
33
XFS_INOBT_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_ialloc_btree.h
36
#define XFS_INOBT_KEY_ADDR(mp, block, index) \
fs/xfs/libxfs/xfs_ialloc_btree.h
39
XFS_INOBT_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_ialloc_btree.h
42
#define XFS_INOBT_PTR_ADDR(mp, block, index, maxrecs) \
fs/xfs/libxfs/xfs_ialloc_btree.h
45
XFS_INOBT_BLOCK_LEN(mp) + \
fs/xfs/libxfs/xfs_ialloc_btree.h
53
unsigned int xfs_inobt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_ialloc_btree.h
63
#define xfs_inobt_rec_check_count(mp, rec) 0
fs/xfs/libxfs/xfs_ialloc_btree.h
68
extern xfs_extlen_t xfs_iallocbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
132
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
139
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, imap->im_blkno,
fs/xfs/libxfs/xfs_inode_buf.c
142
xfs_agno_mark_sick(mp, xfs_daddr_to_agno(mp, imap->im_blkno),
fs/xfs/libxfs/xfs_inode_buf.c
378
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
384
uint32_t fork_size = XFS_DFORK_SIZE(dip, mp, whichfork);
fs/xfs/libxfs/xfs_inode_buf.c
442
if (di_nextents > XFS_DFORK_MAXEXT(dip, mp, whichfork))
fs/xfs/libxfs/xfs_inode_buf.c
453
if (!xfs_has_metadir(mp))
fs/xfs/libxfs/xfs_inode_buf.c
46
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_inode_buf.c
464
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_inode_buf.c
469
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_inode_buf.c
485
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_inode_buf.c
496
if (!xfs_has_metadir(mp) || !xfs_has_parent(mp))
fs/xfs/libxfs/xfs_inode_buf.c
502
if (dip->di_forkoff >= (XFS_LITINO(mp) >> 3))
fs/xfs/libxfs/xfs_inode_buf.c
513
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
517
if (!xfs_has_large_extent_counts(mp))
fs/xfs/libxfs/xfs_inode_buf.c
53
ni = XFS_BB_TO_FSB(mp, bp->b_length) * mp->m_sb.sb_inopblock;
fs/xfs/libxfs/xfs_inode_buf.c
535
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
541
if (!xfs_has_metadir(mp))
fs/xfs/libxfs/xfs_inode_buf.c
59
dip = xfs_buf_offset(bp, (i << mp->m_sb.sb_inodelog));
fs/xfs/libxfs/xfs_inode_buf.c
594
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
612
if (!xfs_has_v3inodes(mp))
fs/xfs/libxfs/xfs_inode_buf.c
614
if (!xfs_verify_cksum((char *)dip, mp->m_sb.sb_inodesize,
fs/xfs/libxfs/xfs_inode_buf.c
619
if (!uuid_equal(&dip->di_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_inode_buf.c
62
xfs_dinode_good_version(mp, dip->di_version) &&
fs/xfs/libxfs/xfs_inode_buf.c
65
XFS_TEST_ERROR(mp, XFS_ERRTAG_ITOBP_INOTOBP))) {
fs/xfs/libxfs/xfs_inode_buf.c
664
fa = xfs_dinode_verify_nrext64(mp, dip);
fs/xfs/libxfs/xfs_inode_buf.c
676
if (S_ISDIR(mode) && nextents > mp->m_dir_geo->max_extents)
fs/xfs/libxfs/xfs_inode_buf.c
679
if (mode && XFS_DFORK_BOFF(dip) > mp->m_sb.sb_inodesize)
fs/xfs/libxfs/xfs_inode_buf.c
684
if (mode && (flags & XFS_DIFLAG_REALTIME) && !mp->m_rtdev_targp)
fs/xfs/libxfs/xfs_inode_buf.c
688
fa = xfs_dinode_verify_forkoff(dip, mp);
fs/xfs/libxfs/xfs_inode_buf.c
704
fa = xfs_dinode_verify_fork(dip, mp, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_inode_buf.c
716
fa = xfs_dinode_verify_fork(dip, mp, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_inode_buf.c
73
xfs_alert(mp,
fs/xfs/libxfs/xfs_inode_buf.c
738
fa = xfs_inode_validate_extsize(mp, be32_to_cpu(dip->di_extsize),
fs/xfs/libxfs/xfs_inode_buf.c
751
!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_inode_buf.c
760
!xfs_has_rtreflink(mp))
fs/xfs/libxfs/xfs_inode_buf.c
763
if (xfs_has_zoned(mp) &&
fs/xfs/libxfs/xfs_inode_buf.c
765
if (be32_to_cpu(dip->di_used_blocks) > mp->m_sb.sb_rgextents)
fs/xfs/libxfs/xfs_inode_buf.c
769
fa = xfs_inode_validate_cowextsize(mp,
fs/xfs/libxfs/xfs_inode_buf.c
778
!xfs_has_bigtime(mp))
fs/xfs/libxfs/xfs_inode_buf.c
782
fa = xfs_dinode_verify_metadir(mp, dip, mode, flags, flags2);
fs/xfs/libxfs/xfs_inode_buf.c
798
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
806
ASSERT(xfs_has_crc(mp));
fs/xfs/libxfs/xfs_inode_buf.c
807
crc = xfs_start_cksum_update((char *)dip, mp->m_sb.sb_inodesize,
fs/xfs/libxfs/xfs_inode_buf.c
829
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
843
extsize_bytes = XFS_FSB_TO_B(mp, extsize);
fs/xfs/libxfs/xfs_inode_buf.c
872
blocksize_bytes = XFS_FSB_TO_B(mp, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_inode_buf.c
874
blocksize_bytes = mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_inode_buf.c
898
if (!rt_flag && extsize > mp->m_sb.sb_agblocks / 2)
fs/xfs/libxfs/xfs_inode_buf.c
919
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.c
932
cowextsize_bytes = XFS_FSB_TO_B(mp, cowextsize);
fs/xfs/libxfs/xfs_inode_buf.c
947
blocksize_bytes = XFS_FSB_TO_B(mp, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_inode_buf.c
949
blocksize_bytes = mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_inode_buf.c
951
if (hint_flag && !xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_inode_buf.c
970
if (!rt_flag && cowextsize > mp->m_sb.sb_agblocks / 2)
fs/xfs/libxfs/xfs_inode_buf.h
22
int xfs_imap_to_bp(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/libxfs/xfs_inode_buf.h
24
void xfs_dinode_calc_crc(struct xfs_mount *mp, struct xfs_dinode *dip);
fs/xfs/libxfs/xfs_inode_buf.h
29
xfs_failaddr_t xfs_dinode_verify(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/libxfs/xfs_inode_buf.h
31
xfs_failaddr_t xfs_dinode_verify_metadir(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.h
34
xfs_failaddr_t xfs_inode_validate_extsize(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.h
36
xfs_failaddr_t xfs_inode_validate_cowextsize(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_buf.h
49
xfs_dinode_good_version(struct xfs_mount *mp, uint8_t version)
fs/xfs/libxfs/xfs_inode_buf.h
51
if (xfs_has_v3inodes(mp))
fs/xfs/libxfs/xfs_inode_fork.c
113
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
127
if (unlikely(size < 0 || size > XFS_DFORK_SIZE(dip, mp, whichfork))) {
fs/xfs/libxfs/xfs_inode_fork.c
180
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
190
size = xfs_bmap_broot_space(mp, dfp);
fs/xfs/libxfs/xfs_inode_fork.c
204
XFS_DFORK_SIZE(dip, mp, whichfork) ||
fs/xfs/libxfs/xfs_inode_fork.c
206
level == 0 || level > XFS_BM_MAXLEVELS(mp, whichfork)) {
fs/xfs/libxfs/xfs_inode_fork.c
207
xfs_warn(mp, "corrupt inode %llu (btree).",
fs/xfs/libxfs/xfs_inode_fork.c
556
xfs_mount_t *mp;
fs/xfs/libxfs/xfs_inode_fork.c
576
mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
602
xfs_bmbt_to_bmdr(mp, ifp->if_broot, ifp->if_broot_bytes,
fs/xfs/libxfs/xfs_inode_fork.c
604
XFS_DFORK_SIZE(dip, mp, whichfork));
fs/xfs/libxfs/xfs_inode_fork.c
678
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
682
fa = xfs_dir2_sf_verify(mp, sfp, ifp->if_bytes);
fs/xfs/libxfs/xfs_inode_fork.c
743
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_fork.c
759
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_REDUCE_MAX_IEXTENTS) && nr_exts > 10)
fs/xfs/libxfs/xfs_inode_fork.c
763
if (has_large || !xfs_has_large_extent_counts(mp))
fs/xfs/libxfs/xfs_inode_util.c
252
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_inode_util.c
258
if (!(args->flags & XFS_ICREATE_UNLINKABLE) && xfs_has_parent(mp))
fs/xfs/libxfs/xfs_inode_util.c
273
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_inode_util.c
295
if (dir && !(dir->i_mode & S_ISGID) && xfs_has_grpid(mp)) {
fs/xfs/libxfs/xfs_inode_util.c
312
if (xfs_has_v3inodes(mp)) {
fs/xfs/libxfs/xfs_inode_util.c
346
if (xfs_icreate_want_attrfork(mp, args)) {
fs/xfs/libxfs/xfs_inode_util.c
350
if (!xfs_has_attr(mp)) {
fs/xfs/libxfs/xfs_inode_util.c
351
spin_lock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_inode_util.c
352
xfs_add_attr(mp);
fs/xfs/libxfs/xfs_inode_util.c
353
spin_unlock(&mp->m_sb_lock);
fs/xfs/libxfs/xfs_inode_util.c
465
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_inode_util.c
468
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
523
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_inode_util.c
532
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/libxfs/xfs_inode_util.c
552
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_inode_util.c
554
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/libxfs/xfs_inode_util.c
567
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/libxfs/xfs_inode_util.c
700
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_inode_util.c
730
ip->i_diflags2 = mp->m_ino_geo.new_diflags2;
fs/xfs/libxfs/xfs_log_format.h
476
#define xfs_log_dinode_size(mp) \
fs/xfs/libxfs/xfs_log_format.h
477
(xfs_has_v3inodes((mp)) ? \
fs/xfs/libxfs/xfs_log_recover.h
130
int xlog_recover_iget(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/libxfs/xfs_log_recover.h
132
int xlog_recover_iget_handle(struct xfs_mount *mp, xfs_ino_t ino, uint32_t gen,
fs/xfs/libxfs/xfs_log_rlimit.c
106
if (xfs_has_rmapbt(mp) && xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_log_rlimit.c
107
mp->m_rmap_maxlevels = XFS_OLD_REFLINK_RMAP_MAXLEVELS;
fs/xfs/libxfs/xfs_log_rlimit.c
109
xfs_trans_resv_calc(mp, resv);
fs/xfs/libxfs/xfs_log_rlimit.c
112
resv->tr_atomic_ioend = M_RES(mp)->tr_atomic_ioend;
fs/xfs/libxfs/xfs_log_rlimit.c
114
if (xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_log_rlimit.c
123
} else if (xfs_has_rmapbt(mp)) {
fs/xfs/libxfs/xfs_log_rlimit.c
139
xfs_calc_write_reservation_minlogsize(mp);
fs/xfs/libxfs/xfs_log_rlimit.c
141
xfs_calc_itruncate_reservation_minlogsize(mp);
fs/xfs/libxfs/xfs_log_rlimit.c
143
xfs_calc_qm_dqalloc_reservation_minlogsize(mp);
fs/xfs/libxfs/xfs_log_rlimit.c
146
mp->m_rmap_maxlevels = rmap_maxlevels;
fs/xfs/libxfs/xfs_log_rlimit.c
156
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_log_rlimit.c
166
attr_space = xfs_log_calc_max_attrsetm_res(mp);
fs/xfs/libxfs/xfs_log_rlimit.c
168
xfs_log_calc_trans_resv_for_minlogblocks(mp, &resv);
fs/xfs/libxfs/xfs_log_rlimit.c
177
trace_xfs_trans_resv_calc_minlogsize(mp, i, resp);
fs/xfs/libxfs/xfs_log_rlimit.c
188
trace_xfs_log_get_max_trans_res(mp, max_resp);
fs/xfs/libxfs/xfs_log_rlimit.c
199
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_log_rlimit.c
206
xfs_log_get_max_trans_res(mp, &tres);
fs/xfs/libxfs/xfs_log_rlimit.c
208
max_logres = xfs_log_calc_unit_res(mp, tres.tr_logres);
fs/xfs/libxfs/xfs_log_rlimit.c
212
if (xfs_has_logv2(mp) && mp->m_sb.sb_logsunit > 1)
fs/xfs/libxfs/xfs_log_rlimit.c
213
lsunit = BTOBB(mp->m_sb.sb_logsunit);
fs/xfs/libxfs/xfs_log_rlimit.c
250
return XFS_BB_TO_FSB(mp, min_logblks);
fs/xfs/libxfs/xfs_log_rlimit.c
53
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_log_rlimit.c
58
size = xfs_attr_leaf_entsize_local_max(mp->m_attr_geo->blksize) -
fs/xfs/libxfs/xfs_log_rlimit.c
60
nblks = XFS_DAENTER_SPACE_RES(mp, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_log_rlimit.c
61
nblks += XFS_B_TO_FSB(mp, size);
fs/xfs/libxfs/xfs_log_rlimit.c
68
if (xfs_want_minlogsize_fixes(&mp->m_sb))
fs/xfs/libxfs/xfs_log_rlimit.c
69
size = XFS_B_TO_FSB(mp, size);
fs/xfs/libxfs/xfs_log_rlimit.c
71
nblks += XFS_NEXTENTADD_SPACE_RES(mp, size, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_log_rlimit.c
73
return M_RES(mp)->tr_attrsetm.tr_logres +
fs/xfs/libxfs/xfs_log_rlimit.c
74
M_RES(mp)->tr_attrsetrt.tr_logres * nblks;
fs/xfs/libxfs/xfs_log_rlimit.c
83
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_log_rlimit.c
86
unsigned int rmap_maxlevels = mp->m_rmap_maxlevels;
fs/xfs/libxfs/xfs_log_rlimit.c
92
if (xfs_want_minlogsize_fixes(&mp->m_sb)) {
fs/xfs/libxfs/xfs_log_rlimit.c
93
xfs_trans_resv_calc(mp, resv);
fs/xfs/libxfs/xfs_log_rlimit.c
94
resv->tr_atomic_ioend = M_RES(mp)->tr_atomic_ioend;
fs/xfs/libxfs/xfs_metadir.c
103
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_metadir.c
106
if (xfs_is_shutdown(mp))
fs/xfs/libxfs/xfs_metadir.c
113
if (!xfs_verify_ino(mp, args.inumber)) {
fs/xfs/libxfs/xfs_metadir.c
114
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_metadir.c
118
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_metadir.c
189
struct xfs_mount *mp = upd->dp->i_mount;
fs/xfs/libxfs/xfs_metadir.c
194
ASSERT(xfs_has_metadir(mp));
fs/xfs/libxfs/xfs_metadir.c
197
error = xfs_parent_start(mp, &upd->ppargs);
fs/xfs/libxfs/xfs_metadir.c
208
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_create,
fs/xfs/libxfs/xfs_metadir.c
209
xfs_create_space_res(mp, MAXNAMELEN), 0, 0, &upd->tp);
fs/xfs/libxfs/xfs_metadir.c
254
struct xfs_mount *mp = upd->dp->i_mount;
fs/xfs/libxfs/xfs_metadir.c
297
resblks = xfs_mkdir_space_res(mp, xname.len);
fs/xfs/libxfs/xfs_metadir.c
299
resblks = xfs_create_space_res(mp, xname.len);
fs/xfs/libxfs/xfs_metadir.c
324
struct xfs_mount *mp = upd->dp->i_mount;
fs/xfs/libxfs/xfs_metadir.c
331
ASSERT(xfs_has_metadir(mp));
fs/xfs/libxfs/xfs_metadir.c
333
error = xfs_parent_start(mp, &upd->ppargs);
fs/xfs/libxfs/xfs_metadir.c
337
resblks = xfs_link_space_res(mp, MAXNAMELEN);
fs/xfs/libxfs/xfs_metadir.c
338
error = xfs_trans_alloc_dir(upd->dp, &M_RES(mp)->tr_link, upd->ip,
fs/xfs/libxfs/xfs_metadir.c
378
struct xfs_mount *mp = upd->dp->i_mount;
fs/xfs/libxfs/xfs_metadir.c
400
resblks = xfs_link_space_res(mp, xname.len);
fs/xfs/libxfs/xfs_metadir.c
88
struct xfs_mount *mp = dp->i_mount;
fs/xfs/libxfs/xfs_metadir.c
92
.geo = mp->m_dir_geo,
fs/xfs/libxfs/xfs_metadir.c
95
.hashval = xfs_dir2_hashname(mp, xname),
fs/xfs/libxfs/xfs_metafile.c
105
return xfs_compare_freecounter(mp, XC_FREE_BLOCKS,
fs/xfs/libxfs/xfs_metafile.c
106
rhs - mp->m_metafile_resv_avail, 2048) >= 0;
fs/xfs/libxfs/xfs_metafile.c
116
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_metafile.c
118
ASSERT(xfs_has_metadir(mp));
fs/xfs/libxfs/xfs_metafile.c
120
trace_xfs_metafile_resv_critical(mp, 0);
fs/xfs/libxfs/xfs_metafile.c
122
if (!xfs_metafile_resv_can_cover(mp, mp->m_rtbtree_maxlevels))
fs/xfs/libxfs/xfs_metafile.c
125
if (!xfs_metafile_resv_can_cover(mp,
fs/xfs/libxfs/xfs_metafile.c
126
div_u64(mp->m_metafile_resv_target, 10)))
fs/xfs/libxfs/xfs_metafile.c
129
return XFS_TEST_ERROR(mp, XFS_ERRTAG_METAFILE_RESV_CRITICAL);
fs/xfs/libxfs/xfs_metafile.c
138
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_metafile.c
144
trace_xfs_metafile_resv_alloc_space(mp, args->len);
fs/xfs/libxfs/xfs_metafile.c
150
mutex_lock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
151
if (mp->m_metafile_resv_avail > 0) {
fs/xfs/libxfs/xfs_metafile.c
154
from_resv = min_t(int64_t, len, mp->m_metafile_resv_avail);
fs/xfs/libxfs/xfs_metafile.c
155
mp->m_metafile_resv_avail -= from_resv;
fs/xfs/libxfs/xfs_metafile.c
182
mp->m_metafile_resv_used += args->len;
fs/xfs/libxfs/xfs_metafile.c
183
mutex_unlock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
196
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_metafile.c
201
trace_xfs_metafile_resv_free_space(mp, len);
fs/xfs/libxfs/xfs_metafile.c
206
mutex_lock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
207
mp->m_metafile_resv_used -= len;
fs/xfs/libxfs/xfs_metafile.c
213
to_resv = mp->m_metafile_resv_target -
fs/xfs/libxfs/xfs_metafile.c
214
(mp->m_metafile_resv_used + mp->m_metafile_resv_avail);
fs/xfs/libxfs/xfs_metafile.c
217
mp->m_metafile_resv_avail += to_resv;
fs/xfs/libxfs/xfs_metafile.c
222
mutex_unlock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
234
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_metafile.c
236
if (mp->m_metafile_resv_avail) {
fs/xfs/libxfs/xfs_metafile.c
237
xfs_mod_sb_delalloc(mp, -(int64_t)mp->m_metafile_resv_avail);
fs/xfs/libxfs/xfs_metafile.c
238
xfs_add_fdblocks(mp, mp->m_metafile_resv_avail);
fs/xfs/libxfs/xfs_metafile.c
240
mp->m_metafile_resv_avail = 0;
fs/xfs/libxfs/xfs_metafile.c
241
mp->m_metafile_resv_used = 0;
fs/xfs/libxfs/xfs_metafile.c
242
mp->m_metafile_resv_target = 0;
fs/xfs/libxfs/xfs_metafile.c
248
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_metafile.c
250
if (!xfs_has_metadir(mp))
fs/xfs/libxfs/xfs_metafile.c
253
trace_xfs_metafile_resv_free(mp, 0);
fs/xfs/libxfs/xfs_metafile.c
255
mutex_lock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
256
__xfs_metafile_resv_free(mp);
fs/xfs/libxfs/xfs_metafile.c
257
mutex_unlock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
263
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_metafile.c
268
xfs_rfsblock_t dblocks_avail = mp->m_sb.sb_dblocks / 4;
fs/xfs/libxfs/xfs_metafile.c
271
if (!xfs_has_metadir(mp))
fs/xfs/libxfs/xfs_metafile.c
277
mutex_lock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
278
__xfs_metafile_resv_free(mp);
fs/xfs/libxfs/xfs_metafile.c
285
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/libxfs/xfs_metafile.c
286
if (xfs_has_rtrmapbt(mp)) {
fs/xfs/libxfs/xfs_metafile.c
288
target += xfs_rtrmapbt_calc_reserves(mp);
fs/xfs/libxfs/xfs_metafile.c
290
if (xfs_has_rtreflink(mp)) {
fs/xfs/libxfs/xfs_metafile.c
292
target += xfs_rtrefcountbt_calc_reserves(mp);
fs/xfs/libxfs/xfs_metafile.c
310
error = xfs_dec_fdblocks(mp, hidden_space, true);
fs/xfs/libxfs/xfs_metafile.c
312
trace_xfs_metafile_resv_init_error(mp, 0);
fs/xfs/libxfs/xfs_metafile.c
316
xfs_mod_sb_delalloc(mp, hidden_space);
fs/xfs/libxfs/xfs_metafile.c
318
mp->m_metafile_resv_target = target;
fs/xfs/libxfs/xfs_metafile.c
319
mp->m_metafile_resv_used = used;
fs/xfs/libxfs/xfs_metafile.c
320
mp->m_metafile_resv_avail = hidden_space;
fs/xfs/libxfs/xfs_metafile.c
322
trace_xfs_metafile_resv_init(mp, target);
fs/xfs/libxfs/xfs_metafile.c
325
mutex_unlock(&mp->m_metafile_resv_lock);
fs/xfs/libxfs/xfs_metafile.c
89
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_metafile.c
98
if (mp->m_metafile_resv_avail >= rhs)
fs/xfs/libxfs/xfs_metafile.h
29
bool xfs_metafile_resv_critical(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_metafile.h
34
void xfs_metafile_resv_free(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_metafile.h
35
int xfs_metafile_resv_init(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_metafile.h
41
int xfs_metafile_iget(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/libxfs/xfs_parent.c
102
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.c
117
return xfs_dir2_hashname(mp, &xname) ^
fs/xfs/libxfs/xfs_parent.c
124
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.c
143
return xfs_parent_hashval(mp, name, namelen, be64_to_cpu(rec->p_ino));
fs/xfs/libxfs/xfs_parent.c
272
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.c
287
if (!xfs_parent_valuecheck(mp, value, valuelen))
fs/xfs/libxfs/xfs_parent.c
322
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.c
330
if (!xfs_parent_valuecheck(mp, pptr, sizeof(*pptr)))
fs/xfs/libxfs/xfs_parent.c
75
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.c
81
if (!xfs_has_parent(mp))
fs/xfs/libxfs/xfs_parent.c
93
if (!xfs_verify_dir_ino(mp, be64_to_cpu(rec->p_ino)))
fs/xfs/libxfs/xfs_parent.h
12
bool xfs_parent_valuecheck(struct xfs_mount *mp, const void *value,
fs/xfs/libxfs/xfs_parent.h
15
xfs_dahash_t xfs_parent_hashval(struct xfs_mount *mp, const uint8_t *name,
fs/xfs/libxfs/xfs_parent.h
17
xfs_dahash_t xfs_parent_hashattr(struct xfs_mount *mp, const uint8_t *name,
fs/xfs/libxfs/xfs_parent.h
58
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.h
61
if (!xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_parent.h
75
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_parent.h
94
int xfs_parent_from_attr(struct xfs_mount *mp, unsigned int attr_flags,
fs/xfs/libxfs/xfs_quota_defs.h
131
extern xfs_failaddr_t xfs_dquot_verify(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_quota_defs.h
133
extern xfs_failaddr_t xfs_dqblk_verify(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_quota_defs.h
136
extern void xfs_dqblk_repair(struct xfs_mount *mp, struct xfs_dqblk *dqb,
fs/xfs/libxfs/xfs_quota_defs.h
184
int xfs_dqinode_mkdir_parent(struct xfs_mount *mp, struct xfs_inode **dpp);
fs/xfs/libxfs/xfs_quota_defs.h
60
#define XFS_IS_QUOTA_ON(mp) ((mp)->m_qflags & XFS_ALL_QUOTA_ACCT)
fs/xfs/libxfs/xfs_quota_defs.h
61
#define XFS_IS_UQUOTA_ON(mp) ((mp)->m_qflags & XFS_UQUOTA_ACCT)
fs/xfs/libxfs/xfs_quota_defs.h
62
#define XFS_IS_PQUOTA_ON(mp) ((mp)->m_qflags & XFS_PQUOTA_ACCT)
fs/xfs/libxfs/xfs_quota_defs.h
63
#define XFS_IS_GQUOTA_ON(mp) ((mp)->m_qflags & XFS_GQUOTA_ACCT)
fs/xfs/libxfs/xfs_quota_defs.h
64
#define XFS_IS_UQUOTA_ENFORCED(mp) ((mp)->m_qflags & XFS_UQUOTA_ENFD)
fs/xfs/libxfs/xfs_quota_defs.h
65
#define XFS_IS_GQUOTA_ENFORCED(mp) ((mp)->m_qflags & XFS_GQUOTA_ENFD)
fs/xfs/libxfs/xfs_quota_defs.h
66
#define XFS_IS_PQUOTA_ENFORCED(mp) ((mp)->m_qflags & XFS_PQUOTA_ENFD)
fs/xfs/libxfs/xfs_refcount.c
1358
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_refcount.c
1361
if (XFS_IS_CORRUPT(mp, !xfs_verify_agbext(pag, new_agbno,
fs/xfs/libxfs/xfs_refcount.c
1369
ASSERT(xfs_verify_fsbext(mp, ri->ri_startblock, ri->ri_blockcount));
fs/xfs/libxfs/xfs_refcount.c
1370
ASSERT(pag_agno(pag) == XFS_FSB_TO_AGNO(mp, ri->ri_startblock));
fs/xfs/libxfs/xfs_refcount.c
1388
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_refcount.c
1396
bno = XFS_FSB_TO_AGBNO(mp, ri->ri_startblock);
fs/xfs/libxfs/xfs_refcount.c
1398
trace_xfs_refcount_deferred(mp, ri);
fs/xfs/libxfs/xfs_refcount.c
1400
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_REFCOUNT_FINISH_ONE))
fs/xfs/libxfs/xfs_refcount.c
1422
*pcur = rcur = xfs_refcountbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/libxfs/xfs_refcount.c
1461
trace_xfs_refcount_finish_one_leftover(mp, ri);
fs/xfs/libxfs/xfs_refcount.c
1476
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_refcount.c
1479
if (XFS_IS_CORRUPT(mp, !xfs_verify_rgbext(rtg, new_agbno,
fs/xfs/libxfs/xfs_refcount.c
1487
ASSERT(xfs_verify_rtbext(mp, ri->ri_startblock, ri->ri_blockcount));
fs/xfs/libxfs/xfs_refcount.c
1501
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_refcount.c
1509
bno = xfs_rtb_to_rgbno(mp, ri->ri_startblock);
fs/xfs/libxfs/xfs_refcount.c
1511
trace_xfs_refcount_deferred(mp, ri);
fs/xfs/libxfs/xfs_refcount.c
1513
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_REFCOUNT_FINISH_ONE))
fs/xfs/libxfs/xfs_refcount.c
1570
trace_xfs_refcount_finish_one_leftover(mp, ri);
fs/xfs/libxfs/xfs_refcount.c
187
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_refcount.c
190
xfs_warn(mp,
fs/xfs/libxfs/xfs_refcount.c
194
xfs_warn(mp,
fs/xfs/libxfs/xfs_refcount.c
198
xfs_warn(mp,
fs/xfs/libxfs/xfs_refcount.c
1986
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_refcount.c
1988
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_refcount.c
2005
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_refcount.c
2007
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_refcount.c
2058
struct xfs_mount *mp = xg->xg_mount;
fs/xfs/libxfs/xfs_refcount.c
2080
if (!xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_refcount.c
2101
tp = xfs_trans_alloc_empty(mp);
fs/xfs/libxfs/xfs_refcount.c
2110
cur = xfs_refcountbt_init_cursor(mp, tp, agbp, to_perag(xg));
fs/xfs/libxfs/xfs_refcount.c
2128
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, 0, 0, 0, &tp);
fs/xfs/libxfs/xfs_refcount_btree.c
103
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_refcount_btree.c
106
xfs_fsblock_t fsbno = XFS_DADDR_TO_FSB(mp, xfs_buf_daddr(bp));
fs/xfs/libxfs/xfs_refcount_btree.c
208
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_refcount_btree.c
217
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_refcount_btree.c
238
} else if (level >= mp->m_refc_maxlevels)
fs/xfs/libxfs/xfs_refcount_btree.c
241
return xfs_btree_agblock_verify(bp, mp->m_refc_mxr[level != 0]);
fs/xfs/libxfs/xfs_refcount_btree.c
357
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.c
364
ASSERT(pag_agno(pag) < mp->m_sb.sb_agcount);
fs/xfs/libxfs/xfs_refcount_btree.c
366
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_refcountbt_ops,
fs/xfs/libxfs/xfs_refcount_btree.c
367
mp->m_refc_maxlevels, xfs_refcountbt_cur_cache);
fs/xfs/libxfs/xfs_refcount_btree.c
421
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.c
447
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_refcount_btree.c
449
if (!xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_refcount_btree.c
450
mp->m_refc_maxlevels = 0;
fs/xfs/libxfs/xfs_refcount_btree.c
454
mp->m_refc_maxlevels = xfs_btree_compute_maxlevels(
fs/xfs/libxfs/xfs_refcount_btree.c
455
mp->m_refc_mnr, mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_refcount_btree.c
456
ASSERT(mp->m_refc_maxlevels <= xfs_refcountbt_maxlevels_ondisk());
fs/xfs/libxfs/xfs_refcount_btree.c
462
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.c
465
return xfs_btree_calc_size(mp->m_refc_mnr, len);
fs/xfs/libxfs/xfs_refcount_btree.c
473
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.c
477
if (mp->m_refc_mxr[0] == 0)
fs/xfs/libxfs/xfs_refcount_btree.c
480
return xfs_refcountbt_calc_size(mp, agblocks);
fs/xfs/libxfs/xfs_refcount_btree.c
488
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.c
500
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_refcount_btree.c
517
if (xfs_ag_contains_log(mp, pag_agno(pag)))
fs/xfs/libxfs/xfs_refcount_btree.c
518
agblocks -= mp->m_sb.sb_logblocks;
fs/xfs/libxfs/xfs_refcount_btree.c
520
*ask += xfs_refcountbt_max_size(mp, agblocks);
fs/xfs/libxfs/xfs_refcount_btree.c
70
args.mp = cur->bc_mp;
fs/xfs/libxfs/xfs_refcount_btree.c
77
xfs_agbno_to_fsb(args.pag, xfs_refc_block(args.mp)));
fs/xfs/libxfs/xfs_refcount_btree.h
48
extern struct xfs_btree_cur *xfs_refcountbt_init_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.h
51
unsigned int xfs_refcountbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_refcount_btree.h
53
extern void xfs_refcountbt_compute_maxlevels(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_refcount_btree.h
55
extern xfs_extlen_t xfs_refcountbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.h
57
extern xfs_extlen_t xfs_refcountbt_max_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_refcount_btree.h
60
extern int xfs_refcountbt_calc_reserves(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap.c
1027
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
1066
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
1086
if (XFS_IS_CORRUPT(mp, have_gt != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1091
if (XFS_IS_CORRUPT(mp, bno + len > gtrec.rm_startblock)) {
fs/xfs/libxfs/xfs_rmap.c
1142
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1189
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1215
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rmap.c
1219
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_rmap.c
1222
cur = xfs_rmapbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/libxfs/xfs_rmap.c
1255
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
1284
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1321
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1326
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
1350
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1363
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1368
if (XFS_IS_CORRUPT(mp, bno + len > RIGHT.rm_startblock)) {
fs/xfs/libxfs/xfs_rmap.c
1398
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1418
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1429
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1437
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1448
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1456
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1479
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1487
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1507
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1518
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1526
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1596
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1639
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_rmap.c
1654
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1687
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1701
if (XFS_IS_CORRUPT(mp, i != 0)) {
fs/xfs/libxfs/xfs_rmap.c
1713
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1753
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
1783
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1810
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
1831
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1836
if (XFS_IS_CORRUPT(mp, bno + len > RIGHT.rm_startblock)) {
fs/xfs/libxfs/xfs_rmap.c
1886
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1913
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1940
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
1964
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2000
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2046
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2082
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2121
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
218
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/libxfs/xfs_rmap.c
2185
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
2208
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2216
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
2226
if (XFS_IS_CORRUPT(mp, owner != ltrec.rm_owner)) {
fs/xfs/libxfs/xfs_rmap.c
2233
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
2242
if (XFS_IS_CORRUPT(mp, ltrec.rm_offset > offset)) {
fs/xfs/libxfs/xfs_rmap.c
2247
if (XFS_IS_CORRUPT(mp, offset > ltoff + ltrec.rm_blockcount)) {
fs/xfs/libxfs/xfs_rmap.c
226
if (irec->rm_startblock <= XFS_AGFL_BLOCK(mp)) {
fs/xfs/libxfs/xfs_rmap.c
229
if (irec->rm_blockcount != XFS_AGFL_BLOCK(mp) + 1)
fs/xfs/libxfs/xfs_rmap.c
2304
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2334
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2377
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
238
if (!(xfs_verify_ino(mp, irec->rm_owner) ||
fs/xfs/libxfs/xfs_rmap.c
2411
if (XFS_IS_CORRUPT(mp, have_gt != 1)) {
fs/xfs/libxfs/xfs_rmap.c
2464
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
263
!xfs_verify_fileext(mp, irec->rm_offset, irec->rm_blockcount))
fs/xfs/libxfs/xfs_rmap.c
2686
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rmap.c
2691
trace_xfs_rmap_deferred(mp, ri);
fs/xfs/libxfs/xfs_rmap.c
2693
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_RMAP_FINISH_ONE))
fs/xfs/libxfs/xfs_rmap.c
2717
bno = xfs_fsb_to_gbno(mp, ri->ri_bmap.br_startblock,
fs/xfs/libxfs/xfs_rmap.c
2734
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap.c
2737
return xfs_has_rmapbt(mp) && whichfork != XFS_COW_FORK;
fs/xfs/libxfs/xfs_rmap.c
274
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rmap.c
2814
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap.c
2823
if (!xfs_rmap_update_is_needed(mp, whichfork))
fs/xfs/libxfs/xfs_rmap.c
285
if (irec->rm_blockcount != mp->m_sb.sb_rextsize)
fs/xfs/libxfs/xfs_rmap.c
289
if (!xfs_has_rtreflink(mp))
fs/xfs/libxfs/xfs_rmap.c
2947
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap.c
2950
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_rmap.c
307
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rmap.c
309
if (!xfs_verify_ino(mp, irec->rm_owner))
fs/xfs/libxfs/xfs_rmap.c
313
if (!xfs_verify_fileext(mp, irec->rm_offset, irec->rm_blockcount))
fs/xfs/libxfs/xfs_rmap.c
349
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
352
xfs_warn(mp,
fs/xfs/libxfs/xfs_rmap.c
355
xfs_warn(mp,
fs/xfs/libxfs/xfs_rmap.c
359
xfs_warn(mp,
fs/xfs/libxfs/xfs_rmap.c
362
xfs_warn(mp,
fs/xfs/libxfs/xfs_rmap.c
605
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
612
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
621
if (XFS_IS_CORRUPT(mp, owner != rec->rm_owner)) {
fs/xfs/libxfs/xfs_rmap.c
632
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
639
if (XFS_IS_CORRUPT(mp, rec->rm_offset > offset)) {
fs/xfs/libxfs/xfs_rmap.c
644
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
682
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rmap.c
707
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
726
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
755
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
765
if (XFS_IS_CORRUPT(mp,
fs/xfs/libxfs/xfs_rmap.c
788
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/libxfs/xfs_rmap.c
970
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rmap.c
974
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_rmap.c
977
cur = xfs_rmapbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/libxfs/xfs_rmap.h
186
void xfs_rmap_convert_extent(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/libxfs/xfs_rmap_btree.c
303
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rmap_btree.c
324
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_rmap_btree.c
345
} else if (level >= mp->m_rmap_maxlevels)
fs/xfs/libxfs/xfs_rmap_btree.c
348
return xfs_btree_agblock_verify(bp, mp->m_rmap_mxr[level != 0]);
fs/xfs/libxfs/xfs_rmap_btree.c
512
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
519
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_rmapbt_ops,
fs/xfs/libxfs/xfs_rmap_btree.c
520
mp->m_rmap_maxlevels, xfs_rmapbt_cur_cache);
fs/xfs/libxfs/xfs_rmap_btree.c
643
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
649
return xfbtree_init(mp, xfbt, btp, &xfs_rmapbt_mem_ops);
fs/xfs/libxfs/xfs_rmap_btree.c
715
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
753
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rmap_btree.c
755
if (!xfs_has_rmapbt(mp)) {
fs/xfs/libxfs/xfs_rmap_btree.c
756
mp->m_rmap_maxlevels = 0;
fs/xfs/libxfs/xfs_rmap_btree.c
760
if (xfs_has_reflink(mp)) {
fs/xfs/libxfs/xfs_rmap_btree.c
774
mp->m_rmap_maxlevels = xfs_btree_space_to_height(mp->m_rmap_mnr,
fs/xfs/libxfs/xfs_rmap_btree.c
775
mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_rmap_btree.c
781
mp->m_rmap_maxlevels = xfs_btree_compute_maxlevels(
fs/xfs/libxfs/xfs_rmap_btree.c
782
mp->m_rmap_mnr, mp->m_sb.sb_agblocks);
fs/xfs/libxfs/xfs_rmap_btree.c
784
ASSERT(mp->m_rmap_maxlevels <= xfs_rmapbt_maxlevels_ondisk());
fs/xfs/libxfs/xfs_rmap_btree.c
790
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
793
return xfs_btree_calc_size(mp->m_rmap_mnr, len);
fs/xfs/libxfs/xfs_rmap_btree.c
801
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
805
if (mp->m_rmap_mxr[0] == 0)
fs/xfs/libxfs/xfs_rmap_btree.c
808
return xfs_rmapbt_calc_size(mp, agblocks);
fs/xfs/libxfs/xfs_rmap_btree.c
816
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.c
828
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_rmap_btree.c
845
if (xfs_ag_contains_log(mp, pag_agno(pag)))
fs/xfs/libxfs/xfs_rmap_btree.c
846
agblocks -= mp->m_sb.sb_logblocks;
fs/xfs/libxfs/xfs_rmap_btree.c
849
*ask += max(agblocks / 100, xfs_rmapbt_max_size(mp, agblocks));
fs/xfs/libxfs/xfs_rmap_btree.h
45
struct xfs_btree_cur *xfs_rmapbt_init_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.h
50
unsigned int xfs_rmapbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_rmap_btree.h
52
extern void xfs_rmapbt_compute_maxlevels(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rmap_btree.h
54
extern xfs_extlen_t xfs_rmapbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.h
56
extern xfs_extlen_t xfs_rmapbt_max_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rmap_btree.h
59
extern int xfs_rmapbt_calc_reserves(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/libxfs/xfs_rmap_btree.h
69
int xfs_rmapbt_mem_init(struct xfs_mount *mp, struct xfbtree *xfbtree,
fs/xfs/libxfs/xfs_rtbitmap.c
1057
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rtbitmap.c
1060
.mp = mp,
fs/xfs/libxfs/xfs_rtbitmap.c
1070
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_FREE_EXTENT))
fs/xfs/libxfs/xfs_rtbitmap.c
1093
if (!xfs_has_rtgroups(mp) &&
fs/xfs/libxfs/xfs_rtbitmap.c
1094
tp->t_frextents_delta + mp->m_sb.sb_frextents ==
fs/xfs/libxfs/xfs_rtbitmap.c
1095
mp->m_sb.sb_rextents) {
fs/xfs/libxfs/xfs_rtbitmap.c
1122
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rtbitmap.c
1126
ASSERT(!xfs_has_zoned(mp));
fs/xfs/libxfs/xfs_rtbitmap.c
1129
mod = xfs_blen_to_rtxoff(mp, rtlen);
fs/xfs/libxfs/xfs_rtbitmap.c
1135
mod = xfs_rtb_to_rtxoff(mp, rtbno);
fs/xfs/libxfs/xfs_rtbitmap.c
1141
error = xfs_rtfree_extent(tp, rtg, xfs_rtb_to_rtx(mp, rtbno),
fs/xfs/libxfs/xfs_rtbitmap.c
1142
xfs_extlen_to_rtxlen(mp, rtlen));
fs/xfs/libxfs/xfs_rtbitmap.c
1146
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
1148
xfs_rtb_to_rgbno(mp, rtbno), rtlen, 0);
fs/xfs/libxfs/xfs_rtbitmap.c
1163
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.c
1166
.mp = mp,
fs/xfs/libxfs/xfs_rtbitmap.c
1178
if (xfs_has_zoned(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
1236
.mp = rtg_mount(rtg),
fs/xfs/libxfs/xfs_rtbitmap.c
1256
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtbitmap.c
1258
unsigned int rbmblock_bytes = mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_rtbitmap.c
1260
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
1272
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.c
1275
if (xfs_has_zoned(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
1277
return howmany_64(rtextents, xfs_rtbitmap_rtx_per_rbmblock(mp));
fs/xfs/libxfs/xfs_rtbitmap.c
1283
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtbitmap.c
1285
if (!mp->m_sb.sb_rextents)
fs/xfs/libxfs/xfs_rtbitmap.c
1289
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
1290
return mp->m_sb.sb_rgextents;
fs/xfs/libxfs/xfs_rtbitmap.c
1292
return mp->m_sb.sb_rextents;
fs/xfs/libxfs/xfs_rtbitmap.c
1300
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtbitmap.c
1302
return xfs_rtbitmap_blockcount_len(mp, xfs_rtbitmap_bitcount(mp));
fs/xfs/libxfs/xfs_rtbitmap.c
1311
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.c
1314
xfs_rtbxlen_t rextents = xfs_rtbitmap_bitcount(mp);
fs/xfs/libxfs/xfs_rtbitmap.c
1317
if (xfs_has_zoned(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.c
1323
rsumwords = xfs_rtbitmap_blockcount_len(mp, rextents) * (*rsumlevels);
fs/xfs/libxfs/xfs_rtbitmap.c
1324
return howmany_64(rsumwords, mp->m_blockwsize);
fs/xfs/libxfs/xfs_rtbitmap.c
1334
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtbitmap.c
1339
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_growrtalloc,
fs/xfs/libxfs/xfs_rtbitmap.c
1340
XFS_GROWFSRT_SPACE_RES(mp, count_fsb), 0, 0, &tp);
fs/xfs/libxfs/xfs_rtbitmap.c
1372
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.c
1377
const size_t copylen = mp->m_blockwsize << XFS_WORDLOG;
fs/xfs/libxfs/xfs_rtbitmap.c
1388
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_growrtzero, 0, 0, 0, &tp);
fs/xfs/libxfs/xfs_rtbitmap.c
1394
error = xfs_trans_get_buf(tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_rtbitmap.c
1395
XFS_FSB_TO_DADDR(mp, fsbno), mp->m_bsize, 0, &bp);
fs/xfs/libxfs/xfs_rtbitmap.c
1403
bp->b_ops = xfs_rtblock_ops(mp, type);
fs/xfs/libxfs/xfs_rtbitmap.c
1405
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.c
1413
hdr->rt_blkno = cpu_to_be64(XFS_FSB_TO_DADDR(mp, fsbno));
fs/xfs/libxfs/xfs_rtbitmap.c
1415
uuid_copy(&hdr->rt_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_rtbitmap.c
1424
xfs_trans_log_buf(tp, bp, 0, mp->m_sb.sb_blocksize - 1);
fs/xfs/libxfs/xfs_rtbitmap.c
1441
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.c
1442
const size_t copylen = mp->m_blockwsize << XFS_WORDLOG;
fs/xfs/libxfs/xfs_rtbitmap.c
1481
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.c
1483
ip->i_disk_size = mp->m_sb.sb_rbmblocks * mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_rtbitmap.c
1484
if (init && !xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.c
1499
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.c
1501
ip->i_disk_size = mp->m_rsumblocks * mp->m_sb.sb_blocksize;
fs/xfs/libxfs/xfs_rtbitmap.c
156
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
199
if (XFS_IS_CORRUPT(mp, nmap == 0 || !xfs_bmap_is_written_extent(&map))) {
fs/xfs/libxfs/xfs_rtbitmap.c
205
error = xfs_trans_read_buf(mp, args->tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_rtbitmap.c
206
XFS_FSB_TO_DADDR(mp, map.br_startblock),
fs/xfs/libxfs/xfs_rtbitmap.c
207
mp->m_bsize, 0, &bp,
fs/xfs/libxfs/xfs_rtbitmap.c
208
xfs_rtblock_ops(mp, type));
fs/xfs/libxfs/xfs_rtbitmap.c
214
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.c
236
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
238
if (XFS_IS_CORRUPT(mp, block >= mp->m_sb.sb_rbmblocks)) {
fs/xfs/libxfs/xfs_rtbitmap.c
251
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
253
if (XFS_IS_CORRUPT(mp, block >= mp->m_rsumblocks)) {
fs/xfs/libxfs/xfs_rtbitmap.c
270
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
286
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
294
word = xfs_rtx_to_rbmword(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
340
word = mp->m_blockwsize - 1;
fs/xfs/libxfs/xfs_rtbitmap.c
37
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rtbitmap.c
378
word = mp->m_blockwsize - 1;
fs/xfs/libxfs/xfs_rtbitmap.c
42
if (!xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
424
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
44
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
442
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
450
word = xfs_rtx_to_rbmword(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
46
if (!uuid_equal(&hdr->rt_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_rtbitmap.c
487
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.c
525
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.c
57
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rtbitmap.c
592
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
593
xfs_rtsumoff_t so = xfs_rtsumoffs(mp, log, bbno);
fs/xfs/libxfs/xfs_rtbitmap.c
599
error = xfs_rtsummary_read_buf(args, xfs_rtsumoffs_to_block(mp, so));
fs/xfs/libxfs/xfs_rtbitmap.c
603
infoword = xfs_rtsumoffs_to_infoword(mp, so);
fs/xfs/libxfs/xfs_rtbitmap.c
61
if (!xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
628
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
629
xfs_rtsumoff_t so = xfs_rtsumoffs(mp, log, bbno);
fs/xfs/libxfs/xfs_rtbitmap.c
632
error = xfs_rtsummary_read_buf(args, xfs_rtsumoffs_to_block(mp, so));
fs/xfs/libxfs/xfs_rtbitmap.c
634
*sum = xfs_suminfo_get(args, xfs_rtsumoffs_to_infoword(mp, so));
fs/xfs/libxfs/xfs_rtbitmap.c
64
if (!xfs_log_check_lsn(mp, be64_to_cpu(hdr->rt_lsn))) {
fs/xfs/libxfs/xfs_rtbitmap.c
665
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
679
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
690
firstword = word = xfs_rtx_to_rbmword(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
720
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.c
752
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.c
803
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
840
xfs_rtx_to_rbmblock(mp, preblock), -1);
fs/xfs/libxfs/xfs_rtbitmap.c
852
xfs_rtx_to_rbmblock(mp, end + 1), -1);
fs/xfs/libxfs/xfs_rtbitmap.c
863
xfs_rtx_to_rbmblock(mp, preblock), 1);
fs/xfs/libxfs/xfs_rtbitmap.c
87
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rtbitmap.c
879
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.c
893
block = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
904
word = xfs_rtx_to_rbmword(mp, start);
fs/xfs/libxfs/xfs_rtbitmap.c
92
if (!xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.c
941
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.c
980
if (++word == mp->m_blockwsize && i < len) {
fs/xfs/libxfs/xfs_rtbitmap.h
101
return div_u64(blen, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
107
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
110
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtbitmap.h
111
return blen & mp->m_rtxblkmask;
fs/xfs/libxfs/xfs_rtbitmap.h
113
return do_div(blen, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
119
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
122
return roundup_64(blen, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
128
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
13
struct xfs_mount *mp;
fs/xfs/libxfs/xfs_rtbitmap.h
132
rtbno &= mp->m_groups[XG_TYPE_RTG].blkmask;
fs/xfs/libxfs/xfs_rtbitmap.h
133
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtbitmap.h
134
return rtbno >> mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtbitmap.h
135
return div_u64(rtbno, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
141
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
144
return rgbno % mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
150
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
154
rtbno &= mp->m_groups[XG_TYPE_RTG].blkmask;
fs/xfs/libxfs/xfs_rtbitmap.h
155
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtbitmap.h
156
return rtbno & mp->m_rtxblkmask;
fs/xfs/libxfs/xfs_rtbitmap.h
157
return do_div(rtbno, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
163
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
166
return roundup_64(off, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
172
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
175
return rounddown_64(off, mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
181
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
184
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.h
185
return div_u64(rtx, mp->m_rtx_per_rbmblock);
fs/xfs/libxfs/xfs_rtbitmap.h
187
return rtx >> mp->m_blkbit_log;
fs/xfs/libxfs/xfs_rtbitmap.h
193
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
196
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.h
199
div_u64_rem(rtx >> XFS_NBWORDLOG, mp->m_blockwsize, &mod);
fs/xfs/libxfs/xfs_rtbitmap.h
203
return (rtx >> XFS_NBWORDLOG) & (mp->m_blockwsize - 1);
fs/xfs/libxfs/xfs_rtbitmap.h
209
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
212
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.h
213
return rbmoff * mp->m_rtx_per_rbmblock;
fs/xfs/libxfs/xfs_rtbitmap.h
215
return rbmoff << mp->m_blkbit_log;
fs/xfs/libxfs/xfs_rtbitmap.h
224
struct xfs_mount *mp = args->mp;
fs/xfs/libxfs/xfs_rtbitmap.h
228
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.h
244
if (xfs_has_rtgroups(args->mp))
fs/xfs/libxfs/xfs_rtbitmap.h
258
if (xfs_has_rtgroups(args->mp))
fs/xfs/libxfs/xfs_rtbitmap.h
270
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
274
return log2_len * mp->m_sb.sb_rbmblocks + rbmoff;
fs/xfs/libxfs/xfs_rtbitmap.h
28
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtbitmap.h
283
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
286
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.h
287
return rsumoff / mp->m_blockwsize;
fs/xfs/libxfs/xfs_rtbitmap.h
289
return XFS_B_TO_FSBT(mp, rsumoff * sizeof(xfs_suminfo_t));
fs/xfs/libxfs/xfs_rtbitmap.h
298
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
301
unsigned int mask = mp->m_blockmask >> XFS_SUMINFOLOG;
fs/xfs/libxfs/xfs_rtbitmap.h
303
if (xfs_has_rtgroups(mp))
fs/xfs/libxfs/xfs_rtbitmap.h
304
return rsumoff % mp->m_blockwsize;
fs/xfs/libxfs/xfs_rtbitmap.h
31
if (mp->m_rtxblklog >= 0)
fs/xfs/libxfs/xfs_rtbitmap.h
318
if (xfs_has_rtgroups(args->mp))
fs/xfs/libxfs/xfs_rtbitmap.h
32
return start + (rtx << mp->m_rtxblklog);
fs/xfs/libxfs/xfs_rtbitmap.h
33
return start + (rtx * mp->m_sb.sb_rextsize);
fs/xfs/libxfs/xfs_rtbitmap.h
334
if (xfs_has_rtgroups(args->mp))
fs/xfs/libxfs/xfs_rtbitmap.h
348
if (xfs_has_rtgroups(args->mp)) {
fs/xfs/libxfs/xfs_rtbitmap.h
359
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
362
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtbitmap.h
39
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
415
xfs_rtxnum_t xfs_rtbitmap_rtx_per_rbmblock(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtbitmap.h
416
xfs_filblks_t xfs_rtbitmap_blockcount(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtbitmap.h
417
xfs_filblks_t xfs_rtbitmap_blockcount_len(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
419
xfs_filblks_t xfs_rtsummary_blockcount(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
42
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtbitmap.h
43
return rgbno >> mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtbitmap.h
44
return rgbno / mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
446
xfs_rtbitmap_blockcount_len(struct xfs_mount *mp, xfs_rtbxlen_t rtextents)
fs/xfs/libxfs/xfs_rtbitmap.h
49
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
52
if (mp->m_rtxblklog >= 0)
fs/xfs/libxfs/xfs_rtbitmap.h
53
return rtbxlen << mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtbitmap.h
55
return rtbxlen * mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
60
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
63
if (mp->m_rtxblklog >= 0)
fs/xfs/libxfs/xfs_rtbitmap.h
64
return rtxlen << mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtbitmap.h
66
return rtxlen * mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
72
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
75
if (mp->m_rtxblklog >= 0)
fs/xfs/libxfs/xfs_rtbitmap.h
76
return len & mp->m_rtxblkmask;
fs/xfs/libxfs/xfs_rtbitmap.h
78
return len % mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
83
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
86
if (mp->m_rtxblklog >= 0)
fs/xfs/libxfs/xfs_rtbitmap.h
87
return len >> mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtbitmap.h
89
return len / mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtbitmap.h
95
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtbitmap.h
98
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtbitmap.h
99
return blen >> mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtgroup.c
105
xfs_rtgroup_calc_geometry(mp, rtg, rgno, rgcount, rextents);
fs/xfs/libxfs/xfs_rtgroup.c
107
error = xfs_group_insert(mp, rtg_group(rtg), rgno, XG_TYPE_RTG);
fs/xfs/libxfs/xfs_rtgroup.c
119
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
122
xfs_group_free(mp, rgno, XG_TYPE_RTG, NULL);
fs/xfs/libxfs/xfs_rtgroup.c
128
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
135
xfs_rtgroup_free(mp, rgno);
fs/xfs/libxfs/xfs_rtgroup.c
141
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
153
error = xfs_rtgroup_alloc(mp, index, end_rgno, rextents);
fs/xfs/libxfs/xfs_rtgroup.c
161
xfs_free_rtgroups(mp, first_rgno, index);
fs/xfs/libxfs/xfs_rtgroup.c
171
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
178
rtg = xfs_rtgroup_grab(mp, prev_rgcount - 1);
fs/xfs/libxfs/xfs_rtgroup.c
181
rtg->rtg_extents = __xfs_rtgroup_extents(mp, prev_rgcount - 1,
fs/xfs/libxfs/xfs_rtgroup.c
182
mp->m_sb.sb_rgcount, mp->m_sb.sb_rextents);
fs/xfs/libxfs/xfs_rtgroup.c
183
rtg_group(rtg)->xg_block_count = rtg->rtg_extents * mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtgroup.c
347
bool (*enabled)(const struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtgroup.c
42
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
446
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rtgroup.c
45
if (xfs_has_rtsb(mp) && rgno == 0)
fs/xfs/libxfs/xfs_rtgroup.c
454
if (!xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_rtgroup.c
459
ino = mp->m_sb.sb_rbmino;
fs/xfs/libxfs/xfs_rtgroup.c
46
return mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtgroup.c
462
ino = mp->m_sb.sb_rsumino;
fs/xfs/libxfs/xfs_rtgroup.c
474
if (!mp->m_rtdirip) {
fs/xfs/libxfs/xfs_rtgroup.c
475
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_rtgroup.c
482
error = xfs_metadir_load(tp, mp->m_rtdirip, path,
fs/xfs/libxfs/xfs_rtgroup.c
493
if (XFS_IS_CORRUPT(mp, !((1U << ip->i_df.if_format) & ops->fmt_mask))) {
fs/xfs/libxfs/xfs_rtgroup.c
499
if (XFS_IS_CORRUPT(mp, ip->i_projid != rtg_rgno(rtg))) {
fs/xfs/libxfs/xfs_rtgroup.c
528
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtgroup.c
530
.dp = mp->m_rtdirip,
fs/xfs/libxfs/xfs_rtgroup.c
538
if (!mp->m_rtdirip) {
fs/xfs/libxfs/xfs_rtgroup.c
539
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_rtgroup.c
54
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
586
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtgroup.c
588
if (!mp->m_metadirip) {
fs/xfs/libxfs/xfs_rtgroup.c
589
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_rtgroup.c
593
return xfs_metadir_mkdir(mp->m_metadirip, "rtgroups", &mp->m_rtdirip);
fs/xfs/libxfs/xfs_rtgroup.c
601
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_rtgroup.c
603
if (!mp->m_metadirip) {
fs/xfs/libxfs/xfs_rtgroup.c
604
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/libxfs/xfs_rtgroup.c
608
return xfs_metadir_load(tp, mp->m_metadirip, "rtgroups",
fs/xfs/libxfs/xfs_rtgroup.c
609
XFS_METAFILE_DIR, &mp->m_rtdirip);
fs/xfs/libxfs/xfs_rtgroup.c
61
return rextents - ((xfs_rtxnum_t)rgno * mp->m_sb.sb_rgextents);
fs/xfs/libxfs/xfs_rtgroup.c
63
ASSERT(xfs_has_rtgroups(mp));
fs/xfs/libxfs/xfs_rtgroup.c
637
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rtgroup.c
64
return mp->m_sb.sb_rgextents;
fs/xfs/libxfs/xfs_rtgroup.c
644
if (memcmp(&rsb->rsb_fname, &mp->m_sb.sb_fname, XFSLABEL_MAX))
fs/xfs/libxfs/xfs_rtgroup.c
646
if (!uuid_equal(&rsb->rsb_uuid, &mp->m_sb.sb_uuid))
fs/xfs/libxfs/xfs_rtgroup.c
648
if (!uuid_equal(&rsb->rsb_meta_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_rtgroup.c
69
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
72
return __xfs_rtgroup_extents(mp, rgno, mp->m_sb.sb_rgcount,
fs/xfs/libxfs/xfs_rtgroup.c
73
mp->m_sb.sb_rextents);
fs/xfs/libxfs/xfs_rtgroup.c
79
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.c
85
rtg->rtg_extents = __xfs_rtgroup_extents(mp, rgno, rgcount, rextents);
fs/xfs/libxfs/xfs_rtgroup.c
87
rtg->rtg_extents * mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtgroup.c
88
rtg_group(rtg)->xg_min_gbno = xfs_rtgroup_min_block(mp, rgno);
fs/xfs/libxfs/xfs_rtgroup.c
93
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
115
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
118
return to_rtg(xfs_group_get(mp, rgno, XG_TYPE_RTG));
fs/xfs/libxfs/xfs_rtgroup.h
138
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
141
return to_rtg(xfs_group_grab(mp, rgno, XG_TYPE_RTG));
fs/xfs/libxfs/xfs_rtgroup.h
153
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
158
return to_rtg(xfs_group_next_range(mp, rtg ? rtg_group(rtg) : NULL,
fs/xfs/libxfs/xfs_rtgroup.h
164
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
167
return xfs_rtgroup_next_range(mp, rtg, 0, mp->m_sb.sb_rgcount - 1);
fs/xfs/libxfs/xfs_rtgroup.h
206
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
209
return xfs_fsb_to_gno(mp, rtbno, XG_TYPE_RTG);
fs/xfs/libxfs/xfs_rtgroup.h
214
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
217
return xfs_fsb_to_gbno(mp, rtbno, XG_TYPE_RTG);
fs/xfs/libxfs/xfs_rtgroup.h
223
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
226
return (rtbno & mp->m_groups[XG_TYPE_RTG].blkmask) == 0;
fs/xfs/libxfs/xfs_rtgroup.h
235
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtgroup.h
237
if (likely(mp->m_rtxblklog >= 0))
fs/xfs/libxfs/xfs_rtgroup.h
238
return rtx << mp->m_rtxblklog;
fs/xfs/libxfs/xfs_rtgroup.h
239
return rtx * mp->m_sb.sb_rextsize;
fs/xfs/libxfs/xfs_rtgroup.h
244
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
247
struct xfs_groups *g = &mp->m_groups[XG_TYPE_RTG];
fs/xfs/libxfs/xfs_rtgroup.h
249
if (xfs_has_rtgroups(mp) && !g->has_daddr_gaps) {
fs/xfs/libxfs/xfs_rtgroup.h
250
xfs_rgnumber_t rgno = xfs_rtb_to_rgno(mp, rtbno);
fs/xfs/libxfs/xfs_rtgroup.h
255
return XFS_FSB_TO_BB(mp, g->start_fsb + rtbno);
fs/xfs/libxfs/xfs_rtgroup.h
260
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
263
struct xfs_groups *g = &mp->m_groups[XG_TYPE_RTG];
fs/xfs/libxfs/xfs_rtgroup.h
266
bno = XFS_BB_TO_FSBT(mp, daddr) - g->start_fsb;
fs/xfs/libxfs/xfs_rtgroup.h
267
if (xfs_has_rtgroups(mp) && !g->has_daddr_gaps) {
fs/xfs/libxfs/xfs_rtgroup.h
279
int xfs_rtgroup_alloc(struct xfs_mount *mp, xfs_rgnumber_t rgno,
fs/xfs/libxfs/xfs_rtgroup.h
281
void xfs_rtgroup_free(struct xfs_mount *mp, xfs_rgnumber_t rgno);
fs/xfs/libxfs/xfs_rtgroup.h
283
void xfs_free_rtgroups(struct xfs_mount *mp, xfs_rgnumber_t first_rgno,
fs/xfs/libxfs/xfs_rtgroup.h
285
int xfs_initialize_rtgroups(struct xfs_mount *mp, xfs_rgnumber_t first_rgno,
fs/xfs/libxfs/xfs_rtgroup.h
288
xfs_rtxnum_t xfs_rtgroup_extents(struct xfs_mount *mp, xfs_rgnumber_t rgno);
fs/xfs/libxfs/xfs_rtgroup.h
289
void xfs_rtgroup_calc_geometry(struct xfs_mount *mp, struct xfs_rtgroup *rtg,
fs/xfs/libxfs/xfs_rtgroup.h
293
int xfs_update_last_rtgroup_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
318
int xfs_rtginode_mkdir_parent(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtgroup.h
344
static inline void xfs_free_rtgroups(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
349
static inline int xfs_initialize_rtgroups(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
356
# define xfs_rtgroup_extents(mp, rgno) (0)
fs/xfs/libxfs/xfs_rtgroup.h
357
# define xfs_update_last_rtgroup_size(mp, rgno) (0)
fs/xfs/libxfs/xfs_rtgroup.h
368
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtgroup.h
371
return xfs_groups_to_rfsbs(mp, nr_groups, XG_TYPE_RTG);
fs/xfs/libxfs/xfs_rtgroup.h
380
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtgroup.h
382
struct xfs_groups *g = &mp->m_groups[XG_TYPE_RTG];
fs/xfs/libxfs/xfs_rtrefcount_btree.c
190
struct xfs_mount *mp = bp->b_target->bt_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
198
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_rtrefcount_btree.c
204
if (level > mp->m_rtrefc_maxlevels)
fs/xfs/libxfs/xfs_rtrefcount_btree.c
207
return xfs_btree_fsblock_verify(bp, mp->m_rtrefc_mxr[level != 0]);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
288
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
297
sptr = xfs_rtrefcount_broot_ptr_addr(mp, broot, 1, old_size);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
298
dptr = xfs_rtrefcount_broot_ptr_addr(mp, broot, 1, new_size);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
307
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
314
new_size = xfs_rtrefcount_broot_space_calc(mp, level, new_numrecs);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
335
old_numrecs = xfs_rtrefcountbt_maxrecs(mp, old_size, level);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
338
xfs_rtrefcountbt_move_ptrs(mp, broot, old_size,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
357
xfs_rtrefcountbt_move_ptrs(mp, ifp->if_broot, old_size,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
406
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
411
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_rtrefcountbt_ops,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
412
mp->m_rtrefc_maxlevels, xfs_rtrefcountbt_cur_cache);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
472
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
519
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtrefcount_btree.c
523
if (!xfs_has_rtreflink(mp)) {
fs/xfs/libxfs/xfs_rtrefcount_btree.c
524
mp->m_rtrefc_maxlevels = 0;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
534
d_maxlevels = xfs_btree_space_to_height(mp->m_rtrefc_mnr,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
535
mp->m_sb.sb_dblocks);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
536
r_maxlevels = xfs_btree_compute_maxlevels(mp->m_rtrefc_mnr,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
537
mp->m_sb.sb_rgextents);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
540
mp->m_rtrefc_maxlevels = min(d_maxlevels, r_maxlevels) + 1;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
546
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
549
return xfs_btree_calc_size(mp->m_rtrefc_mnr, len);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
557
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
561
if (mp->m_rtrefc_mxr[0] == 0)
fs/xfs/libxfs/xfs_rtrefcount_btree.c
564
return xfs_rtrefcountbt_calc_size(mp, rtblocks);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
573
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtrefcount_btree.c
575
if (!xfs_has_rtreflink(mp))
fs/xfs/libxfs/xfs_rtrefcount_btree.c
578
return xfs_rtrefcountbt_max_size(mp, mp->m_sb.sb_rgextents);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
591
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
602
rblocklen = xfs_rtrefcount_broot_space(mp, dblock);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
604
xfs_btree_init_block(mp, rblock, &xfs_rtrefcountbt_ops, 0, 0,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
615
tpp = xfs_rtrefcount_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
633
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
650
dsize = XFS_DFORK_SIZE(dip, mp, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
654
if (level > mp->m_rtrefc_maxlevels ||
fs/xfs/libxfs/xfs_rtrefcount_btree.c
661
xfs_rtrefcount_broot_space_calc(mp, level, numrecs));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
672
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.c
688
ASSERT(uuid_equal(&rblock->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
700
fpp = xfs_rtrefcount_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_rtrefcount_btree.c
742
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrefcount_btree.c
751
xfs_rtrefcount_broot_space_calc(mp, 0, 0));
fs/xfs/libxfs/xfs_rtrefcount_btree.c
753
xfs_btree_init_block(mp, broot, &xfs_rtrefcountbt_ops, 0, 0,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
117
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
123
xfs_rtrefcountbt_maxrecs(mp, block_size, false));
fs/xfs/libxfs/xfs_rtrefcount_btree.h
132
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
149
xfs_rtrefcount_broot_space(struct xfs_mount *mp, struct xfs_rtrefcount_root *bb)
fs/xfs/libxfs/xfs_rtrefcount_btree.h
151
return xfs_rtrefcount_broot_space_calc(mp, be16_to_cpu(bb->bb_level),
fs/xfs/libxfs/xfs_rtrefcount_btree.h
181
void xfs_rtrefcountbt_to_disk(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
20
struct xfs_btree_cur *xfs_rtrefcountbt_stage_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
25
unsigned int xfs_rtrefcountbt_maxrecs(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrefcount_btree.h
27
void xfs_rtrefcountbt_compute_maxlevels(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
71
xfs_filblks_t xfs_rtrefcountbt_calc_reserves(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtrefcount_btree.h
72
unsigned long long xfs_rtrefcountbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
1001
.rm_blockcount = mp->m_sb.sb_rextsize,
fs/xfs/libxfs/xfs_rtrmap_btree.c
1007
ASSERT(xfs_has_rtsb(mp));
fs/xfs/libxfs/xfs_rtrmap_btree.c
245
struct xfs_mount *mp = bp->b_target->bt_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
253
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_rtrmap_btree.c
259
if (level > mp->m_rtrmap_maxlevels)
fs/xfs/libxfs/xfs_rtrmap_btree.c
262
return xfs_btree_fsblock_verify(bp, mp->m_rtrmap_mxr[level != 0]);
fs/xfs/libxfs/xfs_rtrmap_btree.c
389
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
398
sptr = xfs_rtrmap_broot_ptr_addr(mp, broot, 1, old_size);
fs/xfs/libxfs/xfs_rtrmap_btree.c
399
dptr = xfs_rtrmap_broot_ptr_addr(mp, broot, 1, new_size);
fs/xfs/libxfs/xfs_rtrmap_btree.c
408
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/libxfs/xfs_rtrmap_btree.c
415
new_size = xfs_rtrmap_broot_space_calc(mp, level, new_numrecs);
fs/xfs/libxfs/xfs_rtrmap_btree.c
436
old_numrecs = xfs_rtrmapbt_maxrecs(mp, old_size, level == 0);
fs/xfs/libxfs/xfs_rtrmap_btree.c
439
xfs_rtrmapbt_move_ptrs(mp, broot, old_size, new_size,
fs/xfs/libxfs/xfs_rtrmap_btree.c
458
xfs_rtrmapbt_move_ptrs(mp, ifp->if_broot, old_size, new_size,
fs/xfs/libxfs/xfs_rtrmap_btree.c
509
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtrmap_btree.c
514
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_rtrmapbt_ops,
fs/xfs/libxfs/xfs_rtrmap_btree.c
515
mp->m_rtrmap_maxlevels, xfs_rtrmapbt_cur_cache);
fs/xfs/libxfs/xfs_rtrmap_btree.c
535
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
549
if (xfs_has_rmapbt(mp)) {
fs/xfs/libxfs/xfs_rtrmap_btree.c
550
if (level >= mp->m_rtrmap_maxlevels)
fs/xfs/libxfs/xfs_rtrmap_btree.c
557
maxrecs = xfs_rtrmapbt_maxrecs(mp, XFBNO_BLOCKSIZE, level == 0);
fs/xfs/libxfs/xfs_rtrmap_btree.c
617
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/libxfs/xfs_rtrmap_btree.c
620
cur = xfs_btree_alloc_cursor(mp, tp, &xfs_rtrmapbt_mem_ops,
fs/xfs/libxfs/xfs_rtrmap_btree.c
621
mp->m_rtrmap_maxlevels, xfs_rtrmapbt_cur_cache);
fs/xfs/libxfs/xfs_rtrmap_btree.c
631
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
637
return xfbtree_init(mp, xfbt, btp, &xfs_rtrmapbt_mem_ops);
fs/xfs/libxfs/xfs_rtrmap_btree.c
688
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
747
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtrmap_btree.c
751
if (!xfs_has_rtrmapbt(mp)) {
fs/xfs/libxfs/xfs_rtrmap_btree.c
752
mp->m_rtrmap_maxlevels = 0;
fs/xfs/libxfs/xfs_rtrmap_btree.c
768
d_maxlevels = xfs_btree_space_to_height(mp->m_rtrmap_mnr,
fs/xfs/libxfs/xfs_rtrmap_btree.c
769
mp->m_sb.sb_dblocks);
fs/xfs/libxfs/xfs_rtrmap_btree.c
770
if (xfs_has_rtreflink(mp)) {
fs/xfs/libxfs/xfs_rtrmap_btree.c
771
mp->m_rtrmap_maxlevels = d_maxlevels + 1;
fs/xfs/libxfs/xfs_rtrmap_btree.c
775
r_maxlevels = xfs_btree_compute_maxlevels(mp->m_rtrmap_mnr,
fs/xfs/libxfs/xfs_rtrmap_btree.c
776
mp->m_groups[XG_TYPE_RTG].blocks);
fs/xfs/libxfs/xfs_rtrmap_btree.c
779
mp->m_rtrmap_maxlevels = min(d_maxlevels, r_maxlevels) + 1;
fs/xfs/libxfs/xfs_rtrmap_btree.c
785
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
788
return xfs_btree_calc_size(mp->m_rtrmap_mnr, len);
fs/xfs/libxfs/xfs_rtrmap_btree.c
796
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
800
if (mp->m_rtrmap_mxr[0] == 0)
fs/xfs/libxfs/xfs_rtrmap_btree.c
803
return xfs_rtrmapbt_calc_size(mp, rtblocks);
fs/xfs/libxfs/xfs_rtrmap_btree.c
811
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_rtrmap_btree.c
813
uint32_t blocks = mp->m_groups[XG_TYPE_RTG].blocks;
fs/xfs/libxfs/xfs_rtrmap_btree.c
815
if (!xfs_has_rtrmapbt(mp))
fs/xfs/libxfs/xfs_rtrmap_btree.c
820
xfs_rtrmapbt_max_size(mp, blocks));
fs/xfs/libxfs/xfs_rtrmap_btree.c
831
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
838
unsigned int rblocklen = xfs_rtrmap_broot_space(mp, dblock);
fs/xfs/libxfs/xfs_rtrmap_btree.c
842
xfs_btree_init_block(mp, rblock, &xfs_rtrmapbt_ops, 0, 0, ip->i_ino);
fs/xfs/libxfs/xfs_rtrmap_btree.c
853
tpp = xfs_rtrmap_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_rtrmap_btree.c
869
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
885
dsize = XFS_DFORK_SIZE(dip, mp, XFS_DATA_FORK);
fs/xfs/libxfs/xfs_rtrmap_btree.c
889
if (level > mp->m_rtrmap_maxlevels ||
fs/xfs/libxfs/xfs_rtrmap_btree.c
896
xfs_rtrmap_broot_space_calc(mp, level, numrecs));
fs/xfs/libxfs/xfs_rtrmap_btree.c
905
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.c
921
ASSERT(uuid_equal(&rblock->bb_u.l.bb_uuid, &mp->m_sb.sb_meta_uuid));
fs/xfs/libxfs/xfs_rtrmap_btree.c
934
fpp = xfs_rtrmap_broot_ptr_addr(mp, rblock, 1, rblocklen);
fs/xfs/libxfs/xfs_rtrmap_btree.c
973
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_rtrmap_btree.c
981
broot = xfs_broot_realloc(ifp, xfs_rtrmap_broot_space_calc(mp, 0, 0));
fs/xfs/libxfs/xfs_rtrmap_btree.c
983
xfs_btree_init_block(mp, broot, &xfs_rtrmapbt_ops, 0, 0,
fs/xfs/libxfs/xfs_rtrmap_btree.c
996
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.h
128
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.h
134
xfs_rtrmapbt_maxrecs(mp, block_size, false));
fs/xfs/libxfs/xfs_rtrmap_btree.h
143
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.h
160
xfs_rtrmap_broot_space(struct xfs_mount *mp, struct xfs_rtrmap_root *bb)
fs/xfs/libxfs/xfs_rtrmap_btree.h
162
return xfs_rtrmap_broot_space_calc(mp, be16_to_cpu(bb->bb_level),
fs/xfs/libxfs/xfs_rtrmap_btree.h
192
void xfs_rtrmapbt_to_disk(struct xfs_mount *mp, struct xfs_btree_block *rblock,
fs/xfs/libxfs/xfs_rtrmap_btree.h
199
int xfs_rtrmapbt_init_rtsb(struct xfs_mount *mp, struct xfs_rtgroup *rtg,
fs/xfs/libxfs/xfs_rtrmap_btree.h
202
unsigned long long xfs_rtrmapbt_calc_size(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.h
207
int xfs_rtrmapbt_mem_init(struct xfs_mount *mp, struct xfbtree *xfbtree,
fs/xfs/libxfs/xfs_rtrmap_btree.h
21
struct xfs_btree_cur *xfs_rtrmapbt_stage_cursor(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_rtrmap_btree.h
26
unsigned int xfs_rtrmapbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/libxfs/xfs_rtrmap_btree.h
28
void xfs_rtrmapbt_compute_maxlevels(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_rtrmap_btree.h
84
xfs_filblks_t xfs_rtrmapbt_calc_reserves(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_sb.c
1093
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_sb.c
1109
xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_sb.c
1121
error = xfs_validate_sb_common(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1124
error = xfs_validate_sb_read(mp, &sb);
fs/xfs/libxfs/xfs_sb.c
1159
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_sb.c
1169
error = xfs_validate_sb_common(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1172
error = xfs_validate_sb_write(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1206
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1209
struct xfs_groups *rgs = &mp->m_groups[XG_TYPE_RTG];
fs/xfs/libxfs/xfs_sb.c
1211
mp->m_rtxblklog = log2_if_power2(sbp->sb_rextsize);
fs/xfs/libxfs/xfs_sb.c
1212
mp->m_rtxblkmask = mask64_if_power2(sbp->sb_rextsize);
fs/xfs/libxfs/xfs_sb.c
1217
rgs->blklog = mp->m_sb.sb_rgblklog;
fs/xfs/libxfs/xfs_sb.c
1218
rgs->blkmask = xfs_mask32lo(mp->m_sb.sb_rgblklog);
fs/xfs/libxfs/xfs_sb.c
1219
rgs->start_fsb = mp->m_sb.sb_rtstart;
fs/xfs/libxfs/xfs_sb.c
1233
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1243
xfs_sb_mount_rextsize(mp, sbp);
fs/xfs/libxfs/xfs_sb.c
1257
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1260
struct xfs_groups *ags = &mp->m_groups[XG_TYPE_AG];
fs/xfs/libxfs/xfs_sb.c
1262
mp->m_agfrotor = 0;
fs/xfs/libxfs/xfs_sb.c
1263
atomic_set(&mp->m_agirotor, 0);
fs/xfs/libxfs/xfs_sb.c
1264
mp->m_maxagi = mp->m_sb.sb_agcount;
fs/xfs/libxfs/xfs_sb.c
1265
mp->m_blkbit_log = sbp->sb_blocklog + XFS_NBBYLOG;
fs/xfs/libxfs/xfs_sb.c
1266
mp->m_blkbb_log = sbp->sb_blocklog - BBSHIFT;
fs/xfs/libxfs/xfs_sb.c
1267
mp->m_sectbb_log = sbp->sb_sectlog - BBSHIFT;
fs/xfs/libxfs/xfs_sb.c
1268
mp->m_agno_log = xfs_highbit32(sbp->sb_agcount - 1) + 1;
fs/xfs/libxfs/xfs_sb.c
1269
mp->m_blockmask = sbp->sb_blocksize - 1;
fs/xfs/libxfs/xfs_sb.c
1270
mp->m_blockwsize = xfs_rtbmblock_size(sbp) >> XFS_WORDLOG;
fs/xfs/libxfs/xfs_sb.c
1271
mp->m_rtx_per_rbmblock = mp->m_blockwsize << XFS_NBWORDLOG;
fs/xfs/libxfs/xfs_sb.c
1273
ags->blocks = mp->m_sb.sb_agblocks;
fs/xfs/libxfs/xfs_sb.c
1274
ags->blklog = mp->m_sb.sb_agblklog;
fs/xfs/libxfs/xfs_sb.c
1275
ags->blkmask = xfs_mask32lo(mp->m_sb.sb_agblklog);
fs/xfs/libxfs/xfs_sb.c
1277
xfs_sb_mount_rextsize(mp, sbp);
fs/xfs/libxfs/xfs_sb.c
1279
mp->m_alloc_mxr[0] = xfs_allocbt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_sb.c
1280
mp->m_alloc_mxr[1] = xfs_allocbt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_sb.c
1281
mp->m_alloc_mnr[0] = mp->m_alloc_mxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1282
mp->m_alloc_mnr[1] = mp->m_alloc_mxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1284
mp->m_bmap_dmxr[0] = xfs_bmbt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_sb.c
1285
mp->m_bmap_dmxr[1] = xfs_bmbt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_sb.c
1286
mp->m_bmap_dmnr[0] = mp->m_bmap_dmxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1287
mp->m_bmap_dmnr[1] = mp->m_bmap_dmxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1289
mp->m_rmap_mxr[0] = xfs_rmapbt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_sb.c
1290
mp->m_rmap_mxr[1] = xfs_rmapbt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_sb.c
1291
mp->m_rmap_mnr[0] = mp->m_rmap_mxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1292
mp->m_rmap_mnr[1] = mp->m_rmap_mxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1294
mp->m_rtrmap_mxr[0] = xfs_rtrmapbt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_sb.c
1295
mp->m_rtrmap_mxr[1] = xfs_rtrmapbt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_sb.c
1296
mp->m_rtrmap_mnr[0] = mp->m_rtrmap_mxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1297
mp->m_rtrmap_mnr[1] = mp->m_rtrmap_mxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1299
mp->m_refc_mxr[0] = xfs_refcountbt_maxrecs(mp, sbp->sb_blocksize, true);
fs/xfs/libxfs/xfs_sb.c
1300
mp->m_refc_mxr[1] = xfs_refcountbt_maxrecs(mp, sbp->sb_blocksize, false);
fs/xfs/libxfs/xfs_sb.c
1301
mp->m_refc_mnr[0] = mp->m_refc_mxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1302
mp->m_refc_mnr[1] = mp->m_refc_mxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1304
mp->m_rtrefc_mxr[0] = xfs_rtrefcountbt_maxrecs(mp, sbp->sb_blocksize,
fs/xfs/libxfs/xfs_sb.c
1306
mp->m_rtrefc_mxr[1] = xfs_rtrefcountbt_maxrecs(mp, sbp->sb_blocksize,
fs/xfs/libxfs/xfs_sb.c
1308
mp->m_rtrefc_mnr[0] = mp->m_rtrefc_mxr[0] / 2;
fs/xfs/libxfs/xfs_sb.c
1309
mp->m_rtrefc_mnr[1] = mp->m_rtrefc_mxr[1] / 2;
fs/xfs/libxfs/xfs_sb.c
1311
mp->m_bsize = XFS_FSB_TO_BB(mp, 1);
fs/xfs/libxfs/xfs_sb.c
1312
mp->m_alloc_set_aside = xfs_alloc_set_aside(mp);
fs/xfs/libxfs/xfs_sb.c
1313
mp->m_ag_max_usable = xfs_alloc_ag_max_usable(mp);
fs/xfs/libxfs/xfs_sb.c
1326
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_sb.c
1337
if (xfs_has_lazysbcount(mp)) {
fs/xfs/libxfs/xfs_sb.c
1338
mp->m_sb.sb_icount = percpu_counter_sum_positive(&mp->m_icount);
fs/xfs/libxfs/xfs_sb.c
1339
mp->m_sb.sb_ifree = min_t(uint64_t,
fs/xfs/libxfs/xfs_sb.c
1340
percpu_counter_sum_positive(&mp->m_ifree),
fs/xfs/libxfs/xfs_sb.c
1341
mp->m_sb.sb_icount);
fs/xfs/libxfs/xfs_sb.c
1342
mp->m_sb.sb_fdblocks = xfs_sum_freecounter(mp, XC_FREE_BLOCKS);
fs/xfs/libxfs/xfs_sb.c
1354
if (xfs_has_rtgroups(mp) && !xfs_has_zoned(mp)) {
fs/xfs/libxfs/xfs_sb.c
1355
mp->m_sb.sb_frextents =
fs/xfs/libxfs/xfs_sb.c
1356
xfs_sum_freecounter(mp, XC_FREE_RTEXTENTS);
fs/xfs/libxfs/xfs_sb.c
1359
xfs_sb_to_disk(bp->b_addr, &mp->m_sb);
fs/xfs/libxfs/xfs_sb.c
1377
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1383
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_sb, 0, 0,
fs/xfs/libxfs/xfs_sb.c
1406
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_sb.c
1414
while ((pag = xfs_perag_next_from(mp, pag, 1))) {
fs/xfs/libxfs/xfs_sb.c
1417
error = xfs_buf_get(mp->m_ddev_targp,
fs/xfs/libxfs/xfs_sb.c
1418
XFS_AG_DADDR(mp, pag_agno(pag), XFS_SB_DADDR),
fs/xfs/libxfs/xfs_sb.c
1419
XFS_FSS_TO_BB(mp, 1), &bp);
fs/xfs/libxfs/xfs_sb.c
1428
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
1439
xfs_sb_to_disk(bp->b_addr, &mp->m_sb);
fs/xfs/libxfs/xfs_sb.c
1449
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
1459
xfs_warn(mp, "error %d writing secondary superblocks", error);
fs/xfs/libxfs/xfs_sb.c
1469
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1477
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_sb, 0, 0, 0, &tp);
fs/xfs/libxfs/xfs_sb.c
1508
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1512
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/libxfs/xfs_sb.c
1545
if (xfs_has_attr(mp))
fs/xfs/libxfs/xfs_sb.c
1547
if (xfs_has_quota(mp))
fs/xfs/libxfs/xfs_sb.c
1549
if (xfs_has_align(mp))
fs/xfs/libxfs/xfs_sb.c
1551
if (xfs_has_dalign(mp))
fs/xfs/libxfs/xfs_sb.c
1553
if (xfs_has_asciici(mp))
fs/xfs/libxfs/xfs_sb.c
1555
if (xfs_has_lazysbcount(mp))
fs/xfs/libxfs/xfs_sb.c
1557
if (xfs_has_projid32(mp))
fs/xfs/libxfs/xfs_sb.c
1559
if (xfs_has_crc(mp))
fs/xfs/libxfs/xfs_sb.c
1561
if (xfs_has_ftype(mp))
fs/xfs/libxfs/xfs_sb.c
1563
if (xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_sb.c
1565
if (xfs_has_sparseinodes(mp))
fs/xfs/libxfs/xfs_sb.c
1567
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_sb.c
1569
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_sb.c
1571
if (xfs_has_bigtime(mp))
fs/xfs/libxfs/xfs_sb.c
1573
if (xfs_has_inobtcounts(mp))
fs/xfs/libxfs/xfs_sb.c
1575
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_sb.c
1577
if (xfs_has_sector(mp)) {
fs/xfs/libxfs/xfs_sb.c
1583
if (xfs_has_large_extent_counts(mp))
fs/xfs/libxfs/xfs_sb.c
1585
if (xfs_has_exchange_range(mp))
fs/xfs/libxfs/xfs_sb.c
1587
if (xfs_has_metadir(mp))
fs/xfs/libxfs/xfs_sb.c
1589
if (xfs_has_zoned(mp))
fs/xfs/libxfs/xfs_sb.c
1597
if (xfs_has_logv2(mp))
fs/xfs/libxfs/xfs_sb.c
1607
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_sb.c
1611
if (xfs_has_zoned(mp)) {
fs/xfs/libxfs/xfs_sb.c
1620
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1629
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_sb.c
1630
XFS_AG_DADDR(mp, agno, XFS_SB_BLOCK(mp)),
fs/xfs/libxfs/xfs_sb.c
1631
XFS_FSS_TO_BB(mp, 1), 0, &bp, &xfs_sb_buf_ops);
fs/xfs/libxfs/xfs_sb.c
1633
xfs_agno_mark_sick(mp, agno, XFS_SICK_AG_SB);
fs/xfs/libxfs/xfs_sb.c
1644
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1653
error = xfs_trans_get_buf(tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_sb.c
1654
XFS_AG_DADDR(mp, agno, XFS_SB_BLOCK(mp)),
fs/xfs/libxfs/xfs_sb.c
1655
XFS_FSS_TO_BB(mp, 1), 0, &bp);
fs/xfs/libxfs/xfs_sb.c
1672
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
1681
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1688
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1695
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1703
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1710
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1717
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1734
if (!mp->m_dalign)
fs/xfs/libxfs/xfs_sb.c
1737
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
1739
BBTOB(mp->m_dalign), BBTOB(mp->m_swidth));
fs/xfs/libxfs/xfs_sb.c
195
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
206
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
209
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
214
xfs_alert(mp,
fs/xfs/libxfs/xfs_sb.c
218
if (!xfs_is_readonly(mp)) {
fs/xfs/libxfs/xfs_sb.c
219
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
221
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
228
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
232
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
325
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
341
!xfs_verify_icount(mp, sbp->sb_icount) ||
fs/xfs/libxfs/xfs_sb.c
343
xfs_warn(mp, "SB summary counter sanity check failed");
fs/xfs/libxfs/xfs_sb.c
355
if (!xfs_is_readonly(mp) &&
fs/xfs/libxfs/xfs_sb.c
357
xfs_alert(mp,
fs/xfs/libxfs/xfs_sb.c
364
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
372
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
384
if (!xfs_log_check_lsn(mp, sbp->sb_lsn))
fs/xfs/libxfs/xfs_sb.c
402
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
409
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
415
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
423
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
430
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
438
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
446
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
453
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
464
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
468
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
474
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
481
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
493
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.c
504
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
511
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
521
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
529
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
545
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
554
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
561
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
569
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
574
error = xfs_validate_sb_rtgroups(mp, sbp);
fs/xfs/libxfs/xfs_sb.c
579
error = xfs_validate_sb_zoned(mp, sbp);
fs/xfs/libxfs/xfs_sb.c
585
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
591
sbp->sb_logstart == 0 && mp->m_logdev_targp == mp->m_ddev_targp)) {
fs/xfs/libxfs/xfs_sb.c
592
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
599
sbp->sb_logstart != 0 && mp->m_logdev_targp != mp->m_ddev_targp)) {
fs/xfs/libxfs/xfs_sb.c
600
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
636
XFS_FSB_TO_B(mp, sbp->sb_agblocks) < XFS_MIN_AG_BYTES ||
fs/xfs/libxfs/xfs_sb.c
637
XFS_FSB_TO_B(mp, sbp->sb_agblocks) > XFS_MAX_AG_BYTES ||
fs/xfs/libxfs/xfs_sb.c
648
xfs_notice(mp, "SB sanity check failed");
fs/xfs/libxfs/xfs_sb.c
659
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
665
if (XFS_FSB_TO_B(mp, sbp->sb_logblocks) > XFS_MAX_LOG_BYTES) {
fs/xfs/libxfs/xfs_sb.c
666
xfs_warn(mp,
fs/xfs/libxfs/xfs_sb.c
668
XFS_FSB_TO_B(mp, sbp->sb_logblocks),
fs/xfs/libxfs/xfs_sb.c
680
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
686
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
694
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
700
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
708
xfs_notice(mp,
fs/xfs/libxfs/xfs_sb.c
720
xfs_notice(mp, "SB stripe alignment sanity check failed");
fs/xfs/libxfs/xfs_sb.c
724
if (!xfs_validate_stripe_geometry(mp, XFS_FSB_TO_B(mp, sbp->sb_unit),
fs/xfs/libxfs/xfs_sb.c
725
XFS_FSB_TO_B(mp, sbp->sb_width), 0,
fs/xfs/libxfs/xfs_sb.c
739
xfs_warn(mp, "inode size of %d bytes not supported",
fs/xfs/libxfs/xfs_sb.h
17
extern int xfs_sync_sb(struct xfs_mount *mp, bool wait);
fs/xfs/libxfs/xfs_sb.h
18
extern int xfs_sync_sb_buf(struct xfs_mount *mp, bool update_rtsb);
fs/xfs/libxfs/xfs_sb.h
19
extern void xfs_sb_mount_common(struct xfs_mount *mp, struct xfs_sb *sbp);
fs/xfs/libxfs/xfs_sb.h
20
void xfs_sb_mount_rextsize(struct xfs_mount *mp, struct xfs_sb *sbp);
fs/xfs/libxfs/xfs_sb.h
21
void xfs_mount_sb_set_rextsize(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.h
29
extern int xfs_update_secondary_sbs(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_sb.h
32
extern void xfs_fs_geometry(struct xfs_mount *mp, struct xfs_fsop_geom *geo,
fs/xfs/libxfs/xfs_sb.h
34
extern int xfs_sb_read_secondary(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.h
37
extern int xfs_sb_get_secondary(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_sb.h
41
bool xfs_validate_stripe_geometry(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_shared.h
125
int xfs_log_calc_unit_res(struct xfs_mount *mp, int unit_bytes);
fs/xfs/libxfs/xfs_shared.h
129
void xfs_log_get_max_trans_res(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_symlink_remote.c
101
if (!uuid_equal(&dsl->sl_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/libxfs/xfs_symlink_remote.c
110
if (!xfs_log_check_lsn(mp, be64_to_cpu(dsl->sl_lsn)))
fs/xfs/libxfs/xfs_symlink_remote.c
120
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
124
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_symlink_remote.c
140
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
145
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_symlink_remote.c
177
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
182
if (!xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_symlink_remote.c
199
buf += xfs_symlink_hdr_set(mp, ip->i_ino, 0, ifp->if_bytes, bp);
fs/xfs/libxfs/xfs_symlink_remote.c
243
struct xfs_mount *mp = ip->i_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
258
fsblocks = xfs_symlink_blocks(mp, pathlen);
fs/xfs/libxfs/xfs_symlink_remote.c
265
d = XFS_FSB_TO_DADDR(mp, mval[n].br_startblock);
fs/xfs/libxfs/xfs_symlink_remote.c
266
byte_cnt = XFS_FSB_TO_B(mp, mval[n].br_blockcount);
fs/xfs/libxfs/xfs_symlink_remote.c
268
error = xfs_buf_read(mp->m_ddev_targp, d, BTOBB(byte_cnt), 0,
fs/xfs/libxfs/xfs_symlink_remote.c
274
byte_cnt = XFS_SYMLINK_BUF_SPACE(mp, byte_cnt);
fs/xfs/libxfs/xfs_symlink_remote.c
279
if (xfs_has_crc(mp)) {
fs/xfs/libxfs/xfs_symlink_remote.c
284
xfs_alert(mp,
fs/xfs/libxfs/xfs_symlink_remote.c
30
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_symlink_remote.c
323
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_symlink_remote.c
33
int buflen = XFS_SYMLINK_BUF_SPACE(mp, mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_symlink_remote.c
359
d = XFS_FSB_TO_DADDR(mp, mval[n].br_startblock);
fs/xfs/libxfs/xfs_symlink_remote.c
360
byte_cnt = XFS_FSB_TO_B(mp, mval[n].br_blockcount);
fs/xfs/libxfs/xfs_symlink_remote.c
361
error = xfs_trans_get_buf(tp, mp->m_ddev_targp, d,
fs/xfs/libxfs/xfs_symlink_remote.c
367
byte_cnt = XFS_SYMLINK_BUF_SPACE(mp, byte_cnt);
fs/xfs/libxfs/xfs_symlink_remote.c
371
buf += xfs_symlink_hdr_set(mp, owner, offset, byte_cnt, bp);
fs/xfs/libxfs/xfs_symlink_remote.c
394
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/libxfs/xfs_symlink_remote.c
40
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_symlink_remote.c
410
error = xfs_trans_get_buf(tp, mp->m_ddev_targp,
fs/xfs/libxfs/xfs_symlink_remote.c
411
XFS_FSB_TO_DADDR(mp, mval[i].br_startblock),
fs/xfs/libxfs/xfs_symlink_remote.c
412
XFS_FSB_TO_BB(mp, mval[i].br_blockcount), 0,
fs/xfs/libxfs/xfs_symlink_remote.c
48
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_symlink_remote.c
55
uuid_copy(&dsl->sl_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/libxfs/xfs_symlink_remote.c
92
struct xfs_mount *mp = bp->b_mount;
fs/xfs/libxfs/xfs_symlink_remote.c
96
if (!xfs_has_crc(mp))
fs/xfs/libxfs/xfs_symlink_remote.h
13
int xfs_symlink_blocks(struct xfs_mount *mp, int pathlen);
fs/xfs/libxfs/xfs_symlink_remote.h
14
int xfs_symlink_hdr_set(struct xfs_mount *mp, xfs_ino_t ino, uint32_t offset,
fs/xfs/libxfs/xfs_trans_resv.c
1001
return xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1002
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
fs/xfs/libxfs/xfs_trans_resv.c
1003
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
1018
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1020
return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1021
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK),
fs/xfs/libxfs/xfs_trans_resv.c
1022
XFS_FSB_TO_B(mp, 1)) +
fs/xfs/libxfs/xfs_trans_resv.c
1023
xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
1024
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
fs/xfs/libxfs/xfs_trans_resv.c
1025
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
103
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1035
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1037
return xfs_calc_buf_res(1, mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
1052
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1054
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1055
xfs_calc_inode_res(mp, 2) +
fs/xfs/libxfs/xfs_trans_resv.c
1056
xfs_calc_buf_res(1, mp->m_sb.sb_blocksize) +
fs/xfs/libxfs/xfs_trans_resv.c
1057
xfs_calc_buf_res(1, XFS_FSB_TO_B(mp, mp->m_rsumblocks));
fs/xfs/libxfs/xfs_trans_resv.c
106
return num_ops * (2 * mp->m_rtrefc_maxlevels - 1);
fs/xfs/libxfs/xfs_trans_resv.c
1066
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1068
return xfs_calc_inode_res(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
1077
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1079
return xfs_calc_inode_res(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
1092
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1095
xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
1096
xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1097
xfs_calc_buf_res(1, mp->m_dir_geo->blksize) +
fs/xfs/libxfs/xfs_trans_resv.c
1098
xfs_calc_buf_res(XFS_DAENTER_BMAP1B(mp, XFS_DATA_FORK) + 1,
fs/xfs/libxfs/xfs_trans_resv.c
1099
XFS_FSB_TO_B(mp, 1)) +
fs/xfs/libxfs/xfs_trans_resv.c
1100
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
fs/xfs/libxfs/xfs_trans_resv.c
1101
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
1117
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1119
return max((xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
1120
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK),
fs/xfs/libxfs/xfs_trans_resv.c
1121
XFS_FSB_TO_B(mp, 1))),
fs/xfs/libxfs/xfs_trans_resv.c
1122
(xfs_calc_buf_res(9, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1123
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 4),
fs/xfs/libxfs/xfs_trans_resv.c
1124
XFS_FSB_TO_B(mp, 1))));
fs/xfs/libxfs/xfs_trans_resv.c
1140
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1143
xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
1144
xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1145
xfs_calc_buf_res(XFS_DA_NODE_MAXDEPTH, XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
1160
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1162
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1163
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK),
fs/xfs/libxfs/xfs_trans_resv.c
1164
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
1180
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1183
max((xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
1185
XFS_FSB_TO_B(mp, 1)) +
fs/xfs/libxfs/xfs_trans_resv.c
1186
(uint)XFS_FSB_TO_B(mp,
fs/xfs/libxfs/xfs_trans_resv.c
1187
XFS_BM_MAXLEVELS(mp, XFS_ATTR_FORK)) +
fs/xfs/libxfs/xfs_trans_resv.c
1188
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), 0)),
fs/xfs/libxfs/xfs_trans_resv.c
1189
(xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1190
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2),
fs/xfs/libxfs/xfs_trans_resv.c
1191
XFS_FSB_TO_B(mp, 1))));
fs/xfs/libxfs/xfs_trans_resv.c
1199
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1201
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
fs/xfs/libxfs/xfs_trans_resv.c
1221
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1224
return xfs_calc_write_reservation(mp, for_minlogsize) +
fs/xfs/libxfs/xfs_trans_resv.c
1226
XFS_FSB_TO_B(mp, XFS_DQUOT_CLUSTER_SIZE_FSB) - 1);
fs/xfs/libxfs/xfs_trans_resv.c
1231
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1233
return xfs_calc_qm_dqalloc_reservation(mp, true);
fs/xfs/libxfs/xfs_trans_resv.c
1242
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1244
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
fs/xfs/libxfs/xfs_trans_resv.c
1263
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1268
resp->tr_rename.tr_logres = xfs_calc_rename_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1269
resp->tr_rename.tr_logcount = xfs_rename_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1272
resp->tr_link.tr_logres = xfs_calc_link_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1273
resp->tr_link.tr_logcount = xfs_link_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1276
resp->tr_remove.tr_logres = xfs_calc_remove_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1277
resp->tr_remove.tr_logcount = xfs_remove_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1280
resp->tr_symlink.tr_logres = xfs_calc_symlink_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1281
resp->tr_symlink.tr_logcount = xfs_symlink_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1284
resp->tr_create.tr_logres = xfs_calc_icreate_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1285
resp->tr_create.tr_logcount = xfs_icreate_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1288
resp->tr_mkdir.tr_logres = xfs_calc_mkdir_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1289
resp->tr_mkdir.tr_logcount = xfs_mkdir_log_count(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1295
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1304
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1313
resp->tr_write.tr_logres = xfs_calc_write_reservation(mp, false);
fs/xfs/libxfs/xfs_trans_resv.c
1317
resp->tr_itruncate.tr_logres = xfs_calc_itruncate_reservation(mp, false);
fs/xfs/libxfs/xfs_trans_resv.c
1322
xfs_calc_create_tmpfile_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1326
resp->tr_ifree.tr_logres = xfs_calc_ifree_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1330
resp->tr_addafork.tr_logres = xfs_calc_addafork_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1334
resp->tr_attrinval.tr_logres = xfs_calc_attrinval_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1338
resp->tr_attrsetm.tr_logres = xfs_calc_attrsetm_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1342
resp->tr_attrrm.tr_logres = xfs_calc_attrrm_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1346
resp->tr_growrtalloc.tr_logres = xfs_calc_growrtalloc_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1350
resp->tr_qm_dqalloc.tr_logres = xfs_calc_qm_dqalloc_reservation(mp,
fs/xfs/libxfs/xfs_trans_resv.c
1355
xfs_calc_namespace_reservations(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1364
resp->tr_sb.tr_logres = xfs_calc_sb_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1368
resp->tr_growdata.tr_logres = xfs_calc_growdata_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1373
resp->tr_ichange.tr_logres = xfs_calc_ichange_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1374
resp->tr_fsyncts.tr_logres = xfs_calc_swrite_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1375
resp->tr_writeid.tr_logres = xfs_calc_writeid_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1376
resp->tr_attrsetrt.tr_logres = xfs_calc_attrsetrt_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1377
resp->tr_clearagi.tr_logres = xfs_calc_clear_agi_bucket_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1378
resp->tr_growrtzero.tr_logres = xfs_calc_growrtzero_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1379
resp->tr_growrtfree.tr_logres = xfs_calc_growrtfree_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1386
if (xfs_has_reflink(mp) || xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
1388
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_trans_resv.c
1390
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
140
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1401
xfs_calc_default_atomic_ioend_reservation(mp, resp);
fs/xfs/libxfs/xfs_trans_resv.c
1410
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1454
const unsigned int f1 = xfs_calc_finish_efi_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
1455
const unsigned int f2 = xfs_calc_finish_rui_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
1456
const unsigned int f3 = xfs_calc_finish_cui_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
1457
const unsigned int f4 = xfs_calc_finish_bui_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
146
mp->m_sb.sb_inodesize +
fs/xfs/libxfs/xfs_trans_resv.c
147
2 * xfs_bmbt_block_len(mp));
fs/xfs/libxfs/xfs_trans_resv.c
1471
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
1473
const struct xfs_trans_res *resv = &M_RES(mp)->tr_atomic_ioend;
fs/xfs/libxfs/xfs_trans_resv.c
1479
per_intent = xfs_calc_atomic_write_ioend_geometry(mp,
fs/xfs/libxfs/xfs_trans_resv.c
1486
trace_xfs_calc_max_atomic_write_fsblocks(mp, per_intent, step_size,
fs/xfs/libxfs/xfs_trans_resv.c
1499
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1503
struct xfs_trans_res *curr_res = &M_RES(mp)->tr_atomic_ioend;
fs/xfs/libxfs/xfs_trans_resv.c
1511
xfs_calc_default_atomic_ioend_reservation(mp, M_RES(mp));
fs/xfs/libxfs/xfs_trans_resv.c
1513
per_intent = xfs_calc_atomic_write_ioend_geometry(mp, &step_size);
fs/xfs/libxfs/xfs_trans_resv.c
1521
min_logblocks = xfs_log_calc_minimum_size(mp);
fs/xfs/libxfs/xfs_trans_resv.c
1524
trace_xfs_calc_max_atomic_write_log_geometry(mp, per_intent, step_size,
fs/xfs/libxfs/xfs_trans_resv.c
1537
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
1548
xfs_calc_default_atomic_ioend_reservation(mp, M_RES(mp));
fs/xfs/libxfs/xfs_trans_resv.c
1552
min_logblocks = xfs_calc_atomic_write_log_geometry(mp, blockcount,
fs/xfs/libxfs/xfs_trans_resv.c
1554
if (!min_logblocks || min_logblocks > mp->m_sb.sb_logblocks)
fs/xfs/libxfs/xfs_trans_resv.c
1557
M_RES(mp)->tr_atomic_ioend.tr_logres = new_logres;
fs/xfs/libxfs/xfs_trans_resv.c
162
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
164
return xfs_calc_buf_res(M_IGEO(mp)->inobt_maxlevels,
fs/xfs/libxfs/xfs_trans_resv.c
165
XFS_FSB_TO_B(mp, 1)) +
fs/xfs/libxfs/xfs_trans_resv.c
166
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
fs/xfs/libxfs/xfs_trans_resv.c
167
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
183
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
185
if (!xfs_has_finobt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
188
return xfs_calc_inobt_res(mp);
fs/xfs/libxfs/xfs_trans_resv.c
208
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
213
res = xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1),
fs/xfs/libxfs/xfs_trans_resv.c
214
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
217
if (xfs_has_v3inodes(mp))
fs/xfs/libxfs/xfs_trans_resv.c
219
size = XFS_FSB_TO_B(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
222
res += xfs_calc_buf_res(M_IGEO(mp)->ialloc_blks, size);
fs/xfs/libxfs/xfs_trans_resv.c
236
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
243
rtxlen = xfs_extlen_to_rtxlen(mp, XFS_MAX_BMBT_EXTLEN);
fs/xfs/libxfs/xfs_trans_resv.c
244
rtbmp_blocks = xfs_rtbitmap_blockcount_len(mp, rtxlen);
fs/xfs/libxfs/xfs_trans_resv.c
247
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
248
t2 = num_ops * (2 * mp->m_rtrmap_maxlevels - 1);
fs/xfs/libxfs/xfs_trans_resv.c
279
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
282
if (!xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_trans_resv.c
285
return xfs_calc_buf_res(nr_ops, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
286
xfs_calc_buf_res(xfs_refcountbt_block_count(mp, nr_ops),
fs/xfs/libxfs/xfs_trans_resv.c
287
mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
297
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
300
if (!xfs_has_rtreflink(mp))
fs/xfs/libxfs/xfs_trans_resv.c
303
return xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
304
xfs_calc_buf_res(xfs_rtrefcountbt_block_count(mp, nr_ops),
fs/xfs/libxfs/xfs_trans_resv.c
305
mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
322
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
327
t1 = xfs_calc_finish_cui_reservation(mp, nr_ops);
fs/xfs/libxfs/xfs_trans_resv.c
328
t2 = xfs_calc_finish_rt_cui_reservation(mp, nr_ops);
fs/xfs/libxfs/xfs_trans_resv.c
358
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
362
unsigned int blksz = XFS_FSB_TO_B(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
364
t1 = xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
365
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK), blksz) +
fs/xfs/libxfs/xfs_trans_resv.c
366
xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
367
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz);
fs/xfs/libxfs/xfs_trans_resv.c
369
if (xfs_has_realtime(mp)) {
fs/xfs/libxfs/xfs_trans_resv.c
370
t2 = xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
371
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK),
fs/xfs/libxfs/xfs_trans_resv.c
373
xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
374
xfs_calc_buf_res(xfs_rtalloc_block_count(mp, 1), blksz) +
fs/xfs/libxfs/xfs_trans_resv.c
375
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 1), blksz);
fs/xfs/libxfs/xfs_trans_resv.c
380
t3 = xfs_calc_buf_res(5, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
381
xfs_calc_buf_res(xfs_allocfree_block_count(mp, 2), blksz);
fs/xfs/libxfs/xfs_trans_resv.c
395
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_trans_resv.c
397
xfs_refcountbt_block_count(mp, 2),
fs/xfs/libxfs/xfs_trans_resv.c
404
t4 = xfs_calc_refcountbt_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
410
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
412
return xfs_calc_write_reservation(mp, true);
fs/xfs/libxfs/xfs_trans_resv.c
425
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
428
return xfs_calc_buf_res((2 * nr) + 1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
429
xfs_calc_buf_res(xfs_allocfree_block_count(mp, nr),
fs/xfs/libxfs/xfs_trans_resv.c
430
mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
446
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
449
if (!xfs_has_realtime(mp))
fs/xfs/libxfs/xfs_trans_resv.c
452
return xfs_calc_buf_res((2 * nr) + 1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
453
xfs_calc_buf_res(xfs_rtalloc_block_count(mp, nr),
fs/xfs/libxfs/xfs_trans_resv.c
454
mp->m_sb.sb_blocksize) +
fs/xfs/libxfs/xfs_trans_resv.c
455
xfs_calc_buf_res(xfs_allocfree_block_count(mp, nr),
fs/xfs/libxfs/xfs_trans_resv.c
456
mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
466
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
469
if (!xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
471
return xfs_calc_finish_efi_reservation(mp, nr);
fs/xfs/libxfs/xfs_trans_resv.c
481
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
484
if (!xfs_has_rtrmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
486
return xfs_calc_finish_rt_efi_reservation(mp, nr);
fs/xfs/libxfs/xfs_trans_resv.c
497
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
500
return xfs_calc_inode_res(mp, 1) + XFS_DQUOT_LOGRES +
fs/xfs/libxfs/xfs_trans_resv.c
501
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) + 1,
fs/xfs/libxfs/xfs_trans_resv.c
502
mp->m_sb.sb_blocksize);
fs/xfs/libxfs/xfs_trans_resv.c
528
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
532
unsigned int blksz = XFS_FSB_TO_B(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
534
t1 = xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
535
xfs_calc_buf_res(XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) + 1, blksz);
fs/xfs/libxfs/xfs_trans_resv.c
537
t2 = xfs_calc_finish_efi_reservation(mp, 4);
fs/xfs/libxfs/xfs_trans_resv.c
538
t3 = xfs_calc_finish_rt_efi_reservation(mp, 2);
fs/xfs/libxfs/xfs_trans_resv.c
550
if (xfs_has_reflink(mp))
fs/xfs/libxfs/xfs_trans_resv.c
552
xfs_refcountbt_block_count(mp, 4),
fs/xfs/libxfs/xfs_trans_resv.c
558
t4 = xfs_calc_refcountbt_reservation(mp, 2);
fs/xfs/libxfs/xfs_trans_resv.c
564
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
566
return xfs_calc_itruncate_reservation(mp, true);
fs/xfs/libxfs/xfs_trans_resv.c
609
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
612
struct xfs_trans_resv *resp = M_RES(mp);
fs/xfs/libxfs/xfs_trans_resv.c
615
t1 = xfs_calc_inode_res(mp, 5) +
fs/xfs/libxfs/xfs_trans_resv.c
616
xfs_calc_buf_res(2 * XFS_DIROP_LOG_COUNT(mp),
fs/xfs/libxfs/xfs_trans_resv.c
617
XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
619
t2 = xfs_calc_finish_efi_reservation(mp, 3);
fs/xfs/libxfs/xfs_trans_resv.c
621
if (xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_trans_resv.c
656
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
666
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
681
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
683
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
684
2 * M_IGEO(mp)->inode_cluster_size;
fs/xfs/libxfs/xfs_trans_resv.c
689
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
698
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
718
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
721
struct xfs_trans_resv *resp = M_RES(mp);
fs/xfs/libxfs/xfs_trans_resv.c
724
overhead += xfs_calc_iunlink_remove_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
725
t1 = xfs_calc_inode_res(mp, 2) +
fs/xfs/libxfs/xfs_trans_resv.c
726
xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
727
t2 = xfs_calc_finish_efi_reservation(mp, 1);
fs/xfs/libxfs/xfs_trans_resv.c
729
if (xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_trans_resv.c
743
xfs_calc_iunlink_add_reservation(xfs_mount_t *mp)
fs/xfs/libxfs/xfs_trans_resv.c
745
return xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
746
M_IGEO(mp)->inode_cluster_size;
fs/xfs/libxfs/xfs_trans_resv.c
751
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
76
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
760
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
780
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
783
struct xfs_trans_resv *resp = M_RES(mp);
fs/xfs/libxfs/xfs_trans_resv.c
786
overhead += xfs_calc_iunlink_add_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
788
t1 = xfs_calc_inode_res(mp, 2) +
fs/xfs/libxfs/xfs_trans_resv.c
789
xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1));
fs/xfs/libxfs/xfs_trans_resv.c
790
t2 = xfs_calc_finish_efi_reservation(mp, 2);
fs/xfs/libxfs/xfs_trans_resv.c
792
if (xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_trans_resv.c
81
blocks = num_ops * 2 * (2 * mp->m_alloc_maxlevels - 1);
fs/xfs/libxfs/xfs_trans_resv.c
818
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
82
if (xfs_has_rmapbt(mp))
fs/xfs/libxfs/xfs_trans_resv.c
820
return xfs_calc_inode_res(mp, 2) +
fs/xfs/libxfs/xfs_trans_resv.c
821
xfs_calc_buf_res(1, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
822
(uint)XFS_FSB_TO_B(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
823
xfs_calc_buf_res(XFS_DIROP_LOG_COUNT(mp), XFS_FSB_TO_B(mp, 1)) +
fs/xfs/libxfs/xfs_trans_resv.c
824
xfs_calc_finobt_res(mp);
fs/xfs/libxfs/xfs_trans_resv.c
83
blocks += num_ops * (2 * mp->m_rmap_maxlevels - 1);
fs/xfs/libxfs/xfs_trans_resv.c
837
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
839
return xfs_calc_buf_res(2, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
840
mp->m_sb.sb_sectsize +
fs/xfs/libxfs/xfs_trans_resv.c
841
xfs_calc_inode_chunk_res(mp, _ALLOC) +
fs/xfs/libxfs/xfs_trans_resv.c
842
xfs_calc_inobt_res(mp) +
fs/xfs/libxfs/xfs_trans_resv.c
843
xfs_calc_finobt_res(mp);
fs/xfs/libxfs/xfs_trans_resv.c
848
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
857
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
865
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
867
struct xfs_trans_resv *resp = M_RES(mp);
fs/xfs/libxfs/xfs_trans_resv.c
871
t1 = xfs_calc_icreate_resv_alloc(mp);
fs/xfs/libxfs/xfs_trans_resv.c
872
t2 = xfs_calc_create_resv_modify(mp);
fs/xfs/libxfs/xfs_trans_resv.c
874
if (xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_trans_resv.c
884
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
888
res += xfs_calc_icreate_resv_alloc(mp);
fs/xfs/libxfs/xfs_trans_resv.c
889
return res + xfs_calc_iunlink_add_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
894
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
903
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
914
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
916
return xfs_calc_icreate_reservation(mp);
fs/xfs/libxfs/xfs_trans_resv.c
921
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
930
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_resv.c
943
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
945
return xfs_calc_icreate_reservation(mp) +
fs/xfs/libxfs/xfs_trans_resv.c
95
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.c
966
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
969
xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
970
xfs_calc_buf_res(3, mp->m_sb.sb_sectsize) +
fs/xfs/libxfs/xfs_trans_resv.c
971
xfs_calc_iunlink_remove_reservation(mp) +
fs/xfs/libxfs/xfs_trans_resv.c
972
xfs_calc_inode_chunk_res(mp, _FREE) +
fs/xfs/libxfs/xfs_trans_resv.c
973
xfs_calc_inobt_res(mp) +
fs/xfs/libxfs/xfs_trans_resv.c
974
xfs_calc_finobt_res(mp);
fs/xfs/libxfs/xfs_trans_resv.c
98
return num_ops * (2 * mp->m_refc_maxlevels - 1);
fs/xfs/libxfs/xfs_trans_resv.c
983
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.c
986
xfs_calc_inode_res(mp, 1) +
fs/xfs/libxfs/xfs_trans_resv.c
987
xfs_calc_buf_res(1, mp->m_sb.sb_sectsize);
fs/xfs/libxfs/xfs_trans_resv.c
999
struct xfs_mount *mp)
fs/xfs/libxfs/xfs_trans_resv.h
100
uint xfs_allocfree_block_count(struct xfs_mount *mp, uint num_ops);
fs/xfs/libxfs/xfs_trans_resv.h
102
unsigned int xfs_calc_finish_bui_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
105
unsigned int xfs_calc_finish_efi_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
107
unsigned int xfs_calc_finish_rt_efi_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
110
unsigned int xfs_calc_finish_rui_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
112
unsigned int xfs_calc_finish_rt_rui_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
115
unsigned int xfs_calc_finish_cui_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
117
unsigned int xfs_calc_finish_rt_cui_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
120
unsigned int xfs_calc_itruncate_reservation_minlogsize(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_trans_resv.h
121
unsigned int xfs_calc_write_reservation_minlogsize(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_trans_resv.h
122
unsigned int xfs_calc_qm_dqalloc_reservation_minlogsize(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_trans_resv.h
124
xfs_extlen_t xfs_calc_max_atomic_write_fsblocks(struct xfs_mount *mp);
fs/xfs/libxfs/xfs_trans_resv.h
125
xfs_extlen_t xfs_calc_atomic_write_log_geometry(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
127
int xfs_calc_atomic_write_reservation(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_resv.h
55
#define M_RES(mp) (&(mp)->m_resv)
fs/xfs/libxfs/xfs_trans_resv.h
64
#define XFS_DIROP_LOG_RES(mp) \
fs/xfs/libxfs/xfs_trans_resv.h
65
(XFS_FSB_TO_B(mp, XFS_DAENTER_BLOCKS(mp, XFS_DATA_FORK)) + \
fs/xfs/libxfs/xfs_trans_resv.h
66
(XFS_FSB_TO_B(mp, XFS_DAENTER_BMAPS(mp, XFS_DATA_FORK) + 1)))
fs/xfs/libxfs/xfs_trans_resv.h
67
#define XFS_DIROP_LOG_COUNT(mp) \
fs/xfs/libxfs/xfs_trans_resv.h
68
(XFS_DAENTER_BLOCKS(mp, XFS_DATA_FORK) + \
fs/xfs/libxfs/xfs_trans_resv.h
69
XFS_DAENTER_BMAPS(mp, XFS_DATA_FORK) + 1)
fs/xfs/libxfs/xfs_trans_resv.h
99
void xfs_trans_resv_calc(struct xfs_mount *mp, struct xfs_trans_resv *resp);
fs/xfs/libxfs/xfs_trans_space.c
100
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
108
ret = XFS_DIRREMOVE_SPACE_RES(mp) +
fs/xfs/libxfs/xfs_trans_space.c
109
XFS_DIRENTER_SPACE_RES(mp, target_namelen);
fs/xfs/libxfs/xfs_trans_space.c
111
if (xfs_has_parent(mp)) {
fs/xfs/libxfs/xfs_trans_space.c
113
ret += xfs_parent_calc_space_res(mp, src_namelen);
fs/xfs/libxfs/xfs_trans_space.c
114
ret += 2 * xfs_parent_calc_space_res(mp, target_namelen);
fs/xfs/libxfs/xfs_trans_space.c
118
ret += xfs_parent_calc_space_res(mp, target_namelen);
fs/xfs/libxfs/xfs_trans_space.c
21
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
28
return XFS_DAENTER_SPACE_RES(mp, XFS_ATTR_FORK) +
fs/xfs/libxfs/xfs_trans_space.c
29
XFS_NEXTENTADD_SPACE_RES(mp, namelen, XFS_ATTR_FORK);
fs/xfs/libxfs/xfs_trans_space.c
34
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
39
ret = XFS_IALLOC_SPACE_RES(mp) + XFS_DIRENTER_SPACE_RES(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
40
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_space.c
41
ret += xfs_parent_calc_space_res(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
48
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
51
return xfs_create_space_res(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
56
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
61
ret = XFS_DIRENTER_SPACE_RES(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
62
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_space.c
63
ret += xfs_parent_calc_space_res(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
70
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
76
ret = XFS_IALLOC_SPACE_RES(mp) + XFS_DIRENTER_SPACE_RES(mp, namelen) +
fs/xfs/libxfs/xfs_trans_space.c
79
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_space.c
80
ret += xfs_parent_calc_space_res(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.c
87
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.c
90
unsigned int ret = XFS_DIRREMOVE_SPACE_RES(mp);
fs/xfs/libxfs/xfs_trans_space.c
92
if (xfs_has_parent(mp))
fs/xfs/libxfs/xfs_trans_space.c
93
ret += xfs_parent_calc_space_res(mp, namelen);
fs/xfs/libxfs/xfs_trans_space.h
100
#define XFS_GROWFSRT_SPACE_RES(mp,b) \
fs/xfs/libxfs/xfs_trans_space.h
101
((b) + XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK))
fs/xfs/libxfs/xfs_trans_space.h
102
#define XFS_QM_DQALLOC_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
103
(XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK) + \
fs/xfs/libxfs/xfs_trans_space.h
105
#define XFS_QM_QINOCREATE_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
106
XFS_IALLOC_SPACE_RES(mp)
fs/xfs/libxfs/xfs_trans_space.h
107
#define XFS_IFREE_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
108
(xfs_has_finobt(mp) ? M_IGEO(mp)->inobt_maxlevels : 0)
fs/xfs/libxfs/xfs_trans_space.h
110
unsigned int xfs_parent_calc_space_res(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.h
113
unsigned int xfs_create_space_res(struct xfs_mount *mp, unsigned int namelen);
fs/xfs/libxfs/xfs_trans_space.h
114
unsigned int xfs_mkdir_space_res(struct xfs_mount *mp, unsigned int namelen);
fs/xfs/libxfs/xfs_trans_space.h
115
unsigned int xfs_link_space_res(struct xfs_mount *mp, unsigned int namelen);
fs/xfs/libxfs/xfs_trans_space.h
116
unsigned int xfs_symlink_space_res(struct xfs_mount *mp, unsigned int namelen,
fs/xfs/libxfs/xfs_trans_space.h
118
unsigned int xfs_remove_space_res(struct xfs_mount *mp, unsigned int namelen);
fs/xfs/libxfs/xfs_trans_space.h
120
unsigned int xfs_rename_space_res(struct xfs_mount *mp,
fs/xfs/libxfs/xfs_trans_space.h
14
#define XFS_MAX_CONTIG_BMAPS_PER_BLOCK(mp) \
fs/xfs/libxfs/xfs_trans_space.h
15
(((mp)->m_bmap_dmxr[0]) - ((mp)->m_bmap_dmnr[0]))
fs/xfs/libxfs/xfs_trans_space.h
18
#define XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp) \
fs/xfs/libxfs/xfs_trans_space.h
19
(((mp)->m_rtrmap_mxr[0]) - ((mp)->m_rtrmap_mnr[0]))
fs/xfs/libxfs/xfs_trans_space.h
22
#define XFS_RTRMAPADD_SPACE_RES(mp) ((mp)->m_rtrmap_maxlevels)
fs/xfs/libxfs/xfs_trans_space.h
25
#define XFS_NRTRMAPADD_SPACE_RES(mp, b) \
fs/xfs/libxfs/xfs_trans_space.h
26
((((b) + XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp) - 1) / \
fs/xfs/libxfs/xfs_trans_space.h
27
XFS_MAX_CONTIG_RTRMAPS_PER_BLOCK(mp)) * \
fs/xfs/libxfs/xfs_trans_space.h
28
XFS_RTRMAPADD_SPACE_RES(mp))
fs/xfs/libxfs/xfs_trans_space.h
31
#define XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp) \
fs/xfs/libxfs/xfs_trans_space.h
32
(((mp)->m_rmap_mxr[0]) - ((mp)->m_rmap_mnr[0]))
fs/xfs/libxfs/xfs_trans_space.h
35
#define XFS_RMAPADD_SPACE_RES(mp) ((mp)->m_rmap_maxlevels)
fs/xfs/libxfs/xfs_trans_space.h
45
#define XFS_NRMAPADD_SPACE_RES(mp, b)\
fs/xfs/libxfs/xfs_trans_space.h
46
(((b + XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp) - 1) / \
fs/xfs/libxfs/xfs_trans_space.h
47
XFS_MAX_CONTIG_RMAPS_PER_BLOCK(mp)) * \
fs/xfs/libxfs/xfs_trans_space.h
48
XFS_RMAPADD_SPACE_RES(mp))
fs/xfs/libxfs/xfs_trans_space.h
50
#define XFS_MAX_CONTIG_EXTENTS_PER_BLOCK(mp) \
fs/xfs/libxfs/xfs_trans_space.h
51
(((mp)->m_alloc_mxr[0]) - ((mp)->m_alloc_mnr[0]))
fs/xfs/libxfs/xfs_trans_space.h
52
#define XFS_EXTENTADD_SPACE_RES(mp,w) (XFS_BM_MAXLEVELS(mp,w) - 1)
fs/xfs/libxfs/xfs_trans_space.h
53
#define XFS_NEXTENTADD_SPACE_RES(mp,b,w)\
fs/xfs/libxfs/xfs_trans_space.h
54
(((b + XFS_MAX_CONTIG_EXTENTS_PER_BLOCK(mp) - 1) / \
fs/xfs/libxfs/xfs_trans_space.h
55
XFS_MAX_CONTIG_EXTENTS_PER_BLOCK(mp)) * \
fs/xfs/libxfs/xfs_trans_space.h
56
XFS_EXTENTADD_SPACE_RES(mp,w))
fs/xfs/libxfs/xfs_trans_space.h
59
#define XFS_SWAP_RMAP_SPACE_RES(mp,b,w)\
fs/xfs/libxfs/xfs_trans_space.h
60
(XFS_NEXTENTADD_SPACE_RES((mp), (b), (w)) + \
fs/xfs/libxfs/xfs_trans_space.h
61
XFS_NRMAPADD_SPACE_RES((mp), (b)))
fs/xfs/libxfs/xfs_trans_space.h
63
#define XFS_DAENTER_1B(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
64
((w) == XFS_DATA_FORK ? (mp)->m_dir_geo->fsbcount : 1)
fs/xfs/libxfs/xfs_trans_space.h
65
#define XFS_DAENTER_DBS(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
67
#define XFS_DAENTER_BLOCKS(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
68
(XFS_DAENTER_1B(mp,w) * XFS_DAENTER_DBS(mp,w))
fs/xfs/libxfs/xfs_trans_space.h
69
#define XFS_DAENTER_BMAP1B(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
70
XFS_NEXTENTADD_SPACE_RES(mp, XFS_DAENTER_1B(mp, w), w)
fs/xfs/libxfs/xfs_trans_space.h
71
#define XFS_DAENTER_BMAPS(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
72
(XFS_DAENTER_DBS(mp,w) * XFS_DAENTER_BMAP1B(mp,w))
fs/xfs/libxfs/xfs_trans_space.h
73
#define XFS_DAENTER_SPACE_RES(mp,w) \
fs/xfs/libxfs/xfs_trans_space.h
74
(XFS_DAENTER_BLOCKS(mp,w) + XFS_DAENTER_BMAPS(mp,w))
fs/xfs/libxfs/xfs_trans_space.h
75
#define XFS_DAREMOVE_SPACE_RES(mp,w) XFS_DAENTER_BMAPS(mp,w)
fs/xfs/libxfs/xfs_trans_space.h
76
#define XFS_DIRENTER_MAX_SPLIT(mp,nl) 1
fs/xfs/libxfs/xfs_trans_space.h
77
#define XFS_DIRENTER_SPACE_RES(mp,nl) \
fs/xfs/libxfs/xfs_trans_space.h
78
(XFS_DAENTER_SPACE_RES(mp, XFS_DATA_FORK) * \
fs/xfs/libxfs/xfs_trans_space.h
79
XFS_DIRENTER_MAX_SPLIT(mp,nl))
fs/xfs/libxfs/xfs_trans_space.h
80
#define XFS_DIRREMOVE_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
81
XFS_DAREMOVE_SPACE_RES(mp, XFS_DATA_FORK)
fs/xfs/libxfs/xfs_trans_space.h
82
#define XFS_IALLOC_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
83
(M_IGEO(mp)->ialloc_blks + \
fs/xfs/libxfs/xfs_trans_space.h
84
((xfs_has_finobt(mp) ? 2 : 1) * M_IGEO(mp)->inobt_maxlevels))
fs/xfs/libxfs/xfs_trans_space.h
89
#define XFS_ADDAFORK_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
90
((mp)->m_dir_geo->fsbcount + XFS_DAENTER_BMAP1B(mp, XFS_DATA_FORK))
fs/xfs/libxfs/xfs_trans_space.h
91
#define XFS_ATTRRM_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
92
XFS_DAREMOVE_SPACE_RES(mp, XFS_ATTR_FORK)
fs/xfs/libxfs/xfs_trans_space.h
94
#define XFS_ATTRSET_SPACE_RES(mp, v) \
fs/xfs/libxfs/xfs_trans_space.h
95
(XFS_DAENTER_SPACE_RES(mp, XFS_ATTR_FORK) + XFS_B_TO_FSB(mp, v))
fs/xfs/libxfs/xfs_trans_space.h
96
#define XFS_DIOSTRAT_SPACE_RES(mp, v) \
fs/xfs/libxfs/xfs_trans_space.h
97
(XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK) + (v))
fs/xfs/libxfs/xfs_trans_space.h
98
#define XFS_GROWFS_SPACE_RES(mp) \
fs/xfs/libxfs/xfs_trans_space.h
99
(2 * (mp)->m_alloc_maxlevels)
fs/xfs/libxfs/xfs_types.c
101
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
104
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, ino);
fs/xfs/libxfs/xfs_types.c
105
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ino);
fs/xfs/libxfs/xfs_types.c
107
if (agno >= mp->m_sb.sb_agcount)
fs/xfs/libxfs/xfs_types.c
109
if (XFS_AGINO_TO_INO(mp, agno, agino) != ino)
fs/xfs/libxfs/xfs_types.c
111
return xfs_verify_agno_agino(mp, agno, agino);
fs/xfs/libxfs/xfs_types.c
117
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
120
return ino == mp->m_sb.sb_rbmino || ino == mp->m_sb.sb_rsumino ||
fs/xfs/libxfs/xfs_types.c
121
(xfs_has_quota(mp) &&
fs/xfs/libxfs/xfs_types.c
122
xfs_is_quota_inode(&mp->m_sb, ino));
fs/xfs/libxfs/xfs_types.c
131
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
134
if (xfs_is_sb_inum(mp, ino))
fs/xfs/libxfs/xfs_types.c
136
return xfs_verify_ino(mp, ino);
fs/xfs/libxfs/xfs_types.c
146
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
149
if (xfs_has_rtgroups(mp)) {
fs/xfs/libxfs/xfs_types.c
150
xfs_rgnumber_t rgno = xfs_rtb_to_rgno(mp, rtbno);
fs/xfs/libxfs/xfs_types.c
151
xfs_rtxnum_t rtx = xfs_rtb_to_rtx(mp, rtbno);
fs/xfs/libxfs/xfs_types.c
153
if (rgno >= mp->m_sb.sb_rgcount)
fs/xfs/libxfs/xfs_types.c
155
if (rtx >= xfs_rtgroup_extents(mp, rgno))
fs/xfs/libxfs/xfs_types.c
157
if (xfs_has_rtsb(mp) && rgno == 0 && rtx == 0)
fs/xfs/libxfs/xfs_types.c
162
return rtbno < mp->m_sb.sb_rblocks;
fs/xfs/libxfs/xfs_types.c
172
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
179
if (!xfs_verify_rtbno(mp, rtbno))
fs/xfs/libxfs/xfs_types.c
182
if (!xfs_verify_rtbno(mp, rtbno + len - 1))
fs/xfs/libxfs/xfs_types.c
185
if (xfs_has_rtgroups(mp) &&
fs/xfs/libxfs/xfs_types.c
186
xfs_rtb_to_rgno(mp, rtbno) != xfs_rtb_to_rgno(mp, rtbno + len - 1))
fs/xfs/libxfs/xfs_types.c
195
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
205
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/libxfs/xfs_types.c
213
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
218
xfs_icount_range(mp, &min, &max);
fs/xfs/libxfs/xfs_types.c
225
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
236
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
245
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
25
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
252
if (!xfs_verify_fileoff(mp, off))
fs/xfs/libxfs/xfs_types.c
255
return xfs_verify_fileoff(mp, off + len - 1);
fs/xfs/libxfs/xfs_types.c
31
eoag = xfs_ag_block_count(mp, agno);
fs/xfs/libxfs/xfs_types.c
34
if (agbno <= XFS_AGFL_BLOCK(mp))
fs/xfs/libxfs/xfs_types.c
45
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
48
xfs_agnumber_t agno = XFS_FSB_TO_AGNO(mp, fsbno);
fs/xfs/libxfs/xfs_types.c
50
if (agno >= mp->m_sb.sb_agcount)
fs/xfs/libxfs/xfs_types.c
52
return xfs_verify_agno_agbno(mp, agno, XFS_FSB_TO_AGBNO(mp, fsbno));
fs/xfs/libxfs/xfs_types.c
61
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
68
if (!xfs_verify_fsbno(mp, fsbno))
fs/xfs/libxfs/xfs_types.c
71
if (!xfs_verify_fsbno(mp, fsbno + len - 1))
fs/xfs/libxfs/xfs_types.c
74
return XFS_FSB_TO_AGNO(mp, fsbno) ==
fs/xfs/libxfs/xfs_types.c
75
XFS_FSB_TO_AGNO(mp, fsbno + len - 1);
fs/xfs/libxfs/xfs_types.c
84
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_types.c
91
xfs_agino_range(mp, agno, &first, &last);
fs/xfs/libxfs/xfs_types.h
269
bool xfs_verify_fsbno(struct xfs_mount *mp, xfs_fsblock_t fsbno);
fs/xfs/libxfs/xfs_types.h
270
bool xfs_verify_fsbext(struct xfs_mount *mp, xfs_fsblock_t fsbno,
fs/xfs/libxfs/xfs_types.h
273
bool xfs_verify_ino(struct xfs_mount *mp, xfs_ino_t ino);
fs/xfs/libxfs/xfs_types.h
274
bool xfs_is_sb_inum(struct xfs_mount *mp, xfs_ino_t ino);
fs/xfs/libxfs/xfs_types.h
275
bool xfs_verify_dir_ino(struct xfs_mount *mp, xfs_ino_t ino);
fs/xfs/libxfs/xfs_types.h
276
bool xfs_verify_rtbno(struct xfs_mount *mp, xfs_rtblock_t rtbno);
fs/xfs/libxfs/xfs_types.h
277
bool xfs_verify_rtbext(struct xfs_mount *mp, xfs_rtblock_t rtbno,
fs/xfs/libxfs/xfs_types.h
279
bool xfs_verify_icount(struct xfs_mount *mp, unsigned long long icount);
fs/xfs/libxfs/xfs_types.h
280
bool xfs_verify_dablk(struct xfs_mount *mp, xfs_fileoff_t off);
fs/xfs/libxfs/xfs_types.h
281
void xfs_icount_range(struct xfs_mount *mp, unsigned long long *min,
fs/xfs/libxfs/xfs_types.h
283
bool xfs_verify_fileoff(struct xfs_mount *mp, xfs_fileoff_t off);
fs/xfs/libxfs/xfs_types.h
284
bool xfs_verify_fileext(struct xfs_mount *mp, xfs_fileoff_t off,
fs/xfs/libxfs/xfs_zones.c
100
zone_no, XFS_BB_TO_FSB(mp, zone->len),
fs/xfs/libxfs/xfs_zones.c
107
return xfs_validate_blk_zone_conv(mp, zone, zone_no);
fs/xfs/libxfs/xfs_zones.c
109
return xfs_validate_blk_zone_seq(mp, zone, zone_no,
fs/xfs/libxfs/xfs_zones.c
112
xfs_warn(mp, "zoned %u has unsupported type 0x%x.",
fs/xfs/libxfs/xfs_zones.c
19
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_zones.c
34
xfs_warn(mp,
fs/xfs/libxfs/xfs_zones.c
40
*write_pointer = XFS_BB_TO_FSB(mp, zone->wp - zone->start);
fs/xfs/libxfs/xfs_zones.c
43
*write_pointer = XFS_BB_TO_FSB(mp, zone->capacity);
fs/xfs/libxfs/xfs_zones.c
48
xfs_warn(mp, "zone %u has unsupported zone condition 0x%x.",
fs/xfs/libxfs/xfs_zones.c
52
xfs_warn(mp, "zone %u has unknown zone condition 0x%x.",
fs/xfs/libxfs/xfs_zones.c
60
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_zones.c
68
xfs_warn(mp,
fs/xfs/libxfs/xfs_zones.c
77
struct xfs_mount *mp,
fs/xfs/libxfs/xfs_zones.c
89
if (XFS_BB_TO_FSB(mp, zone->capacity) != expected_capacity) {
fs/xfs/libxfs/xfs_zones.c
90
xfs_warn(mp,
fs/xfs/libxfs/xfs_zones.c
92
zone_no, XFS_BB_TO_FSB(mp, zone->capacity),
fs/xfs/libxfs/xfs_zones.c
97
if (XFS_BB_TO_FSB(mp, zone->len) != expected_size) {
fs/xfs/libxfs/xfs_zones.c
98
xfs_warn(mp,
fs/xfs/libxfs/xfs_zones.h
40
bool xfs_validate_blk_zone(struct xfs_mount *mp, struct blk_zone *zone,
fs/xfs/scrub/agheader.c
1014
if (xfs_has_finobt(mp)) {
fs/xfs/scrub/agheader.c
102
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
1025
xfs_agino_range(mp, agno, &first_agino, &last_agino);
fs/xfs/scrub/agheader.c
122
pag = xfs_perag_get(mp, agno);
fs/xfs/scrub/agheader.c
126
error = xfs_sb_read_secondary(mp, sc->tp, agno, &bp);
fs/xfs/scrub/agheader.c
144
if (!xchk_process_error(sc, agno, XFS_SB_BLOCK(mp), &error))
fs/xfs/scrub/agheader.c
155
if (sb->sb_blocksize != cpu_to_be32(mp->m_sb.sb_blocksize))
fs/xfs/scrub/agheader.c
158
if (sb->sb_dblocks != cpu_to_be64(mp->m_sb.sb_dblocks))
fs/xfs/scrub/agheader.c
161
if (sb->sb_rblocks != cpu_to_be64(mp->m_sb.sb_rblocks))
fs/xfs/scrub/agheader.c
164
if (sb->sb_rextents != cpu_to_be64(mp->m_sb.sb_rextents))
fs/xfs/scrub/agheader.c
167
if (!uuid_equal(&sb->sb_uuid, &mp->m_sb.sb_uuid))
fs/xfs/scrub/agheader.c
170
if (sb->sb_logstart != cpu_to_be64(mp->m_sb.sb_logstart))
fs/xfs/scrub/agheader.c
173
if (sb->sb_rootino != cpu_to_be64(mp->m_sb.sb_rootino))
fs/xfs/scrub/agheader.c
176
if (xfs_has_metadir(sc->mp)) {
fs/xfs/scrub/agheader.c
183
if (sb->sb_rbmino != cpu_to_be64(mp->m_sb.sb_rbmino))
fs/xfs/scrub/agheader.c
186
if (sb->sb_rsumino != cpu_to_be64(mp->m_sb.sb_rsumino))
fs/xfs/scrub/agheader.c
190
if (sb->sb_rextsize != cpu_to_be32(mp->m_sb.sb_rextsize))
fs/xfs/scrub/agheader.c
193
if (sb->sb_agblocks != cpu_to_be32(mp->m_sb.sb_agblocks))
fs/xfs/scrub/agheader.c
196
if (sb->sb_agcount != cpu_to_be32(mp->m_sb.sb_agcount))
fs/xfs/scrub/agheader.c
199
if (sb->sb_rbmblocks != cpu_to_be32(mp->m_sb.sb_rbmblocks))
fs/xfs/scrub/agheader.c
202
if (sb->sb_logblocks != cpu_to_be32(mp->m_sb.sb_logblocks))
fs/xfs/scrub/agheader.c
215
(cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask))
fs/xfs/scrub/agheader.c
223
(cpu_to_be16(mp->m_sb.sb_versionnum) & vernum_mask))
fs/xfs/scrub/agheader.c
226
if (sb->sb_sectsize != cpu_to_be16(mp->m_sb.sb_sectsize))
fs/xfs/scrub/agheader.c
229
if (sb->sb_inodesize != cpu_to_be16(mp->m_sb.sb_inodesize))
fs/xfs/scrub/agheader.c
232
if (sb->sb_inopblock != cpu_to_be16(mp->m_sb.sb_inopblock))
fs/xfs/scrub/agheader.c
235
if (memcmp(sb->sb_fname, mp->m_sb.sb_fname, sizeof(sb->sb_fname)))
fs/xfs/scrub/agheader.c
238
if (sb->sb_blocklog != mp->m_sb.sb_blocklog)
fs/xfs/scrub/agheader.c
241
if (sb->sb_sectlog != mp->m_sb.sb_sectlog)
fs/xfs/scrub/agheader.c
244
if (sb->sb_inodelog != mp->m_sb.sb_inodelog)
fs/xfs/scrub/agheader.c
247
if (sb->sb_inopblog != mp->m_sb.sb_inopblog)
fs/xfs/scrub/agheader.c
250
if (sb->sb_agblklog != mp->m_sb.sb_agblklog)
fs/xfs/scrub/agheader.c
253
if (sb->sb_rextslog != mp->m_sb.sb_rextslog)
fs/xfs/scrub/agheader.c
256
if (sb->sb_imax_pct != mp->m_sb.sb_imax_pct)
fs/xfs/scrub/agheader.c
264
if (xfs_has_metadir(mp)) {
fs/xfs/scrub/agheader.c
271
if (sb->sb_uquotino != cpu_to_be64(mp->m_sb.sb_uquotino))
fs/xfs/scrub/agheader.c
274
if (sb->sb_gquotino != cpu_to_be64(mp->m_sb.sb_gquotino))
fs/xfs/scrub/agheader.c
283
if (sb->sb_flags != mp->m_sb.sb_flags)
fs/xfs/scrub/agheader.c
286
if (sb->sb_shared_vn != mp->m_sb.sb_shared_vn)
fs/xfs/scrub/agheader.c
289
if (sb->sb_inoalignmt != cpu_to_be32(mp->m_sb.sb_inoalignmt))
fs/xfs/scrub/agheader.c
292
if (sb->sb_unit != cpu_to_be32(mp->m_sb.sb_unit))
fs/xfs/scrub/agheader.c
295
if (sb->sb_width != cpu_to_be32(mp->m_sb.sb_width))
fs/xfs/scrub/agheader.c
298
if (sb->sb_dirblklog != mp->m_sb.sb_dirblklog)
fs/xfs/scrub/agheader.c
301
if (sb->sb_logsectlog != mp->m_sb.sb_logsectlog)
fs/xfs/scrub/agheader.c
304
if (sb->sb_logsectsize != cpu_to_be16(mp->m_sb.sb_logsectsize))
fs/xfs/scrub/agheader.c
307
if (sb->sb_logsunit != cpu_to_be32(mp->m_sb.sb_logsunit))
fs/xfs/scrub/agheader.c
311
if (!xfs_sb_version_hasmorebits(&mp->m_sb)) {
fs/xfs/scrub/agheader.c
316
if (xfs_sb_is_v5(&mp->m_sb))
fs/xfs/scrub/agheader.c
332
(cpu_to_be32(mp->m_sb.sb_features2) & features_mask))
fs/xfs/scrub/agheader.c
338
(cpu_to_be32(mp->m_sb.sb_features2) & features_mask))
fs/xfs/scrub/agheader.c
341
if (!xfs_has_crc(mp)) {
fs/xfs/scrub/agheader.c
350
cpu_to_be32(mp->m_sb.sb_features_compat))
fs/xfs/scrub/agheader.c
355
cpu_to_be32(mp->m_sb.sb_features_ro_compat))
fs/xfs/scrub/agheader.c
363
if ((cpu_to_be32(mp->m_sb.sb_features_incompat) ^
fs/xfs/scrub/agheader.c
368
if ((cpu_to_be32(mp->m_sb.sb_features_incompat) ^
fs/xfs/scrub/agheader.c
382
if (sb->sb_spino_align != cpu_to_be32(mp->m_sb.sb_spino_align))
fs/xfs/scrub/agheader.c
385
if (xfs_has_metadir(mp)) {
fs/xfs/scrub/agheader.c
389
if (sb->sb_pquotino != cpu_to_be64(mp->m_sb.sb_pquotino))
fs/xfs/scrub/agheader.c
39
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
396
if (xfs_has_metauuid(mp)) {
fs/xfs/scrub/agheader.c
398
if (!uuid_equal(&sb->sb_meta_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/scrub/agheader.c
402
if (xfs_has_metadir(mp)) {
fs/xfs/scrub/agheader.c
403
if (sb->sb_metadirino != cpu_to_be64(mp->m_sb.sb_metadirino))
fs/xfs/scrub/agheader.c
406
if (sb->sb_rgcount != cpu_to_be32(mp->m_sb.sb_rgcount))
fs/xfs/scrub/agheader.c
409
if (sb->sb_rgextents != cpu_to_be32(mp->m_sb.sb_rgextents))
fs/xfs/scrub/agheader.c
412
if (sb->sb_rgblklog != mp->m_sb.sb_rgblklog)
fs/xfs/scrub/agheader.c
420
sblen = xchk_superblock_ondisk_size(mp);
fs/xfs/scrub/agheader.c
47
agbno = XFS_SB_BLOCK(mp);
fs/xfs/scrub/agheader.c
503
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
509
if (!xfs_has_lazysbcount(sc->mp))
fs/xfs/scrub/agheader.c
528
if ((xfs_has_rmapbt(mp) && !sc->sa.rmap_cur) ||
fs/xfs/scrub/agheader.c
571
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
577
agbno = XFS_AGF_BLOCK(mp);
fs/xfs/scrub/agheader.c
599
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
613
if (!xchk_process_error(sc, agno, XFS_AGF_BLOCK(sc->mp), &error))
fs/xfs/scrub/agheader.c
635
if (level <= 0 || level > mp->m_alloc_maxlevels)
fs/xfs/scrub/agheader.c
639
if (level <= 0 || level > mp->m_alloc_maxlevels)
fs/xfs/scrub/agheader.c
642
if (xfs_has_rmapbt(mp)) {
fs/xfs/scrub/agheader.c
648
if (level <= 0 || level > mp->m_rmap_maxlevels)
fs/xfs/scrub/agheader.c
652
if (xfs_has_reflink(mp)) {
fs/xfs/scrub/agheader.c
658
if (level <= 0 || level > mp->m_refc_maxlevels)
fs/xfs/scrub/agheader.c
669
fl_count = xfs_agfl_size(mp) - agfl_first + agfl_last + 1;
fs/xfs/scrub/agheader.c
678
if (xfs_has_lazysbcount(sc->mp) &&
fs/xfs/scrub/agheader.c
70
struct xfs_mount *mp)
fs/xfs/scrub/agheader.c
72
if (xfs_has_zoned(mp))
fs/xfs/scrub/agheader.c
722
struct xfs_mount *mp,
fs/xfs/scrub/agheader.c
74
if (xfs_has_metadir(mp))
fs/xfs/scrub/agheader.c
759
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
76
if (xfs_has_metauuid(mp))
fs/xfs/scrub/agheader.c
765
agbno = XFS_AGFL_BLOCK(mp);
fs/xfs/scrub/agheader.c
78
if (xfs_has_crc(mp))
fs/xfs/scrub/agheader.c
796
if (!xchk_process_error(sc, agno, XFS_AGFL_BLOCK(sc->mp), &error))
fs/xfs/scrub/agheader.c
80
if (xfs_sb_version_hasmorebits(&mp->m_sb))
fs/xfs/scrub/agheader.c
803
if (!xchk_process_error(sc, agno, XFS_AGFL_BLOCK(sc->mp), &error))
fs/xfs/scrub/agheader.c
815
if (sai.agflcount > xfs_agfl_size(sc->mp)) {
fs/xfs/scrub/agheader.c
82
if (xfs_has_logv2(mp))
fs/xfs/scrub/agheader.c
827
error = xfs_agfl_walk(sc->mp, sc->sa.agf_bp->b_addr, sai.agfl_bp,
fs/xfs/scrub/agheader.c
84
if (xfs_has_sector(mp))
fs/xfs/scrub/agheader.c
889
if (!xfs_has_inobtcounts(sc->mp))
fs/xfs/scrub/agheader.c
914
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
920
agbno = XFS_AGI_BLOCK(mp);
fs/xfs/scrub/agheader.c
977
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader.c
980
struct xfs_ino_geometry *igeo = M_IGEO(sc->mp);
fs/xfs/scrub/agheader.c
993
if (!xchk_process_error(sc, agno, XFS_AGI_BLOCK(sc->mp), &error))
fs/xfs/scrub/agheader_repair.c
1119
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/scrub/agheader_repair.c
1121
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/scrub/agheader_repair.c
1139
struct xfs_mount *mp = ragi->sc->mp;
fs/xfs/scrub/agheader_repair.c
114
struct xfs_mount *mp,
fs/xfs/scrub/agheader_repair.c
1145
ASSERT(XFS_INO_TO_AGNO(mp, ip->i_ino) == pag_agno(ragi->sc->sa.pag));
fs/xfs/scrub/agheader_repair.c
1148
agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/scrub/agheader_repair.c
1176
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/scrub/agheader_repair.c
1216
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/scrub/agheader_repair.c
1218
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
fs/xfs/scrub/agheader_repair.c
1219
if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
fs/xfs/scrub/agheader_repair.c
1248
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/scrub/agheader_repair.c
1253
xfs_inobt_btrec_to_irec(mp, rec, &irec);
fs/xfs/scrub/agheader_repair.c
1693
xfs_force_summary_recalc(sc->mp);
fs/xfs/scrub/agheader_repair.c
1714
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
1719
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/agheader_repair.c
1731
.maxlevels = M_IGEO(sc->mp)->inobt_maxlevels,
fs/xfs/scrub/agheader_repair.c
1736
.maxlevels = M_IGEO(sc->mp)->inobt_maxlevels,
fs/xfs/scrub/agheader_repair.c
1762
error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp,
fs/xfs/scrub/agheader_repair.c
1763
XFS_AG_DADDR(mp, pag_agno(sc->sa.pag),
fs/xfs/scrub/agheader_repair.c
1764
XFS_AGI_DADDR(mp)),
fs/xfs/scrub/agheader_repair.c
1765
XFS_FSS_TO_BB(mp, 1), 0, &ragi->agi_bp, NULL);
fs/xfs/scrub/agheader_repair.c
186
if (xfs_has_reflink(sc->mp) &&
fs/xfs/scrub/agheader_repair.c
203
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
216
if (xfs_has_crc(mp))
fs/xfs/scrub/agheader_repair.c
217
uuid_copy(&agf->agf_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/agheader_repair.c
240
if (xfs_has_reflink(sc->mp)) {
fs/xfs/scrub/agheader_repair.c
257
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
263
cur = xfs_bnobt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
276
cur = xfs_cntbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
284
cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
295
if (xfs_has_reflink(mp)) {
fs/xfs/scrub/agheader_repair.c
296
cur = xfs_refcountbt_init_cursor(mp, sc->tp, agf_bp,
fs/xfs/scrub/agheader_repair.c
321
xfs_force_summary_recalc(sc->mp);
fs/xfs/scrub/agheader_repair.c
350
.maxlevels = sc->mp->m_alloc_maxlevels,
fs/xfs/scrub/agheader_repair.c
355
.maxlevels = sc->mp->m_alloc_maxlevels,
fs/xfs/scrub/agheader_repair.c
360
.maxlevels = sc->mp->m_rmap_maxlevels,
fs/xfs/scrub/agheader_repair.c
365
.maxlevels = sc->mp->m_refc_maxlevels,
fs/xfs/scrub/agheader_repair.c
372
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
379
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/agheader_repair.c
386
error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp,
fs/xfs/scrub/agheader_repair.c
387
XFS_AG_DADDR(mp, pag_agno(sc->sa.pag),
fs/xfs/scrub/agheader_repair.c
388
XFS_AGF_DADDR(mp)),
fs/xfs/scrub/agheader_repair.c
389
XFS_FSS_TO_BB(mp, 1), 0, &agf_bp, NULL);
fs/xfs/scrub/agheader_repair.c
412
error = xfs_agfl_walk(sc->mp, agf_bp->b_addr, agfl_bp,
fs/xfs/scrub/agheader_repair.c
44
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
538
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
54
error = xfs_sb_get_secondary(mp, sc->tp, agno, &bp);
fs/xfs/scrub/agheader_repair.c
548
cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
555
cur = xfs_bnobt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
562
cur = xfs_cntbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
577
ra.rmap_cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/agheader_repair.c
591
xfs_agfl_size(mp));
fs/xfs/scrub/agheader_repair.c
608
ASSERT(flcount <= xfs_agfl_size(sc->mp));
fs/xfs/scrub/agheader_repair.c
611
xfs_force_summary_recalc(sc->mp);
fs/xfs/scrub/agheader_repair.c
624
agf->agf_fllast = cpu_to_be32(xfs_agfl_size(sc->mp) - 1);
fs/xfs/scrub/agheader_repair.c
64
xfs_sb_to_disk(bp->b_addr, &mp->m_sb);
fs/xfs/scrub/agheader_repair.c
677
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
681
ASSERT(flcount <= xfs_agfl_size(mp));
fs/xfs/scrub/agheader_repair.c
691
uuid_copy(&agfl->agfl_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/agheader_repair.c
70
if (xfs_has_crc(mp)) {
fs/xfs/scrub/agheader_repair.c
718
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
725
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/agheader_repair.c
743
error = xfs_trans_read_buf(mp, sc->tp, mp->m_ddev_targp,
fs/xfs/scrub/agheader_repair.c
744
XFS_AG_DADDR(mp, pag_agno(sc->sa.pag),
fs/xfs/scrub/agheader_repair.c
745
XFS_AGFL_DADDR(mp)),
fs/xfs/scrub/agheader_repair.c
746
XFS_FSS_TO_BB(mp, 1), 0, &agfl_bp, NULL);
fs/xfs/scrub/agheader_repair.c
860
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
878
if (xfs_has_finobt(mp) &&
fs/xfs/scrub/agheader_repair.c
898
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
908
if (xfs_has_crc(mp))
fs/xfs/scrub/agheader_repair.c
909
uuid_copy(&agi->agi_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/agheader_repair.c
928
if (xfs_has_finobt(sc->mp)) {
fs/xfs/scrub/agheader_repair.c
943
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/agheader_repair.c
952
if (xfs_has_inobtcounts(mp)) {
fs/xfs/scrub/agheader_repair.c
965
if (xfs_has_finobt(mp) && xfs_has_inobtcounts(mp)) {
fs/xfs/scrub/alloc_repair.c
264
struct xfs_mount *mp,
fs/xfs/scrub/alloc_repair.c
365
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/alloc_repair.c
396
error = xfs_agfl_walk(mp, agf, agfl_bp, xrep_abt_walk_agfl, ra);
fs/xfs/scrub/alloc_repair.c
738
bno_cur = xfs_bnobt_init_cursor(sc->mp, NULL, NULL, pag);
fs/xfs/scrub/alloc_repair.c
741
cnt_cur = xfs_cntbt_init_cursor(sc->mp, NULL, NULL, pag);
fs/xfs/scrub/alloc_repair.c
851
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/alloc_repair.c
856
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/alloc_repair.c
878
error = xfarray_create("free space records", mp->m_sb.sb_agblocks / 2,
fs/xfs/scrub/attr.c
178
.geo = sc->mp->m_attr_geo,
fs/xfs/scrub/attr.c
214
!xfs_parent_valuecheck(sc->mp, value, valuelen)) {
fs/xfs/scrub/attr.c
273
unsigned int mapsize = sc->mp->m_attr_geo->blksize;
fs/xfs/scrub/attr.c
306
struct xfs_mount *mp = ds->state->mp;
fs/xfs/scrub/attr.c
324
nameidx >= mp->m_attr_geo->blksize) {
fs/xfs/scrub/attr.c
363
struct xfs_mount *mp = ds->state->mp;
fs/xfs/scrub/attr.c
382
bitmap_zero(ab->usedmap, mp->m_attr_geo->blksize);
fs/xfs/scrub/attr.c
383
bitmap_zero(ab->freemap, mp->m_attr_geo->blksize);
fs/xfs/scrub/attr.c
386
if (xfs_has_crc(ds->sc->mp)) {
fs/xfs/scrub/attr.c
398
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &leafhdr, leaf);
fs/xfs/scrub/attr.c
412
if (leafhdr.usedbytes > mp->m_attr_geo->blksize)
fs/xfs/scrub/attr.c
414
if (leafhdr.firstused > mp->m_attr_geo->blksize)
fs/xfs/scrub/attr.c
433
buf_end = (char *)bp->b_addr + mp->m_attr_geo->blksize;
fs/xfs/scrub/attr.c
473
mp->m_attr_geo->blksize))
fs/xfs/scrub/attr.c
489
struct xfs_mount *mp = ds->state->mp;
fs/xfs/scrub/attr.c
521
if (nameidx < hdrsize || nameidx >= mp->m_attr_geo->blksize) {
fs/xfs/scrub/attr.c
544
calc_hash = xfs_attr_hashval(mp, ent->flags, lentry->nameval,
fs/xfs/scrub/attr.c
559
calc_hash = xfs_attr_hashval(mp, ent->flags, rentry->name,
fs/xfs/scrub/attr.c
83
bmp_sz = sizeof(long) * BITS_TO_LONGS(sc->mp->m_attr_geo->blksize);
fs/xfs/scrub/attr_repair.c
1294
.geo = sc->mp->m_attr_geo,
fs/xfs/scrub/attr_repair.c
1388
if (!xfs_has_parent(sc->mp))
fs/xfs/scrub/attr_repair.c
1513
if (xfs_has_parent(rx->sc->mp))
fs/xfs/scrub/attr_repair.c
1514
xfs_dir_hook_del(rx->sc->mp, &rx->dhook);
fs/xfs/scrub/attr_repair.c
152
if (xfs_has_parent(sc->mp))
fs/xfs/scrub/attr_repair.c
1551
max_len = xfs_attr_leaf_entsize_local_max(sc->mp->m_attr_geo->blksize);
fs/xfs/scrub/attr_repair.c
1568
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/attr_repair.c
1583
error = xfs_dir_hook_add(sc->mp, &rx->dhook);
fs/xfs/scrub/attr_repair.c
1623
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/attr_repair.c
1626
if (!xfs_has_exchange_range(sc->mp))
fs/xfs/scrub/attr_repair.c
181
return xfs_parent_valuecheck(rx->sc->mp, value, valuelen);
fs/xfs/scrub/attr_repair.c
331
.geo = rx->sc->mp->m_attr_geo,
fs/xfs/scrub/attr_repair.c
394
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/attr_repair.c
408
bitmap_zero(ab->usedmap, mp->m_attr_geo->blksize);
fs/xfs/scrub/attr_repair.c
412
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &leafhdr, leaf);
fs/xfs/scrub/attr_repair.c
417
buf_end = (char *)bp->b_addr + mp->m_attr_geo->blksize;
fs/xfs/scrub/attr_repair.c
431
nameidx >= mp->m_attr_geo->blksize)
fs/xfs/scrub/attr_repair.c
519
struct xfs_mount *mp,
fs/xfs/scrub/attr_repair.c
526
.daddr = XFS_FSB_TO_DADDR(mp, fsbno),
fs/xfs/scrub/attr_repair.c
527
.max_sectors = xrep_bufscan_max_sectors(mp, max_len),
fs/xfs/scrub/attr_repair.c
528
.daddr_step = XFS_FSB_TO_BB(mp, 1),
fs/xfs/scrub/attr_repair.c
532
while ((bp = xrep_bufscan_advance(mp, &scan)) != NULL) {
fs/xfs/scrub/attr_repair.c
542
return xfs_buf_read(mp->m_ddev_targp, scan.daddr, XFS_FSB_TO_BB(mp, 1),
fs/xfs/scrub/attr_repair.c
578
error = xrep_xattr_find_buf(rx->sc->mp, fsbno, max_len, true, &bp);
fs/xfs/scrub/attr_repair.c
582
*actual_len = XFS_BB_TO_FSB(rx->sc->mp, bp->b_length);
fs/xfs/scrub/attr_repair.c
621
.geo = rx->sc->mp->m_attr_geo,
fs/xfs/scrub/attr_repair.c
783
if (!xfs_has_parent(rx->sc->mp))
fs/xfs/scrub/attr_repair.c
886
struct xfs_da_geometry *geo = sc->mp->m_attr_geo;
fs/xfs/scrub/attr_repair.c
983
xfs_emerg(sc->mp,
fs/xfs/scrub/attr_repair.c
987
xfs_err(sc->mp,
fs/xfs/scrub/bmap.c
1024
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/bmap.c
1042
if (!xfs_has_reflink(mp) && !xfs_has_zoned(mp)) {
fs/xfs/scrub/bmap.c
1054
if (!xfs_has_attr(mp))
fs/xfs/scrub/bmap.c
346
if (!xfs_has_rtrmapbt(info->sc->mp))
fs/xfs/scrub/bmap.c
349
rgbno = xfs_rtb_to_rgbno(info->sc->mp, irec->br_startblock);
fs/xfs/scrub/bmap.c
389
struct xfs_mount *mp = info->sc->mp;
fs/xfs/scrub/bmap.c
395
agno = XFS_FSB_TO_AGNO(mp, irec->br_startblock);
fs/xfs/scrub/bmap.c
396
agbno = XFS_FSB_TO_AGBNO(mp, irec->br_startblock);
fs/xfs/scrub/bmap.c
456
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/bmap.c
462
if (!xfs_verify_dablk(mp, irec->br_startoff))
fs/xfs/scrub/bmap.c
467
if (!xfs_verify_dablk(mp, off))
fs/xfs/scrub/bmap.c
478
struct xfs_mount *mp = info->sc->mp;
fs/xfs/scrub/bmap.c
489
if (!xfs_verify_fileext(mp, irec->br_startoff, irec->br_blockcount))
fs/xfs/scrub/bmap.c
497
!xfs_verify_rtbext(mp, irec->br_startblock, irec->br_blockcount))
fs/xfs/scrub/bmap.c
501
!xfs_verify_fsbext(mp, irec->br_startblock, irec->br_blockcount))
fs/xfs/scrub/bmap.c
589
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/bmap.c
602
cur = xfs_bmbt_init_cursor(mp, sc->tp, ip, whichfork);
fs/xfs/scrub/bmap.c
705
cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, agf, pag);
fs/xfs/scrub/bmap.c
824
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/bmap.c
849
while ((rtg = xfs_rtgroup_next(sc->mp, rtg))) {
fs/xfs/scrub/bmap.c
861
while ((pag = xfs_perag_next(sc->mp, pag))) {
fs/xfs/scrub/bmap.c
880
struct xfs_mount *mp = info->sc->mp;
fs/xfs/scrub/bmap.c
891
if (!xfs_verify_fileext(mp, irec->br_startoff, irec->br_blockcount))
fs/xfs/scrub/bmap.c
922
struct xfs_mount *mp = info->sc->mp;
fs/xfs/scrub/bmap.c
942
if (info->is_rt && xfs_has_rtgroups(mp)) {
fs/xfs/scrub/bmap.c
943
if (xfs_rtb_to_rgno(mp, b1->br_startblock) !=
fs/xfs/scrub/bmap.c
944
xfs_rtb_to_rgno(mp, b2->br_startblock))
fs/xfs/scrub/bmap_repair.c
111
agbno = xfs_rtb_to_rgbno(sc->mp, startblock);
fs/xfs/scrub/bmap_repair.c
114
agbno = XFS_FSB_TO_AGBNO(sc->mp, startblock);
fs/xfs/scrub/bmap_repair.c
214
!xfs_verify_fileext(sc->mp, rec->rm_offset, rec->rm_blockcount))
fs/xfs/scrub/bmap_repair.c
392
if (!xfs_verify_fileext(sc->mp, rec->rm_offset, rec->rm_blockcount))
fs/xfs/scrub/bmap_repair.c
456
if (!xfs_has_rtrmapbt(sc->mp))
fs/xfs/scrub/bmap_repair.c
537
while ((rtg = xfs_rtgroup_next(sc->mp, rtg))) {
fs/xfs/scrub/bmap_repair.c
547
while ((pag = xfs_perag_next(sc->mp, pag))) {
fs/xfs/scrub/bmap_repair.c
777
bmap_cur = xfs_bmbt_init_cursor(sc->mp, NULL, sc->ip, XFS_STAGING_FORK);
fs/xfs/scrub/bmap_repair.c
850
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/bmap_repair.c
900
if (!xfs_has_reflink(sc->mp))
fs/xfs/scrub/bmap_repair.c
945
large_extcount = xfs_has_large_extent_counts(sc->mp);
fs/xfs/scrub/common.c
1018
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, inum));
fs/xfs/scrub/common.c
1024
error = xfs_iget(mp, tp, inum, XFS_IGET_NORETRY | XCHK_IGET_FLAGS, 0,
fs/xfs/scrub/common.c
1122
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.c
1127
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, sc->sm->sm_ino);
fs/xfs/scrub/common.c
1142
if (!xfs_has_metadir(mp) && xfs_is_sb_inum(mp, sc->sm->sm_ino))
fs/xfs/scrub/common.c
1145
if (!xfs_verify_ino(sc->mp, sc->sm->sm_ino))
fs/xfs/scrub/common.c
1210
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, sc->sm->sm_ino));
fs/xfs/scrub/common.c
1227
trace_xchk_op_error(sc, agno, XFS_INO_TO_AGBNO(mp, sc->sm->sm_ino),
fs/xfs/scrub/common.c
1447
if (xfs_has_reflink(sc->mp)) {
fs/xfs/scrub/common.c
1462
if (!xfs_has_metadir(sc->mp)) {
fs/xfs/scrub/common.c
1528
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.c
1548
if (!xfs_verify_ino(mp, ino))
fs/xfs/scrub/common.c
1662
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/common.c
1664
return ip == mp->m_rootip ||
fs/xfs/scrub/common.c
1665
(xfs_has_metadir(mp) && ip == mp->m_metadirip);
fs/xfs/scrub/common.c
1680
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/common.c
1683
return mp->m_metadirip->i_ino;
fs/xfs/scrub/common.c
1684
return mp->m_rootip->i_ino;
fs/xfs/scrub/common.c
570
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.c
573
sa->pag = xfs_perag_get(mp, agno);
fs/xfs/scrub/common.c
612
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.c
616
sa->bno_cur = xfs_bnobt_init_cursor(mp, sc->tp, sa->agf_bp,
fs/xfs/scrub/common.c
622
sa->cnt_cur = xfs_cntbt_init_cursor(mp, sc->tp, sa->agf_bp,
fs/xfs/scrub/common.c
628
if (xfs_has_rmapbt(mp)) {
fs/xfs/scrub/common.c
629
sa->rmap_cur = xfs_rmapbt_init_cursor(mp, sc->tp,
fs/xfs/scrub/common.c
636
if (xfs_has_reflink(mp)) {
fs/xfs/scrub/common.c
637
sa->refc_cur = xfs_refcountbt_init_cursor(mp, sc->tp,
fs/xfs/scrub/common.c
652
if (xfs_has_finobt(mp)) {
fs/xfs/scrub/common.c
723
sr->rtg = xfs_rtgroup_get(sc->mp, rgno);
fs/xfs/scrub/common.c
802
if (xfs_has_rtrmapbt(sc->mp) && (rtglock_flags & XFS_RTGLOCK_RMAP))
fs/xfs/scrub/common.c
805
if (xfs_has_rtreflink(sc->mp) && (rtglock_flags & XFS_RTGLOCK_REFCOUNT))
fs/xfs/scrub/common.c
877
sc->tp = xfs_trans_alloc_empty(sc->mp);
fs/xfs/scrub/common.c
896
return xfs_trans_alloc(sc->mp, &M_RES(sc->mp)->tr_itruncate,
fs/xfs/scrub/common.c
928
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.c
938
error = xchk_checkpoint_log(mp);
fs/xfs/scrub/common.c
953
struct xfs_mount *mp)
fs/xfs/scrub/common.c
957
error = xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/scrub/common.c
960
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/scrub/common.c
973
return xfs_iget(sc->mp, sc->tp, inum, XCHK_IGET_FLAGS, 0, ipp);
fs/xfs/scrub/common.c
999
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/common.h
52
int xchk_checkpoint_log(struct xfs_mount *mp);
fs/xfs/scrub/cow_repair.c
250
agno = XFS_FSB_TO_AGNO(sc->mp, xc->irec.br_startblock);
fs/xfs/scrub/cow_repair.c
251
xc->irec_startbno = XFS_FSB_TO_AGBNO(sc->mp, xc->irec.br_startblock);
fs/xfs/scrub/cow_repair.c
253
pag = xfs_perag_get(sc->mp, agno);
fs/xfs/scrub/cow_repair.c
303
XFS_TEST_ERROR(sc->mp, XFS_ERRTAG_FORCE_SCRUB_REPAIR)) {
fs/xfs/scrub/cow_repair.c
333
xc->irec_startbno = xfs_rtb_to_rgbno(sc->mp, xc->irec.br_startblock);
fs/xfs/scrub/cow_repair.c
335
rtg = xfs_rtgroup_get(sc->mp,
fs/xfs/scrub/cow_repair.c
336
xfs_rtb_to_rgno(sc->mp, xc->irec.br_startblock));
fs/xfs/scrub/cow_repair.c
388
XFS_TEST_ERROR(sc->mp, XFS_ERRTAG_FORCE_SCRUB_REPAIR)) {
fs/xfs/scrub/cow_repair.c
415
.mp = sc->mp,
fs/xfs/scrub/cow_repair.c
430
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/cow_repair.c
453
xfs_rtxlen_t maxrtx = xfs_rtb_to_rtx(sc->mp, maxlen);
fs/xfs/scrub/cow_repair.c
660
if (!xfs_has_rmapbt(sc->mp) || !xfs_has_reflink(sc->mp))
fs/xfs/scrub/dabtree.c
109
xfs_da3_node_hdr_from_disk(ds->sc->mp, &hdr, blk->bp->b_addr);
fs/xfs/scrub/dabtree.c
390
if (xfs_has_crc(ds->sc->mp) && hdr3->hdr.pad)
fs/xfs/scrub/dabtree.c
503
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/dabtree.c
527
ds->dargs.geo = mp->m_attr_geo;
fs/xfs/scrub/dabtree.c
531
ds->dargs.geo = mp->m_dir_geo;
fs/xfs/scrub/dir.c
1104
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/dir.c
1127
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/dir.c
230
struct xfs_mount *mp = dp->i_mount;
fs/xfs/scrub/dir.c
237
offset = xfs_dir2_db_to_da(mp->m_dir_geo,
fs/xfs/scrub/dir.c
238
xfs_dir2_dataptr_to_db(mp->m_dir_geo, dapos));
fs/xfs/scrub/dir.c
244
if (!xfs_verify_dir_ino(mp, ino)) {
fs/xfs/scrub/dir.c
300
if (xfs_has_parent(mp)) {
fs/xfs/scrub/dir.c
322
struct xfs_mount *mp = ds->state->mp;
fs/xfs/scrub/dir.c
324
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/scrub/dir.c
344
xfs_dir2_leaf_hdr_from_disk(mp, &hdr, blk->bp->b_addr);
fs/xfs/scrub/dir.c
404
iter_off += xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/scrub/dir.c
410
tag = be16_to_cpup(xfs_dir2_data_entry_tag_p(mp, dent));
fs/xfs/scrub/dir.c
411
if (!xfs_verify_dir_ino(mp, ino) || tag != off)
fs/xfs/scrub/dir.c
421
calc_hash = xfs_dir2_hashname(mp, &dname);
fs/xfs/scrub/dir.c
471
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/dir.c
483
if (lblk != XFS_B_TO_FSBT(mp, XFS_DIR2_DATA_OFFSET))
fs/xfs/scrub/dir.c
501
bf = xfs_dir2_data_bestfree_p(mp, bp->b_addr);
fs/xfs/scrub/dir.c
507
if (offset >= mp->m_dir_geo->blksize) {
fs/xfs/scrub/dir.c
533
offset = mp->m_dir_geo->data_entry_offset;
fs/xfs/scrub/dir.c
534
end = xfs_dir3_data_end_offset(mp->m_dir_geo, bp->b_addr);
fs/xfs/scrub/dir.c
544
newlen = xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/scrub/dir.c
608
dfp = xfs_dir2_data_bestfree_p(sc->mp, dbp->b_addr);
fs/xfs/scrub/dir.c
630
struct xfs_da_geometry *geo = sc->mp->m_dir_geo;
fs/xfs/scrub/dir.c
652
if (xfs_has_crc(sc->mp)) {
fs/xfs/scrub/dir.c
752
if (xfs_has_crc(sc->mp)) {
fs/xfs/scrub/dir.c
793
.geo = sc->mp->m_dir_geo,
fs/xfs/scrub/dir.c
798
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/dir.c
814
lblk = XFS_B_TO_FSB(mp, XFS_DIR2_DATA_OFFSET);
fs/xfs/scrub/dir.c
815
leaf_lblk = XFS_B_TO_FSB(mp, XFS_DIR2_LEAF_OFFSET);
fs/xfs/scrub/dir.c
816
free_lblk = XFS_B_TO_FSB(mp, XFS_DIR2_FREE_OFFSET);
fs/xfs/scrub/dir.c
93
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/dir.c
95
if (!xfs_has_ftype(mp)) {
fs/xfs/scrub/dir_repair.c
1123
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/dir_repair.c
1158
if (!xfs_verify_dir_ino(rd->sc->mp, ino))
fs/xfs/scrub/dir_repair.c
1467
.geo = sc->mp->m_dir_geo,
fs/xfs/scrub/dir_repair.c
1526
error = xfs_dir_ino_validate(sc->mp, inum);
fs/xfs/scrub/dir_repair.c
1579
pag = xfs_perag_get(sc->mp, XFS_INO_TO_AGNO(sc->mp, dp->i_ino));
fs/xfs/scrub/dir_repair.c
1608
if (!xfs_has_parent(sc->mp))
fs/xfs/scrub/dir_repair.c
1654
rd->pscan.parent_ino = rd->sc->mp->m_sb.sb_rootino;
fs/xfs/scrub/dir_repair.c
1767
error = xrep_dir_reset_fork(rd, sc->mp->m_rootip->i_ino);
fs/xfs/scrub/dir_repair.c
1803
if (xfs_has_parent(sc->mp))
fs/xfs/scrub/dir_repair.c
1926
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/dir_repair.c
1929
if (!xfs_has_exchange_range(sc->mp))
fs/xfs/scrub/dir_repair.c
1936
if (xfs_has_parent(sc->mp))
fs/xfs/scrub/dir_repair.c
226
if (!xfs_verify_dir_ino(sc->mp, ino))
fs/xfs/scrub/dir_repair.c
304
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
309
if (!xfs_verify_dir_ino(mp, ino))
fs/xfs/scrub/dir_repair.c
443
ino = xfs_dir2_sf_get_ino(rd->sc->mp, sfp, sfep);
fs/xfs/scrub/dir_repair.c
471
struct xfs_da_geometry *geo = rd->sc->mp->m_dir_geo;
fs/xfs/scrub/dir_repair.c
498
offset += xfs_dir2_data_entsize(rd->sc->mp, dep->namelen);
fs/xfs/scrub/dir_repair.c
537
next = xfs_dir2_sf_nextentry(rd->sc->mp, hdr, sfep);
fs/xfs/scrub/dir_repair.c
566
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
567
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/scrub/dir_repair.c
571
ASSERT(xfs_has_crc(mp));
fs/xfs/scrub/dir_repair.c
580
if (!error && XFS_FSB_TO_B(mp, last) == geo->blksize &&
fs/xfs/scrub/dir_repair.c
594
XFS_FSB_TO_B(mp, last) > geo->blksize &&
fs/xfs/scrub/dir_repair.c
595
XFS_FSB_TO_B(mp, last) == dp->i_disk_size) {
fs/xfs/scrub/dir_repair.c
677
rd->args.geo = rd->sc->mp->m_dir_geo;
fs/xfs/scrub/dir_repair.c
687
rd->args.hashval = xfs_dir2_hashname(rd->sc->mp, name);
fs/xfs/scrub/dir_repair.c
704
error = xfs_dir_ino_validate(sc->mp, inum);
fs/xfs/scrub/dir_repair.c
747
struct xfs_mount *mp = rd->sc->mp;
fs/xfs/scrub/dir_repair.c
754
resblks = xfs_link_space_res(mp, xname->len);
fs/xfs/scrub/dir_repair.c
961
struct xfs_da_geometry *geo = sc->mp->m_dir_geo;
fs/xfs/scrub/dirtree.c
223
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/dirtree.c
323
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/dirtree.c
76
xfs_dir_hook_del(dl->sc->mp, &dl->dhook);
fs/xfs/scrub/dirtree.c
923
ASSERT(xfs_has_parent(sc->mp));
fs/xfs/scrub/dirtree.c
943
error = xfs_dir_hook_add(sc->mp, &dl->dhook);
fs/xfs/scrub/dirtree_repair.c
385
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/dirtree_repair.c
402
resblks = xfs_remove_space_res(mp, step->name_len);
fs/xfs/scrub/dirtree_repair.c
403
error = xfs_trans_alloc_dir(dp, &M_RES(mp)->tr_remove, sc->ip,
fs/xfs/scrub/dirtree_repair.c
469
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/dqiterate.c
148
struct xfs_quotainfo *qi = cursor->sc->mp->m_quotainfo;
fs/xfs/scrub/dqiterate.c
175
struct xfs_mount *mp = cursor->sc->mp;
fs/xfs/scrub/dqiterate.c
204
error = xfs_qm_dqget(mp, cursor->id, cursor->dqtype, false, &dq);
fs/xfs/scrub/dqiterate.c
34
cursor->quota_ip = xfs_quota_inode(sc->mp, cursor->dqtype);
fs/xfs/scrub/dqiterate.c
46
struct xfs_quotainfo *qi = cursor->sc->mp->m_quotainfo;
fs/xfs/scrub/dqiterate.c
90
struct xfs_quotainfo *qi = cursor->sc->mp->m_quotainfo;
fs/xfs/scrub/findparent.c
276
error = xfs_dir_hook_add(sc->mp, &pscan->dhook);
fs/xfs/scrub/findparent.c
334
xfs_dir_hook_del(pscan->sc->mp, &pscan->dhook);
fs/xfs/scrub/findparent.c
370
if (sc->ip == sc->mp->m_rootip) {
fs/xfs/scrub/findparent.c
371
*parent_ino = sc->mp->m_sb.sb_rootino;
fs/xfs/scrub/findparent.c
376
if (sc->ip == sc->mp->m_metadirip) {
fs/xfs/scrub/findparent.c
377
*parent_ino = sc->mp->m_sb.sb_metadirino;
fs/xfs/scrub/findparent.c
390
if (!xfs_verify_dir_ino(sc->mp, *parent_ino) ||
fs/xfs/scrub/findparent.c
425
if (sc->ip->i_ino == sc->mp->m_sb.sb_rootino)
fs/xfs/scrub/findparent.c
426
return sc->mp->m_sb.sb_rootino;
fs/xfs/scrub/findparent.c
428
if (sc->ip->i_ino == sc->mp->m_sb.sb_metadirino)
fs/xfs/scrub/findparent.c
429
return sc->mp->m_sb.sb_metadirino;
fs/xfs/scrub/fscounters.c
126
error = freeze_super(sc->mp->m_super, FREEZE_HOLDER_KERNEL, NULL);
fs/xfs/scrub/fscounters.c
138
error = thaw_super(sc->mp->m_super, FREEZE_HOLDER_KERNEL, NULL);
fs/xfs/scrub/fscounters.c
191
xfs_emerg(sc->mp, "still frozen after scrub, err=%d", error);
fs/xfs/scrub/fscounters.c
207
if (!xfs_has_lazysbcount(sc->mp))
fs/xfs/scrub/fscounters.c
217
xfs_icount_range(sc->mp, &fsc->icount_min, &fsc->icount_max);
fs/xfs/scrub/fscounters.c
298
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/fscounters.c
309
while ((pag = xfs_perag_next(mp, pag))) {
fs/xfs/scrub/fscounters.c
327
if (xfs_has_lazysbcount(sc->mp)) {
fs/xfs/scrub/fscounters.c
354
fsc->fdblocks -= mp->m_free[XC_FREE_BLOCKS].res_avail;
fs/xfs/scrub/fscounters.c
361
delayed = percpu_counter_sum(&mp->m_delalloc_blks);
fs/xfs/scrub/fscounters.c
364
trace_xchk_fscounters_calc(mp, fsc->icount, fsc->ifree, fsc->fdblocks,
fs/xfs/scrub/fscounters.c
370
fsc->fdblocks > mp->m_sb.sb_dblocks ||
fs/xfs/scrub/fscounters.c
411
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/fscounters.c
423
if (!xfs_has_realtime(mp) || xfs_has_zoned(mp))
fs/xfs/scrub/fscounters.c
426
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/scrub/fscounters.c
438
fsc->frextents_delayed = percpu_counter_sum(&mp->m_delalloc_rtextents);
fs/xfs/scrub/fscounters.c
483
trace_xchk_fscounters_within_range(sc->mp, expected, curr_value,
fs/xfs/scrub/fscounters.c
514
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/fscounters.c
521
icount = percpu_counter_sum(&mp->m_icount);
fs/xfs/scrub/fscounters.c
522
ifree = percpu_counter_sum(&mp->m_ifree);
fs/xfs/scrub/fscounters.c
523
fdblocks = xfs_sum_freecounter_raw(mp, XC_FREE_BLOCKS);
fs/xfs/scrub/fscounters.c
524
frextents = xfs_sum_freecounter_raw(mp, XC_FREE_RTEXTENTS);
fs/xfs/scrub/fscounters.c
552
if (fdblocks > mp->m_sb.sb_dblocks)
fs/xfs/scrub/fscounters.c
556
if (frextents > mp->m_sb.sb_rextents)
fs/xfs/scrub/fscounters.c
568
if (!xchk_process_error(sc, 0, XFS_SB_BLOCK(mp), &error))
fs/xfs/scrub/fscounters.c
573
if (!xchk_process_error(sc, 0, XFS_SB_BLOCK(mp), &error))
fs/xfs/scrub/fscounters.c
584
if (!xchk_fscount_within_range(sc, icount, &mp->m_icount,
fs/xfs/scrub/fscounters.c
592
if (!xchk_fscount_within_range(sc, ifree, &mp->m_ifree, fsc->ifree)) {
fs/xfs/scrub/fscounters.c
600
&mp->m_free[XC_FREE_BLOCKS].count, fsc->fdblocks)) {
fs/xfs/scrub/fscounters.c
607
if (!xfs_has_zoned(mp) &&
fs/xfs/scrub/fscounters.c
609
&mp->m_free[XC_FREE_RTEXTENTS].count,
fs/xfs/scrub/fscounters.c
74
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/fscounters.c
80
while ((pag = xfs_perag_next(mp, pag))) {
fs/xfs/scrub/fscounters_repair.c
50
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/fscounters_repair.c
63
trace_xrep_reset_counters(mp, fsc);
fs/xfs/scrub/fscounters_repair.c
65
percpu_counter_set(&mp->m_icount, fsc->icount);
fs/xfs/scrub/fscounters_repair.c
66
percpu_counter_set(&mp->m_ifree, fsc->ifree);
fs/xfs/scrub/fscounters_repair.c
67
xfs_set_freecounter(mp, XC_FREE_BLOCKS, fsc->fdblocks);
fs/xfs/scrub/fscounters_repair.c
77
if (!xfs_has_zoned(mp)) {
fs/xfs/scrub/fscounters_repair.c
78
xfs_set_freecounter(mp, XC_FREE_RTEXTENTS,
fs/xfs/scrub/fscounters_repair.c
80
if (!xfs_has_rtgroups(mp))
fs/xfs/scrub/fscounters_repair.c
81
mp->m_sb.sb_frextents = fsc->frextents;
fs/xfs/scrub/health.c
168
struct xfs_mount *mp)
fs/xfs/scrub/health.c
173
xfs_fs_mark_healthy(mp, XFS_SICK_FS_INDIRECT);
fs/xfs/scrub/health.c
174
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/scrub/health.c
176
while ((rtg = xfs_rtgroup_next(mp, rtg)))
fs/xfs/scrub/health.c
207
xchk_mark_all_healthy(sc->mp);
fs/xfs/scrub/health.c
221
pag = xfs_perag_get(sc->mp, sc->sm->sm_agno);
fs/xfs/scrub/health.c
249
xfs_fs_mark_corrupt(sc->mp, mask);
fs/xfs/scrub/health.c
251
xfs_fs_mark_healthy(sc->mp, mask);
fs/xfs/scrub/health.c
256
rtg = xfs_rtgroup_get(sc->mp, sc->sm->sm_agno);
fs/xfs/scrub/health.c
314
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/health.c
320
xfs_fs_measure_sickness(mp, &sick, &checked);
fs/xfs/scrub/health.c
324
while ((pag = xfs_perag_next(mp, pag))) {
fs/xfs/scrub/health.c
330
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/scrub/ialloc.c
260
struct xfs_mount *mp = bs->cur->bc_mp;
fs/xfs/scrub/ialloc.c
265
agbno = XFS_AGINO_TO_AGBNO(mp, agino);
fs/xfs/scrub/ialloc.c
266
len = XFS_B_TO_FSB(mp, nr_inodes * mp->m_sb.sb_inodesize);
fs/xfs/scrub/ialloc.c
367
struct xfs_mount *mp = bs->cur->bc_mp;
fs/xfs/scrub/ialloc.c
377
M_IGEO(mp)->inodes_per_cluster);
fs/xfs/scrub/ialloc.c
380
agbno = XFS_AGINO_TO_AGBNO(mp, irec->ir_startino + cluster_base);
fs/xfs/scrub/ialloc.c
398
imap.im_len = XFS_FSB_TO_BB(mp, M_IGEO(mp)->blocks_per_cluster);
fs/xfs/scrub/ialloc.c
399
imap.im_boffset = XFS_INO_TO_OFFSET(mp, irec->ir_startino) <<
fs/xfs/scrub/ialloc.c
400
mp->m_sb.sb_inodelog;
fs/xfs/scrub/ialloc.c
411
XFS_INO_TO_OFFSET(mp, irec->ir_startino +
fs/xfs/scrub/ialloc.c
423
M_IGEO(mp)->blocks_per_cluster,
fs/xfs/scrub/ialloc.c
428
xchk_xref_is_only_owned_by(bs->sc, agbno, M_IGEO(mp)->blocks_per_cluster,
fs/xfs/scrub/ialloc.c
432
error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap, &cluster_bp);
fs/xfs/scrub/ialloc.c
450
imap.im_boffset += mp->m_sb.sb_inodesize;
fs/xfs/scrub/ialloc.c
479
cluster_base += M_IGEO(bs->sc->mp)->inodes_per_cluster) {
fs/xfs/scrub/ialloc.c
499
struct xfs_mount *mp = bs->sc->mp;
fs/xfs/scrub/ialloc.c
501
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/scrub/ialloc.c
575
struct xfs_mount *mp = bs->cur->bc_mp;
fs/xfs/scrub/ialloc.c
585
xfs_inobt_btrec_to_irec(mp, rec, &irec);
fs/xfs/scrub/ialloc.c
658
(xfs_has_finobt(sc->mp) && !sc->sa.fino_cur) ||
fs/xfs/scrub/ialloc.c
702
inode_blocks = XFS_B_TO_FSB(sc->mp, inodes * sc->mp->m_sb.sb_inodesize);
fs/xfs/scrub/ialloc_repair.c
147
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/ialloc_repair.c
157
cluster_buf_base = XFS_INO_TO_OFFSET(mp, cluster_ag_base);
fs/xfs/scrub/ialloc_repair.c
158
offset = (cluster_buf_base + cluster_index) * mp->m_sb.sb_inodesize;
fs/xfs/scrub/ialloc_repair.c
225
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/ialloc_repair.c
232
if (xfs_has_sparseinodes(mp))
fs/xfs/scrub/ialloc_repair.c
293
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/ialloc_repair.c
294
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/scrub/ialloc_repair.c
309
imap.im_len = XFS_FSB_TO_BB(mp, igeo->blocks_per_cluster);
fs/xfs/scrub/ialloc_repair.c
311
error = xfs_imap_to_bp(mp, sc->tp, &imap, &cluster_bp);
fs/xfs/scrub/ialloc_repair.c
319
cluster_ag_base = XFS_AGB_TO_AGINO(mp, cluster_bno);
fs/xfs/scrub/ialloc_repair.c
342
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/ialloc_repair.c
343
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/scrub/ialloc_repair.c
361
if (!xfs_has_sparseinodes(mp) &&
fs/xfs/scrub/ialloc_repair.c
370
if (xfs_has_sparseinodes(mp) && mp->m_sb.sb_spino_align &&
fs/xfs/scrub/ialloc_repair.c
371
(!IS_ALIGNED(agbno, mp->m_sb.sb_spino_align) ||
fs/xfs/scrub/ialloc_repair.c
372
!IS_ALIGNED(agbno + len, mp->m_sb.sb_spino_align)))
fs/xfs/scrub/ialloc_repair.c
376
agino = XFS_AGB_TO_AGINO(mp, agbno);
fs/xfs/scrub/ialloc_repair.c
380
agino = XFS_AGB_TO_AGINO(mp, agbno + len) - 1;
fs/xfs/scrub/ialloc_repair.c
414
struct xfs_mount *mp = ri->sc->mp;
fs/xfs/scrub/ialloc_repair.c
415
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/scrub/ialloc_repair.c
502
xfs_force_summary_recalc(sc->mp);
fs/xfs/scrub/ialloc_repair.c
636
need_finobt = xfs_has_finobt(sc->mp);
fs/xfs/scrub/ialloc_repair.c
655
xfs_agbno_to_fsb(sc->sa.pag, XFS_IBT_BLOCK(sc->mp)),
fs/xfs/scrub/ialloc_repair.c
671
if (sc->mp->m_finobt_nores)
fs/xfs/scrub/ialloc_repair.c
675
xfs_agbno_to_fsb(sc->sa.pag, XFS_FIBT_BLOCK(sc->mp)),
fs/xfs/scrub/ialloc_repair.c
787
if (xfs_has_finobt(sc->mp) && !sc->mp->m_finobt_nores)
fs/xfs/scrub/ialloc_repair.c
799
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/ialloc_repair.c
804
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/ialloc_repair.c
816
xfs_agino_range(mp, pag_agno(sc->sa.pag), &first_agino, &last_agino);
fs/xfs/scrub/ialloc_repair.c
874
if (!xfs_has_finobt(sc->mp) ||
fs/xfs/scrub/inode.c
119
if (!xfs_has_metadir(mp) && xfs_is_sb_inum(mp, sc->sm->sm_ino))
fs/xfs/scrub/inode.c
122
if (!xfs_verify_ino(sc->mp, sc->sm->sm_ino))
fs/xfs/scrub/inode.c
189
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, sc->sm->sm_ino));
fs/xfs/scrub/inode.c
221
trace_xchk_op_error(sc, agno, XFS_INO_TO_AGBNO(mp, sc->sm->sm_ino),
fs/xfs/scrub/inode.c
244
fa = xfs_inode_validate_extsize(sc->mp, value, mode, flags);
fs/xfs/scrub/inode.c
259
xfs_extlen_to_rtxmod(sc->mp, value) > 0)
fs/xfs/scrub/inode.c
279
if (xfs_has_zoned(sc->mp) &&
fs/xfs/scrub/inode.c
283
fa = xfs_inode_validate_cowextsize(sc->mp, value, mode, flags, flags2);
fs/xfs/scrub/inode.c
298
value % sc->mp->m_sb.sb_rextsize > 0)
fs/xfs/scrub/inode.c
311
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode.c
318
if ((flags & XFS_DIFLAG_REALTIME) && !mp->m_rtdev_targp)
fs/xfs/scrub/inode.c
322
if ((flags & XFS_DIFLAG_NEWRTBM) && ino != mp->m_sb.sb_rbmino)
fs/xfs/scrub/inode.c
357
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode.c
365
!xfs_has_reflink(mp))
fs/xfs/scrub/inode.c
380
!xfs_has_rtreflink(mp))
fs/xfs/scrub/inode.c
384
if (xfs_dinode_has_bigtime(dip) && !xfs_has_bigtime(mp))
fs/xfs/scrub/inode.c
388
if ((flags2 & XFS_DIFLAG2_NREXT64) && !xfs_has_large_extent_counts(mp))
fs/xfs/scrub/inode.c
417
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode.c
474
!xfs_has_projid32(mp))
fs/xfs/scrub/inode.c
484
if (xfs_has_projid32(mp))
fs/xfs/scrub/inode.c
563
if (isize > mp->m_super->s_maxbytes)
fs/xfs/scrub/inode.c
578
mp->m_sb.sb_dblocks + mp->m_sb.sb_rblocks)
fs/xfs/scrub/inode.c
581
if (be64_to_cpu(dip->di_nblocks) >= mp->m_sb.sb_dblocks)
fs/xfs/scrub/inode.c
593
fork_recs = XFS_DFORK_DSIZE(dip, mp) / sizeof(struct xfs_bmbt_rec);
fs/xfs/scrub/inode.c
610
if (XFS_DFORK_BOFF(dip) >= mp->m_sb.sb_inodesize)
fs/xfs/scrub/inode.c
624
fork_recs = XFS_DFORK_ASIZE(dip, mp) / sizeof(struct xfs_bmbt_rec);
fs/xfs/scrub/inode.c
665
agino = XFS_INO_TO_AGINO(sc->mp, ino);
fs/xfs/scrub/inode.c
740
agno = XFS_INO_TO_AGNO(sc->mp, ino);
fs/xfs/scrub/inode.c
741
agbno = XFS_INO_TO_AGBNO(sc->mp, ino);
fs/xfs/scrub/inode.c
769
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode.c
773
if (!xfs_has_reflink(mp))
fs/xfs/scrub/inode.c
778
if (!xchk_xref_process_error(sc, XFS_INO_TO_AGNO(mp, ino),
fs/xfs/scrub/inode.c
779
XFS_INO_TO_AGBNO(mp, ino), &error))
fs/xfs/scrub/inode.c
94
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode.c
98
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, sc->sm->sm_ino);
fs/xfs/scrub/inode_repair.c
1086
dfork_size = XFS_DFORK_SIZE(dip, sc->mp, XFS_DATA_FORK);
fs/xfs/scrub/inode_repair.c
1098
xfs_dir2_sf_verify(sc->mp, dfork_ptr, data_size) != NULL)
fs/xfs/scrub/inode_repair.c
1214
afork_size = XFS_DFORK_SIZE(dip, sc->mp, XFS_ATTR_FORK);
fs/xfs/scrub/inode_repair.c
1297
unsigned int lit_sz = XFS_LITINO(sc->mp);
fs/xfs/scrub/inode_repair.c
1319
afork_min = XFS_DFORK_SIZE(dip, sc->mp, XFS_ATTR_FORK);
fs/xfs/scrub/inode_repair.c
1348
afork_min = xfs_bmap_broot_space(sc->mp, bmdr);
fs/xfs/scrub/inode_repair.c
1398
dfork_min = xfs_bmap_broot_space(sc->mp, bmdr);
fs/xfs/scrub/inode_repair.c
1404
dfork_min = xfs_rtrmap_broot_space(sc->mp, rmdr);
fs/xfs/scrub/inode_repair.c
1408
dfork_min = xfs_rtrefcount_broot_space(sc->mp, rcdr);
fs/xfs/scrub/inode_repair.c
1438
bmdr_minsz > XFS_DFORK_DSIZE(dip, sc->mp)) {
fs/xfs/scrub/inode_repair.c
1453
memmove(after, before, XFS_DFORK_ASIZE(dip, sc->mp));
fs/xfs/scrub/inode_repair.c
1464
bmdr_minsz > XFS_DFORK_ASIZE(dip, sc->mp)) {
fs/xfs/scrub/inode_repair.c
1562
error = xfs_trans_read_buf(sc->mp, sc->tp, sc->mp->m_ddev_targp,
fs/xfs/scrub/inode_repair.c
1586
xfs_dinode_calc_crc(sc->mp, dip);
fs/xfs/scrub/inode_repair.c
1589
ri->imap.im_boffset + sc->mp->m_sb.sb_inodesize - 1);
fs/xfs/scrub/inode_repair.c
1643
if (XFS_IS_UQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
1645
if (XFS_IS_GQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
1647
if (XFS_IS_PQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
1683
if (count >= sc->mp->m_sb.sb_rblocks)
fs/xfs/scrub/inode_repair.c
1686
if (count >= sc->mp->m_sb.sb_dblocks)
fs/xfs/scrub/inode_repair.c
1701
if (count >= sc->mp->m_sb.sb_dblocks)
fs/xfs/scrub/inode_repair.c
1728
if (XFS_IS_UQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
1735
if (XFS_IS_GQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
1742
if (XFS_IS_PQUOTA_ON(sc->mp))
fs/xfs/scrub/inode_repair.c
176
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
1799
if (sc->ip->i_ino == sc->mp->m_sb.sb_rbmino)
fs/xfs/scrub/inode_repair.c
1830
if (!xfs_has_reflink(sc->mp) || !S_ISREG(mode))
fs/xfs/scrub/inode_repair.c
1861
off = XFS_B_TO_FSB(sc->mp, XFS_DIR2_SPACE_SIZE);
fs/xfs/scrub/inode_repair.c
1869
XFS_FSB_TO_B(sc->mp, off));
fs/xfs/scrub/inode_repair.c
188
xfs_dinode_good_version(mp, dip->di_version))
fs/xfs/scrub/inode_repair.c
191
if (xfs_verify_cksum((char *)dip, mp->m_sb.sb_inodesize,
fs/xfs/scrub/inode_repair.c
1914
xfs_extlen_to_rtxmod(sc->mp, sc->ip->i_extsize) > 0) {
fs/xfs/scrub/inode_repair.c
1925
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
1929
if (!xfs_has_parent(mp))
fs/xfs/scrub/inode_repair.c
1959
sc->ip->i_extsize % sc->mp->m_sb.sb_rextsize > 0) {
fs/xfs/scrub/inode_repair.c
2014
pag = xfs_perag_get(sc->mp,
fs/xfs/scrub/inode_repair.c
2015
XFS_INO_TO_AGNO(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/inode_repair.c
204
xfs_dinode_calc_crc(mp, dip);
fs/xfs/scrub/inode_repair.c
216
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
220
ni = XFS_BB_TO_FSB(mp, bp->b_length) * mp->m_sb.sb_inopblock;
fs/xfs/scrub/inode_repair.c
222
xrep_dinode_buf_core(sc, bp, i << mp->m_sb.sb_inodelog);
fs/xfs/scrub/inode_repair.c
234
if (!xfs_dinode_good_version(sc->mp, dip->di_version))
fs/xfs/scrub/inode_repair.c
237
uuid_copy(&dip->di_uuid, &sc->mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/inode_repair.c
402
if (!xfs_has_ftype(sc->mp)) {
fs/xfs/scrub/inode_repair.c
547
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
564
if (xfs_has_reflink(mp) && S_ISREG(mode))
fs/xfs/scrub/inode_repair.c
568
if (!xfs_has_bigtime(mp))
fs/xfs/scrub/inode_repair.c
570
if (!xfs_has_large_extent_counts(mp))
fs/xfs/scrub/inode_repair.c
580
fa = xfs_dinode_verify_metadir(sc->mp, dip, mode, flags,
fs/xfs/scrub/inode_repair.c
623
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
630
i8count = mp->m_sb.sb_rootino > XFS_DIR2_MAX_SHORT_INUM;
fs/xfs/scrub/inode_repair.c
634
xfs_dir2_sf_put_parent_ino(sfp, mp->m_sb.sb_rootino);
fs/xfs/scrub/inode_repair.c
696
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/inode_repair.c
705
fa = xfs_inode_validate_extsize(mp, be32_to_cpu(dip->di_extsize),
fs/xfs/scrub/inode_repair.c
714
(xfs_has_zoned(sc->mp) &&
fs/xfs/scrub/inode_repair.c
718
fa = xfs_inode_validate_cowextsize(mp, be32_to_cpu(dip->di_cowextsize),
fs/xfs/scrub/inode_repair.c
772
cur = xfs_rmapbt_init_cursor(ri->sc->mp, ri->sc->tp, agf, pag);
fs/xfs/scrub/inode_repair.c
833
if (!xfs_has_rmapbt(ri->sc->mp))
fs/xfs/scrub/inode_repair.c
836
while ((rtg = xfs_rtgroup_next(ri->sc->mp, rtg))) {
fs/xfs/scrub/inode_repair.c
844
while ((pag = xfs_perag_next(ri->sc->mp, pag))) {
fs/xfs/scrub/inode_repair.c
887
fa = xfs_bmap_validate_extent_raw(sc->mp, isrt, whichfork,
fs/xfs/scrub/inode_repair.c
924
if (level == 0 || level >= XFS_BM_MAXLEVELS(sc->mp, whichfork))
fs/xfs/scrub/inode_repair.c
936
if (!xfs_verify_fileoff(sc->mp, fileoff))
fs/xfs/scrub/inode_repair.c
941
if (!xfs_verify_fsbno(sc->mp, fsbno))
fs/xfs/scrub/inode_repair.c
966
if (level > sc->mp->m_rtrmap_maxlevels)
fs/xfs/scrub/inode_repair.c
994
if (level > sc->mp->m_rtrefc_maxlevels)
fs/xfs/scrub/iscan.c
106
xfs_agino_range(mp, agno, &first, &last);
fs/xfs/scrub/iscan.c
143
XFS_IS_CORRUPT(mp, lastino >= rec.ir_startino)) {
fs/xfs/scrub/iscan.c
205
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/iscan.c
215
cursor = XFS_AGINO_TO_INO(mp, agno, agino);
fs/xfs/scrub/iscan.c
311
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/iscan.c
324
agno = XFS_INO_TO_AGNO(mp, iscan->cursor_ino);
fs/xfs/scrub/iscan.c
325
pag = xfs_perag_get(mp, agno);
fs/xfs/scrub/iscan.c
333
agino = XFS_INO_TO_AGINO(mp, iscan->cursor_ino);
fs/xfs/scrub/iscan.c
355
agno = (agno + 1) % mp->m_sb.sb_agcount;
fs/xfs/scrub/iscan.c
437
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/iscan.c
446
error = xfs_iget(sc->mp, sc->tp, ino, ISCAN_IGET_FLAGS, 0,
fs/xfs/scrub/iscan.c
463
xfs_inodegc_push(mp);
fs/xfs/scrub/iscan.c
465
xfs_inodegc_flush(mp);
fs/xfs/scrub/iscan.c
519
error = xfs_iget(sc->mp, sc->tp, ino, ISCAN_IGET_FLAGS, 0,
fs/xfs/scrub/iscan.c
530
trace_xchk_iscan_iget_batch(sc->mp, iscan, nr_inodes, idx);
fs/xfs/scrub/iscan.c
66
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/iscan.c
67
xfs_agnumber_t skip_agno = XFS_INO_TO_AGNO(mp, iscan->skip_ino);
fs/xfs/scrub/iscan.c
673
struct xfs_mount *mp)
fs/xfs/scrub/iscan.c
68
xfs_agnumber_t skip_agino = XFS_INO_TO_AGINO(mp, iscan->skip_ino);
fs/xfs/scrub/iscan.c
682
r = (r % mp->m_sb.sb_agcount) + 1;
fs/xfs/scrub/iscan.c
684
return XFS_AGINO_TO_INO(mp, mp->m_sb.sb_agcount - r, 0);
fs/xfs/scrub/iscan.c
703
start_ino = xchk_iscan_rotor(sc->mp);
fs/xfs/scrub/iscan.c
96
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/listxattr.c
136
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/listxattr.c
173
xfs_da3_node_hdr_from_disk(mp, &nodehdr, node);
fs/xfs/scrub/listxattr.c
229
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/listxattr.c
250
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &leafhdr, leaf);
fs/xfs/scrub/listxattr.c
62
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/listxattr.c
68
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr, leaf);
fs/xfs/scrub/metapath.c
131
if (!sc->mp->m_rtdirip)
fs/xfs/scrub/metapath.c
134
return xchk_setup_metapath_scan(sc, sc->mp->m_metadirip,
fs/xfs/scrub/metapath.c
135
kstrdup_const("rtgroups", GFP_KERNEL), sc->mp->m_rtdirip);
fs/xfs/scrub/metapath.c
148
rtg = xfs_rtgroup_get(sc->mp, sc->sm->sm_agno);
fs/xfs/scrub/metapath.c
158
error = xchk_setup_metapath_scan(sc, sc->mp->m_rtdirip,
fs/xfs/scrub/metapath.c
176
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/metapath.c
181
return xchk_setup_metapath_scan(sc, sc->mp->m_metadirip,
fs/xfs/scrub/metapath.c
191
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/metapath.c
226
if (!xfs_has_metadir(sc->mp))
fs/xfs/scrub/metapath.c
362
if (xfs_has_parent(sc->mp))
fs/xfs/scrub/metapath.c
381
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/metapath.c
399
if (xfs_has_parent(mp)) {
fs/xfs/scrub/metapath.c
625
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/metapath.c
640
if (xfs_has_parent(mp)) {
fs/xfs/scrub/metapath.c
648
mpath->unlink_resblks = xfs_remove_space_res(mp, MAXNAMELEN);
fs/xfs/scrub/metapath.c
649
mpath->link_resblks = xfs_link_space_res(mp, MAXNAMELEN);
fs/xfs/scrub/newbt.c
108
xnr->bload.max_dirty = XFS_B_TO_FSBT(sc->mp, 256U << 10); /* 256K */
fs/xfs/scrub/newbt.c
127
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino),
fs/xfs/scrub/newbt.c
164
XFS_INO_TO_FSB(sc->mp, sc->ip->i_ino),
fs/xfs/scrub/newbt.c
194
struct xfs_mount *mp = xnr->sc->mp;
fs/xfs/scrub/newbt.c
203
resv->agbno = XFS_FSB_TO_AGBNO(mp, args->fsbno);
fs/xfs/scrub/newbt.c
254
xfs_agnumber_t agno = XFS_FSB_TO_AGNO(sc->mp, xnr->alloc_hint);
fs/xfs/scrub/newbt.c
257
xfs_verify_fsbno(sc->mp, xnr->alloc_hint))
fs/xfs/scrub/newbt.c
261
xfs_agbno_to_fsb(sc->sa.pag, XFS_AGFL_BLOCK(sc->mp) + 1);
fs/xfs/scrub/newbt.c
271
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/newbt.c
280
.mp = mp,
fs/xfs/scrub/newbt.c
301
agno = XFS_FSB_TO_AGNO(mp, args.fsbno);
fs/xfs/scrub/newbt.c
308
XFS_FSB_TO_AGBNO(mp, args.fsbno), args.len,
fs/xfs/scrub/newbt.c
333
if (xfs_verify_fsbno(sc->mp, xnr->alloc_hint))
fs/xfs/scrub/newbt.c
336
xnr->alloc_hint = XFS_AGB_TO_FSB(sc->mp, 0, XFS_AGFL_BLOCK(sc->mp) + 1);
fs/xfs/scrub/newbt.c
346
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/newbt.c
354
.mp = mp,
fs/xfs/scrub/newbt.c
376
agno = XFS_FSB_TO_AGNO(mp, args.fsbno);
fs/xfs/scrub/newbt.c
378
pag = xfs_perag_get(mp, agno);
fs/xfs/scrub/newbt.c
385
XFS_FSB_TO_AGBNO(mp, args.fsbno), args.len,
fs/xfs/scrub/newbt.c
488
if (xfs_is_shutdown(sc->mp))
fs/xfs/scrub/newbt.c
72
sz = xfs_ag_block_count(sc->mp, pag_agno(sc->sa.pag));
fs/xfs/scrub/newbt.c
74
free = xfs_sum_freecounter_raw(sc->mp, XC_FREE_BLOCKS);
fs/xfs/scrub/newbt.c
75
sz = sc->mp->m_sb.sb_dblocks;
fs/xfs/scrub/nlinks.c
1008
xfs_agino_range(mp, last_agno, &first_agino, &last_agino);
fs/xfs/scrub/nlinks.c
1009
max_inos = XFS_AGINO_TO_INO(mp, last_agno, last_agino) + 1;
fs/xfs/scrub/nlinks.c
1025
error = xfs_dir_hook_add(mp, &xnc->dhook);
fs/xfs/scrub/nlinks.c
135
trace_xchk_nlinks_update_incore(xnc->sc->mp, ino, &nl, parents_delta,
fs/xfs/scrub/nlinks.c
178
trace_xchk_nlinks_live_update(xnc->sc->mp, p->dp, action, p->ip->i_ino,
fs/xfs/scrub/nlinks.c
252
if (!xfs_verify_dir_ino(sc->mp, ino)) {
fs/xfs/scrub/nlinks.c
264
trace_xchk_nlinks_collect_dirent(sc->mp, dp, ino, name);
fs/xfs/scrub/nlinks.c
284
else if (!xfs_has_parent(sc->mp))
fs/xfs/scrub/nlinks.c
355
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/nlinks.c
360
trace_xchk_nlinks_collect_pptr(sc->mp, ip, &xname, pptr_rec);
fs/xfs/scrub/nlinks.c
455
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/nlinks.c
493
if (!xfs_verify_ino(xnc->sc->mp, ino))
fs/xfs/scrub/nlinks.c
496
trace_xchk_nlinks_collect_metafile(xnc->sc->mp, ino);
fs/xfs/scrub/nlinks.c
505
struct xfs_mount *mp = xnc->sc->mp;
fs/xfs/scrub/nlinks.c
513
error = xchk_nlinks_collect_metafile(xnc, mp->m_sb.sb_rbmino);
fs/xfs/scrub/nlinks.c
517
error = xchk_nlinks_collect_metafile(xnc, mp->m_sb.sb_rsumino);
fs/xfs/scrub/nlinks.c
521
error = xchk_nlinks_collect_metafile(xnc, mp->m_sb.sb_uquotino);
fs/xfs/scrub/nlinks.c
525
error = xchk_nlinks_collect_metafile(xnc, mp->m_sb.sb_gquotino);
fs/xfs/scrub/nlinks.c
529
error = xchk_nlinks_collect_metafile(xnc, mp->m_sb.sb_pquotino);
fs/xfs/scrub/nlinks.c
707
if (!xfs_has_ftype(sc->mp) && S_ISDIR(VFS_I(ip)->i_mode))
fs/xfs/scrub/nlinks.c
713
trace_xchk_nlinks_compare_inode(sc->mp, ip, &obs);
fs/xfs/scrub/nlinks.c
806
struct xfs_mount *mp = xnc->sc->mp;
fs/xfs/scrub/nlinks.c
851
trace_xchk_nlinks_check_zero(mp, ino, &obs);
fs/xfs/scrub/nlinks.c
972
xfs_dir_hook_del(xnc->sc->mp, &xnc->dhook);
fs/xfs/scrub/nlinks.c
993
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/nlinks.c
995
xfs_agnumber_t last_agno = mp->m_sb.sb_agcount - 1;
fs/xfs/scrub/nlinks_repair.c
137
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_link, 0, 0, 0,
fs/xfs/scrub/nlinks_repair.c
179
trace_xrep_nlinks_unfixable_inode(mp, ip, &obs);
fs/xfs/scrub/nlinks_repair.c
243
trace_xrep_nlinks_update_inode(mp, ip, &obs);
fs/xfs/scrub/nlinks_repair.c
308
if (!xfs_has_ftype(sc->mp))
fs/xfs/scrub/nlinks_repair.c
78
pag = xfs_perag_get(sc->mp, XFS_INO_TO_AGNO(sc->mp, sc->ip->i_ino));
fs/xfs/scrub/nlinks_repair.c
94
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/orphanage.c
101
XFS_STATS_INC(mp, xs_ig_attrchg);
fs/xfs/scrub/orphanage.c
103
if (xfs_has_wsync(mp))
fs/xfs/scrub/orphanage.c
125
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/orphanage.c
127
struct inode *root_inode = VFS_I(sc->mp->m_rootip);
fs/xfs/scrub/orphanage.c
131
if (xfs_is_shutdown(mp))
fs/xfs/scrub/orphanage.c
133
if (xfs_is_readonly(mp)) {
fs/xfs/scrub/orphanage.c
315
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/orphanage.c
330
adopt->orphanage_blkres = xfs_link_space_res(mp, MAXNAMELEN);
fs/xfs/scrub/orphanage.c
332
child_blkres = xfs_rename_space_res(mp, 0, false,
fs/xfs/scrub/orphanage.c
334
if (xfs_has_parent(mp))
fs/xfs/scrub/orphanage.c
335
child_blkres += XFS_ADDAFORK_SPACE_RES(mp);
fs/xfs/scrub/orphanage.c
343
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_link,
fs/xfs/scrub/orphanage.c
451
trace_xrep_adoption_check_child(sc->mp, d_child);
fs/xfs/scrub/orphanage.c
488
trace_xrep_adoption_invalidate_child(sc->mp, d_child);
fs/xfs/scrub/orphanage.c
500
trace_xrep_adoption_invalidate_child(sc->mp, d_child);
fs/xfs/scrub/orphanage.c
546
if (!xfs_inode_has_attr_fork(sc->ip) && xfs_has_parent(sc->mp)) {
fs/xfs/scrub/orphanage.c
55
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/orphanage.c
585
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/orphanage.c
83
if (XFS_IS_UQUOTA_ON(mp))
fs/xfs/scrub/orphanage.c
88
if (XFS_IS_GQUOTA_ON(mp))
fs/xfs/scrub/orphanage.c
93
if (XFS_IS_PQUOTA_ON(mp))
fs/xfs/scrub/parent.c
121
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/parent.c
128
if (sc->ip == mp->m_rootip) {
fs/xfs/scrub/parent.c
129
if (sc->ip->i_ino != mp->m_sb.sb_rootino ||
fs/xfs/scrub/parent.c
136
if (sc->ip == mp->m_metadirip) {
fs/xfs/scrub/parent.c
137
if (sc->ip->i_ino != mp->m_sb.sb_metadirino ||
fs/xfs/scrub/parent.c
288
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/parent.c
311
if (!xfs_verify_dir_ino(sc->mp, pp->parent_ino)) {
fs/xfs/scrub/parent.c
422
error = xfs_dir_ino_validate(sc->mp, parent_ino);
fs/xfs/scrub/parent.c
485
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/parent.c
689
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/parent.c
742
if (xfs_has_metadir(sc->mp) && xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/parent.c
844
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/parent.c
848
if (xfs_has_parent(mp))
fs/xfs/scrub/parent.c
859
if (!xfs_verify_dir_ino(mp, sc->ip->i_ino)) {
fs/xfs/scrub/parent.c
873
if (!xfs_verify_dir_ino(mp, parent_ino)) {
fs/xfs/scrub/parent_repair.c
1355
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/parent_repair.c
1373
if (try_adoption && xfs_has_metadir(sc->mp) &&
fs/xfs/scrub/parent_repair.c
1408
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/parent_repair.c
1443
if (xfs_has_metadir(sc->mp) && xchk_inode_is_sb_rooted(sc->ip))
fs/xfs/scrub/parent_repair.c
1454
pag = xfs_perag_get(sc->mp, XFS_INO_TO_AGNO(sc->mp, ip->i_ino));
fs/xfs/scrub/parent_repair.c
1500
struct xfs_da_geometry *geo = sc->mp->m_attr_geo;
fs/xfs/scrub/parent_repair.c
1504
if (!xfs_has_parent(sc->mp))
fs/xfs/scrub/parent_repair.c
1586
if (xfs_has_parent(sc->mp)) {
fs/xfs/scrub/parent_repair.c
1587
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/parent_repair.c
1589
if (!xfs_has_exchange_range(sc->mp))
fs/xfs/scrub/parent_repair.c
1597
if (xfs_has_parent(sc->mp))
fs/xfs/scrub/parent_repair.c
1611
if (xfs_has_parent(sc->mp) && !S_ISDIR(VFS_I(sc->ip)->i_mode)) {
fs/xfs/scrub/parent_repair.c
693
spaceres = xfs_rename_space_res(sc->mp, 0, false, xfs_name_dotdot.len,
fs/xfs/scrub/parent_repair.c
731
error = xfs_parent_from_attr(sc->mp, attr_flags, name, namelen, value,
fs/xfs/scrub/parent_repair.c
902
.geo = sc->mp->m_attr_geo,
fs/xfs/scrub/parent_repair.c
988
.geo = rp->sc->mp->m_attr_geo,
fs/xfs/scrub/quota.c
111
if (!xfs_verify_fsbno(mp, irec.br_startblock))
fs/xfs/scrub/quota.c
113
if (XFS_FSB_TO_DADDR(mp, irec.br_startblock) != dq->q_blkno)
fs/xfs/scrub/quota.c
145
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota.c
146
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/scrub/quota.c
187
if (dq->q_blk.hardlimit > mp->m_sb.sb_dblocks)
fs/xfs/scrub/quota.c
192
if (dq->q_ino.hardlimit > M_IGEO(mp)->maxicount)
fs/xfs/scrub/quota.c
197
if (dq->q_rtb.hardlimit > mp->m_sb.sb_rblocks)
fs/xfs/scrub/quota.c
203
fs_icount = percpu_counter_sum(&mp->m_icount);
fs/xfs/scrub/quota.c
210
if (xfs_has_reflink(mp)) {
fs/xfs/scrub/quota.c
211
if (mp->m_sb.sb_dblocks < dq->q_blk.count)
fs/xfs/scrub/quota.c
214
if (mp->m_sb.sb_rblocks < dq->q_rtb.count)
fs/xfs/scrub/quota.c
218
if (mp->m_sb.sb_dblocks < dq->q_blk.count)
fs/xfs/scrub/quota.c
221
if (mp->m_sb.sb_rblocks < dq->q_rtb.count)
fs/xfs/scrub/quota.c
267
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/quota.c
307
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota.c
308
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/scrub/quota.c
48
if (!XFS_IS_QUOTA_ON(sc->mp))
fs/xfs/scrub/quota.c
55
if (!xfs_this_quota_on(sc->mp, dqtype))
fs/xfs/scrub/quota.c
65
error = xchk_install_live_inode(sc, xfs_quota_inode(sc->mp, dqtype));
fs/xfs/scrub/quota.c
88
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota.c
92
if (!xfs_verify_fileoff(mp, offset)) {
fs/xfs/scrub/quota_repair.c
112
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota_repair.c
113
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/scrub/quota_repair.c
119
if (!xfs_verify_fileoff(mp, offset)) {
fs/xfs/scrub/quota_repair.c
120
ASSERT(xfs_verify_fileoff(mp, offset));
fs/xfs/scrub/quota_repair.c
138
} else if (dq->q_blkno != XFS_FSB_TO_DADDR(mp, irec.br_startblock)) {
fs/xfs/scrub/quota_repair.c
144
dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec.br_startblock);
fs/xfs/scrub/quota_repair.c
175
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota_repair.c
217
if (!xfs_has_reflink(mp) && dq->q_blk.count > mp->m_sb.sb_dblocks) {
fs/xfs/scrub/quota_repair.c
219
dq->q_blk.reserved += mp->m_sb.sb_dblocks;
fs/xfs/scrub/quota_repair.c
220
dq->q_blk.count = mp->m_sb.sb_dblocks;
fs/xfs/scrub/quota_repair.c
224
fs_icount = percpu_counter_sum(&mp->m_icount);
fs/xfs/scrub/quota_repair.c
232
if (!xfs_has_reflink(mp) && dq->q_rtb.count > mp->m_sb.sb_rblocks) {
fs/xfs/scrub/quota_repair.c
234
dq->q_rtb.reserved += mp->m_sb.sb_rblocks;
fs/xfs/scrub/quota_repair.c
235
dq->q_rtb.count = mp->m_sb.sb_rblocks;
fs/xfs/scrub/quota_repair.c
247
trace_xrep_dquot_item(sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quota_repair.c
266
struct xfs_mount *mp,
fs/xfs/scrub/quota_repair.c
281
new_timer = xfs_dquot_set_timeout(mp,
fs/xfs/scrub/quota_repair.c
301
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/quota_repair.c
308
error = xfs_trans_read_buf(sc->mp, sc->tp, sc->mp->m_ddev_targp, daddr,
fs/xfs/scrub/quota_repair.c
314
error = xfs_trans_read_buf(sc->mp, sc->tp,
fs/xfs/scrub/quota_repair.c
315
sc->mp->m_ddev_targp, daddr, qi->qi_dqchunklen,
fs/xfs/scrub/quota_repair.c
344
trace_xrep_disk_dquot(sc->mp, dqtype, id + i);
fs/xfs/scrub/quota_repair.c
351
if (xfs_has_bigtime(sc->mp) && ddq->d_id)
fs/xfs/scrub/quota_repair.c
354
xrep_quota_fix_timer(sc->mp, ddq, ddq->d_blk_softlimit,
fs/xfs/scrub/quota_repair.c
358
xrep_quota_fix_timer(sc->mp, ddq, ddq->d_ino_softlimit,
fs/xfs/scrub/quota_repair.c
362
xrep_quota_fix_timer(sc->mp, ddq, ddq->d_rtb_softlimit,
fs/xfs/scrub/quota_repair.c
367
uuid_copy(&dqblk->dd_uuid, &sc->mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/quota_repair.c
399
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/quota_repair.c
459
max_dqid_off * sc->mp->m_sb.sb_blocksize,
fs/xfs/scrub/quota_repair.c
485
XFS_FSB_TO_DADDR(sc->mp, fsbno),
fs/xfs/scrub/quota_repair.c
62
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quota_repair.c
69
error = xfs_trans_reserve_more(sc->tp, XFS_QM_DQALLOC_SPACE_RES(mp),
fs/xfs/scrub/quota_repair.c
80
dq->q_blkno = XFS_FSB_TO_DADDR(mp, irec->br_startblock);
fs/xfs/scrub/quota_repair.c
82
trace_xrep_dquot_item_fill_bmap_hole(sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quota_repair.c
85
error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp, dq->q_blkno,
fs/xfs/scrub/quota_repair.c
86
mp->m_quotainfo->qi_dqchunklen, 0, &bp);
fs/xfs/scrub/quotacheck.c
629
error = xfs_qm_dqget(xqc->sc->mp, id, dqtype, false, &dq);
fs/xfs/scrub/quotacheck.c
667
if (!(xfs_quota_chkd_flag(dqtype) & sc->mp->m_qflags)) {
fs/xfs/scrub/quotacheck.c
693
struct xfs_quotainfo *qi = xqc->sc->mp->m_quotainfo;
fs/xfs/scrub/quotacheck.c
744
struct xfs_quotainfo *qi = sc->mp->m_quotainfo;
fs/xfs/scrub/quotacheck.c
757
if (xfs_this_quota_on(sc->mp, XFS_DQTYPE_USER)) {
fs/xfs/scrub/quotacheck.c
764
if (xfs_this_quota_on(sc->mp, XFS_DQTYPE_GROUP)) {
fs/xfs/scrub/quotacheck.c
771
if (xfs_this_quota_on(sc->mp, XFS_DQTYPE_PROJ)) {
fs/xfs/scrub/quotacheck.c
84
if (!XFS_IS_QUOTA_ON(sc->mp))
fs/xfs/scrub/quotacheck_repair.c
109
trace_xrep_quotacheck_dquot(xqc->sc->mp, dq->q_type, dq->q_id);
fs/xfs/scrub/quotacheck_repair.c
134
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/quotacheck_repair.c
172
error = xfs_qm_dqget(mp, id, dqtype, true, &dq);
fs/xfs/scrub/quotacheck_repair.c
191
struct xfs_mount *mp)
fs/xfs/scrub/quotacheck_repair.c
195
if (XFS_IS_UQUOTA_ON(mp))
fs/xfs/scrub/quotacheck_repair.c
197
if (XFS_IS_GQUOTA_ON(mp))
fs/xfs/scrub/quotacheck_repair.c
199
if (XFS_IS_PQUOTA_ON(mp))
fs/xfs/scrub/quotacheck_repair.c
210
unsigned int qflags = xqcheck_chkd_flags(sc->mp);
fs/xfs/scrub/rcbag.c
155
struct xfs_mount *mp = bag->mp;
fs/xfs/scrub/rcbag.c
164
cur = rcbagbt_mem_cursor(mp, tp, &bag->xfbtree);
fs/xfs/scrub/rcbag.c
216
struct xfs_mount *mp = bag->mp;
fs/xfs/scrub/rcbag.c
222
cur = rcbagbt_mem_cursor(mp, tp, &bag->xfbtree);
fs/xfs/scrub/rcbag.c
25
struct xfs_mount *mp;
fs/xfs/scrub/rcbag.c
272
struct xfs_mount *mp = bag->mp;
fs/xfs/scrub/rcbag.c
278
cur = rcbagbt_mem_cursor(mp, tp, &bag->xfbtree);
fs/xfs/scrub/rcbag.c
298
xfs_err(bag->mp, "[%llu]: bno 0x%x fsbcount 0x%x refcount 0x%llx\n",
fs/xfs/scrub/rcbag.c
32
struct xfs_mount *mp,
fs/xfs/scrub/rcbag.c
44
bag->mp = mp;
fs/xfs/scrub/rcbag.c
46
error = rcbagbt_mem_init(mp, &bag->xfbtree, btp);
fs/xfs/scrub/rcbag.c
77
struct xfs_mount *mp = bag->mp;
fs/xfs/scrub/rcbag.c
82
cur = rcbagbt_mem_cursor(mp, tp, &bag->xfbtree);
fs/xfs/scrub/rcbag.h
13
int rcbag_init(struct xfs_mount *mp, struct xfs_buftarg *btp,
fs/xfs/scrub/rcbag_btree.c
126
struct xfs_mount *mp = bp->b_mount;
fs/xfs/scrub/rcbag_btree.c
143
maxrecs = rcbagbt_maxrecs(mp, XFBNO_BLOCKSIZE, level == 0);
fs/xfs/scrub/rcbag_btree.c
196
struct xfs_mount *mp,
fs/xfs/scrub/rcbag_btree.c
202
cur = xfs_btree_alloc_cursor(mp, tp, &rcbagbt_mem_ops,
fs/xfs/scrub/rcbag_btree.c
213
struct xfs_mount *mp,
fs/xfs/scrub/rcbag_btree.c
218
return xfbtree_init(mp, xfbt, btp, &rcbagbt_mem_ops);
fs/xfs/scrub/rcbag_btree.c
238
struct xfs_mount *mp,
fs/xfs/scrub/rcbag_btree.h
54
unsigned int rcbagbt_maxrecs(struct xfs_mount *mp, unsigned int blocklen,
fs/xfs/scrub/rcbag_btree.h
64
struct xfs_btree_cur *rcbagbt_mem_cursor(struct xfs_mount *mp,
fs/xfs/scrub/rcbag_btree.h
66
int rcbagbt_mem_init(struct xfs_mount *mp, struct xfbtree *xfbtree,
fs/xfs/scrub/readdir.c
123
next_off = off + xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/scrub/readdir.c
131
name.type = xfs_dir2_data_get_ftype(mp, dep);
fs/xfs/scrub/readdir.c
190
struct xfs_mount *mp = dp->i_mount;
fs/xfs/scrub/readdir.c
191
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/scrub/readdir.c
239
length = xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/scrub/readdir.c
245
name.type = xfs_dir2_data_get_ftype(mp, dep);
fs/xfs/scrub/readdir.c
37
struct xfs_mount *mp = dp->i_mount;
fs/xfs/scrub/readdir.c
38
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/scrub/readdir.c
60
xfs_dir2_data_entsize(mp, sizeof(".") - 1));
fs/xfs/scrub/readdir.c
74
ino = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/scrub/readdir.c
77
name.type = xfs_dir2_sf_get_ftype(mp, sfep);
fs/xfs/scrub/readdir.c
83
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/scrub/readdir.c
97
struct xfs_mount *mp = dp->i_mount;
fs/xfs/scrub/readdir.c
98
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/scrub/reap.c
1027
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
1064
const unsigned int f1 = xfs_calc_finish_rt_efi_reservation(mp, 1);
fs/xfs/scrub/reap.c
1065
const unsigned int f2 = xfs_calc_finish_rt_rui_reservation(mp, 1);
fs/xfs/scrub/reap.c
1066
const unsigned int f3 = xfs_calc_finish_rt_cui_reservation(mp, 1);
fs/xfs/scrub/reap.c
1096
xfs_rgblock_t rgbno = xfs_rtb_to_rgbno(sc->mp, rtbno);
fs/xfs/scrub/reap.c
1107
sc->sr.rtg = xfs_rtgroup_get(sc->mp, xfs_rtb_to_rgno(sc->mp, rtbno));
fs/xfs/scrub/reap.c
1165
ASSERT(xfs_has_rmapbt(sc->mp));
fs/xfs/scrub/reap.c
1203
ASSERT(xfs_has_rmapbt(sc->mp));
fs/xfs/scrub/reap.c
1246
agbno = XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock);
fs/xfs/scrub/reap.c
1249
cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/reap.c
1320
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
1335
agbno = bno = XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock);
fs/xfs/scrub/reap.c
1350
max_off = off + xfs_attr3_max_rmt_blocks(mp);
fs/xfs/scrub/reap.c
1382
.max_sectors = xrep_bufscan_max_sectors(mp,
fs/xfs/scrub/reap.c
1384
.daddr_step = XFS_FSB_TO_BB(mp, 1),
fs/xfs/scrub/reap.c
1388
while ((bp = xrep_bufscan_advance(mp, &scan)) != NULL) {
fs/xfs/scrub/reap.c
1441
XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock),
fs/xfs/scrub/reap.c
1464
XFS_FSB_TO_AGBNO(sc->mp, imap->br_startblock),
fs/xfs/scrub/reap.c
1497
return sc->mp->m_dir_geo->fsbcount;
fs/xfs/scrub/reap.c
1510
return sc->mp->m_dir_geo->blksize;
fs/xfs/scrub/reap.c
1521
return sc->mp->m_sb.sb_blocksize;
fs/xfs/scrub/reap.c
1533
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
1569
const unsigned int f1 = xfs_calc_finish_efi_reservation(mp, 1);
fs/xfs/scrub/reap.c
1570
const unsigned int f2 = xfs_calc_finish_rui_reservation(mp, 1);
fs/xfs/scrub/reap.c
1571
const unsigned int f3 = xfs_calc_finish_bui_reservation(mp, 1);
fs/xfs/scrub/reap.c
1607
agno = XFS_FSB_TO_AGNO(sc->mp, imap->br_startblock);
fs/xfs/scrub/reap.c
1608
sc->sa.pag = xfs_perag_get(sc->mp, agno);
fs/xfs/scrub/reap.c
1657
ASSERT(xfs_has_rmapbt(sc->mp));
fs/xfs/scrub/reap.c
238
struct xfs_mount *mp)
fs/xfs/scrub/reap.c
241
return xfs_attr3_max_rmt_blocks(mp);
fs/xfs/scrub/reap.c
250
struct xfs_mount *mp,
fs/xfs/scrub/reap.c
253
return XFS_FSB_TO_BB(mp, min_t(xfs_extlen_t, fsblocks,
fs/xfs/scrub/reap.c
254
xrep_binval_max_fsblocks(mp)));
fs/xfs/scrub/reap.c
263
struct xfs_mount *mp,
fs/xfs/scrub/reap.c
271
error = xfs_buf_incore(mp->m_ddev_targp, scan->daddr,
fs/xfs/scrub/reap.c
291
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
313
.max_sectors = xrep_bufscan_max_sectors(mp,
fs/xfs/scrub/reap.c
315
.daddr_step = XFS_FSB_TO_BB(mp, 1),
fs/xfs/scrub/reap.c
319
while ((bp = xrep_bufscan_advance(mp, &scan)) != NULL) {
fs/xfs/scrub/reap.c
365
cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/reap.c
537
xfs_force_shutdown(sc->mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/scrub/reap.c
557
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
595
const unsigned int f1 = xfs_calc_finish_efi_reservation(mp, 1);
fs/xfs/scrub/reap.c
596
const unsigned int f2 = xfs_calc_finish_rui_reservation(mp, 1);
fs/xfs/scrub/reap.c
600
const unsigned int max_binval = xrep_binval_max_fsblocks(mp);
fs/xfs/scrub/reap.c
604
xfs_buf_inval_log_space(1, XFS_B_TO_FSBT(mp, max_binval));
fs/xfs/scrub/reap.c
636
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/reap.c
673
const unsigned int f1 = xfs_calc_finish_efi_reservation(mp, 1);
fs/xfs/scrub/reap.c
674
const unsigned int f2 = xfs_calc_finish_rui_reservation(mp, 1);
fs/xfs/scrub/reap.c
675
const unsigned int f3 = xfs_calc_finish_cui_reservation(mp, 1);
fs/xfs/scrub/reap.c
679
const unsigned int max_binval = xrep_binval_max_fsblocks(mp);
fs/xfs/scrub/reap.c
683
xfs_buf_inval_log_space(1, XFS_B_TO_FSBT(mp, max_binval));
fs/xfs/scrub/reap.c
767
ASSERT(xfs_has_rmapbt(sc->mp));
fs/xfs/scrub/reap.c
794
xfs_agnumber_t agno = XFS_FSB_TO_AGNO(sc->mp, fsbno);
fs/xfs/scrub/reap.c
795
xfs_agblock_t agbno = XFS_FSB_TO_AGBNO(sc->mp, fsbno);
fs/xfs/scrub/reap.c
807
sc->sa.pag = xfs_perag_get(sc->mp, agno);
fs/xfs/scrub/reap.c
880
ASSERT(xfs_has_rmapbt(sc->mp));
fs/xfs/scrub/reap.h
42
xfs_daddr_t xrep_bufscan_max_sectors(struct xfs_mount *mp,
fs/xfs/scrub/reap.h
44
struct xfs_buf *xrep_bufscan_advance(struct xfs_mount *mp,
fs/xfs/scrub/refcount.c
541
xchk_refcountbt_xref_gaps(sc, &rrc, sc->mp->m_sb.sb_agblocks);
fs/xfs/scrub/refcount_repair.c
204
struct xfs_mount *mp,
fs/xfs/scrub/refcount_repair.c
212
if (xfs_is_sb_inum(mp, rmap->rm_owner))
fs/xfs/scrub/refcount_repair.c
234
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/scrub/refcount_repair.c
259
if (XFS_IS_CORRUPT(mp, !have_gt)) {
fs/xfs/scrub/refcount_repair.c
278
} while (!xrep_refc_rmap_shareable(mp, rmap));
fs/xfs/scrub/refcount_repair.c
392
if (XFS_IS_CORRUPT(sc->mp, !have_gt)) {
fs/xfs/scrub/refcount_repair.c
421
error = rcbag_init(sc->mp, sc->xmbtp, &rcstack);
fs/xfs/scrub/refcount_repair.c
600
xfs_agbno_to_fsb(pag, xfs_refc_block(sc->mp)),
fs/xfs/scrub/refcount_repair.c
606
refc_cur = xfs_refcountbt_init_cursor(sc->mp, NULL, NULL, pag);
fs/xfs/scrub/refcount_repair.c
700
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/refcount_repair.c
704
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/refcount_repair.c
713
error = xfarray_create("reference count records", mp->m_sb.sb_agblocks,
fs/xfs/scrub/repair.c
1006
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
1012
xfs_has_rtrmapbt(mp))
fs/xfs/scrub/repair.c
1017
xfs_has_rtreflink(mp))
fs/xfs/scrub/repair.c
1050
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
1056
if (xfs_has_zoned(mp)) {
fs/xfs/scrub/repair.c
1062
startrtx = xfs_rgbno_to_rtx(mp, rgbno);
fs/xfs/scrub/repair.c
1063
endrtx = xfs_rgbno_to_rtx(mp, rgbno + len - 1);
fs/xfs/scrub/repair.c
1094
xfs_err(sc->mp,
fs/xfs/scrub/repair.c
1113
if (XFS_TEST_ERROR(sc->mp, XFS_ERRTAG_FORCE_SCRUB_REPAIR))
fs/xfs/scrub/repair.c
1234
if (xfs_inode_hasattr(sc->ip) && !xfs_has_metadir(sc->mp)) {
fs/xfs/scrub/repair.c
1271
return xmbuf_alloc(sc->mp, descr, &sc->xmbtp);
fs/xfs/scrub/repair.c
129
struct xfs_mount *mp)
fs/xfs/scrub/repair.c
131
xfs_alert_ratelimited(mp,
fs/xfs/scrub/repair.c
1363
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
1367
delta = mp->m_metafile_resv_used + mp->m_metafile_resv_avail -
fs/xfs/scrub/repair.c
1368
mp->m_metafile_resv_target;
fs/xfs/scrub/repair.c
1379
give_back = min_t(uint64_t, delta, mp->m_metafile_resv_avail);
fs/xfs/scrub/repair.c
1381
xfs_mod_sb_delalloc(mp, -give_back);
fs/xfs/scrub/repair.c
1382
xfs_add_fdblocks(mp, give_back);
fs/xfs/scrub/repair.c
1383
mp->m_metafile_resv_avail -= give_back;
fs/xfs/scrub/repair.c
1394
error = xfs_dec_fdblocks(mp, delta, true);
fs/xfs/scrub/repair.c
1398
xfs_warn(sc->mp,
fs/xfs/scrub/repair.c
1402
error = xfs_dec_fdblocks(mp, delta, true);
fs/xfs/scrub/repair.c
1407
xfs_mod_sb_delalloc(mp, delta);
fs/xfs/scrub/repair.c
1408
mp->m_metafile_resv_avail += delta;
fs/xfs/scrub/repair.c
284
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
301
pag = xfs_perag_get(mp, sm->sm_agno);
fs/xfs/scrub/repair.c
351
bnobt_sz = 2 * xfs_allocbt_calc_size(mp, freelen);
fs/xfs/scrub/repair.c
352
if (xfs_has_sparseinodes(mp))
fs/xfs/scrub/repair.c
353
inobt_sz = xfs_iallocbt_calc_size(mp, icount /
fs/xfs/scrub/repair.c
356
inobt_sz = xfs_iallocbt_calc_size(mp, icount /
fs/xfs/scrub/repair.c
358
if (xfs_has_finobt(mp))
fs/xfs/scrub/repair.c
360
if (xfs_has_reflink(mp))
fs/xfs/scrub/repair.c
361
refcbt_sz = xfs_refcountbt_calc_size(mp, usedlen);
fs/xfs/scrub/repair.c
364
if (xfs_has_rmapbt(mp)) {
fs/xfs/scrub/repair.c
373
if (xfs_has_reflink(mp))
fs/xfs/scrub/repair.c
374
rmapbt_sz = xfs_rmapbt_calc_size(mp,
fs/xfs/scrub/repair.c
377
rmapbt_sz = xfs_rmapbt_calc_size(mp, usedlen);
fs/xfs/scrub/repair.c
399
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
406
if (!xfs_has_rtgroups(mp)) {
fs/xfs/scrub/repair.c
411
usedlen = xfs_rtbxlen_to_blen(mp, xfs_rtgroup_extents(mp, sm->sm_agno));
fs/xfs/scrub/repair.c
414
if (xfs_has_rmapbt(mp))
fs/xfs/scrub/repair.c
415
rmapbt_sz = xfs_rtrmapbt_calc_size(mp, usedlen);
fs/xfs/scrub/repair.c
417
trace_xrep_calc_rtgroup_resblks_btsize(mp, sm->sm_agno, usedlen,
fs/xfs/scrub/repair.c
458
args.mp = sc->mp;
fs/xfs/scrub/repair.c
504
struct xfs_mount *mp,
fs/xfs/scrub/repair.c
522
struct xfs_mount *mp = ri->sc->mp;
fs/xfs/scrub/repair.c
538
error = xfs_agfl_walk(mp, ri->agf, ri->agfl_bp,
fs/xfs/scrub/repair.c
564
error = xfs_trans_read_buf(mp, ri->sc->tp, mp->m_ddev_targp, daddr,
fs/xfs/scrub/repair.c
565
mp->m_bsize, 0, &bp, NULL);
fs/xfs/scrub/repair.c
595
&mp->m_sb.sb_meta_uuid))
fs/xfs/scrub/repair.c
712
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
732
cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.pag);
fs/xfs/scrub/repair.c
747
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
750
mutex_lock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/scrub/repair.c
751
if ((mp->m_qflags & clear_flags) == 0 &&
fs/xfs/scrub/repair.c
752
(mp->m_qflags & set_flags) == set_flags)
fs/xfs/scrub/repair.c
755
mp->m_qflags &= ~clear_flags;
fs/xfs/scrub/repair.c
756
mp->m_qflags |= set_flags;
fs/xfs/scrub/repair.c
758
spin_lock(&mp->m_sb_lock);
fs/xfs/scrub/repair.c
759
mp->m_sb.sb_qflags &= ~clear_flags;
fs/xfs/scrub/repair.c
760
mp->m_sb.sb_qflags |= set_flags;
fs/xfs/scrub/repair.c
761
spin_unlock(&mp->m_sb_lock);
fs/xfs/scrub/repair.c
770
xfs_sb_to_disk(bp->b_addr, &mp->m_sb);
fs/xfs/scrub/repair.c
775
mutex_unlock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/scrub/repair.c
787
if (!(flag & sc->mp->m_qflags))
fs/xfs/scrub/repair.c
817
xfs_err_ratelimited(sc->mp,
fs/xfs/scrub/repair.c
820
if (XFS_IS_UQUOTA_ON(sc->mp) && !sc->ip->i_udquot)
fs/xfs/scrub/repair.c
822
if (XFS_IS_GQUOTA_ON(sc->mp) && !sc->ip->i_gdquot)
fs/xfs/scrub/repair.c
824
if (XFS_IS_PQUOTA_ON(sc->mp) && !sc->ip->i_pdquot)
fs/xfs/scrub/repair.c
858
if (!xfs_has_large_extent_counts(sc->mp))
fs/xfs/scrub/repair.c
879
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/repair.c
884
sa->bno_cur = xfs_bnobt_init_cursor(mp, sc->tp, sa->agf_bp,
fs/xfs/scrub/repair.c
886
sa->cnt_cur = xfs_cntbt_init_cursor(mp, sc->tp, sa->agf_bp,
fs/xfs/scrub/repair.c
895
if (xfs_has_finobt(mp))
fs/xfs/scrub/repair.c
902
xfs_has_rmapbt(mp))
fs/xfs/scrub/repair.c
903
sa->rmap_cur = xfs_rmapbt_init_cursor(mp, sc->tp, sa->agf_bp,
fs/xfs/scrub/repair.c
908
xfs_has_reflink(mp))
fs/xfs/scrub/repair.c
909
sa->refc_cur = xfs_refcountbt_init_cursor(mp, sc->tp,
fs/xfs/scrub/repair.h
201
static inline void xrep_failure(struct xfs_mount *mp) {}
fs/xfs/scrub/repair.h
25
void xrep_failure(struct xfs_mount *mp);
fs/xfs/scrub/rmap.c
190
if (!xfs_has_reflink(sc->mp))
fs/xfs/scrub/rmap.c
377
struct xfs_mount *mp,
fs/xfs/scrub/rmap.c
400
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rmap.c
407
error = xagb_bitmap_set(&cr->fs_owned, XFS_SB_BLOCK(mp),
fs/xfs/scrub/rmap.c
408
XFS_AGFL_BLOCK(mp) - XFS_SB_BLOCK(mp) + 1);
fs/xfs/scrub/rmap.c
413
if (xfs_ag_contains_log(mp, pag_agno(sc->sa.pag))) {
fs/xfs/scrub/rmap.c
415
XFS_FSB_TO_AGBNO(mp, mp->m_sb.sb_logstart),
fs/xfs/scrub/rmap.c
416
mp->m_sb.sb_logblocks);
fs/xfs/scrub/rmap.c
424
cur = xfs_bnobt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rmap.c
434
cur = xfs_cntbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rmap.c
450
error = xfs_agfl_walk(sc->mp, agf, agfl_bp, xchk_rmapbt_walk_agfl,
fs/xfs/scrub/rmap.c
466
if (xfs_has_finobt(sc->mp)) {
fs/xfs/scrub/rmap.c
479
if (xfs_has_reflink(sc->mp)) {
fs/xfs/scrub/rmap.c
482
cur = xfs_refcountbt_init_cursor(sc->mp, sc->tp,
fs/xfs/scrub/rmap_repair.c
1017
struct xfs_mount *mp,
fs/xfs/scrub/rmap_repair.c
1079
sc->sa.bno_cur = xfs_bnobt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rmap_repair.c
1088
sc->sa.cnt_cur = xfs_cntbt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rmap_repair.c
1111
error = xfs_agfl_walk(sc->mp, agf, agfl_bp, xrep_rmap_walk_agfl, &ra);
fs/xfs/scrub/rmap_repair.c
1357
xfs_agbno_to_fsb(pag, XFS_RMAP_BLOCK(sc->mp)),
fs/xfs/scrub/rmap_repair.c
1362
rmap_cur = xfs_rmapbt_init_cursor(sc->mp, NULL, NULL, pag);
fs/xfs/scrub/rmap_repair.c
1533
sc->sa.bno_cur = xfs_bnobt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rmap_repair.c
1607
struct xfs_mount *mp;
fs/xfs/scrub/rmap_repair.c
1613
mp = rr->sc->mp;
fs/xfs/scrub/rmap_repair.c
1620
tp = xfs_trans_alloc_empty(mp);
fs/xfs/scrub/rmap_repair.c
1658
error = xfs_rmapbt_mem_init(sc->mp, &rr->rmap_btree, sc->xmbtp,
fs/xfs/scrub/rmap_repair.c
340
struct xfs_mount *mp = rf->rr->sc->mp;
fs/xfs/scrub/rmap_repair.c
346
if (XFS_FSB_TO_AGNO(mp, rec->br_startblock) !=
fs/xfs/scrub/rmap_repair.c
350
agbno = XFS_FSB_TO_AGBNO(mp, rec->br_startblock);
fs/xfs/scrub/rmap_repair.c
453
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, rf->whichfork);
fs/xfs/scrub/rmap_repair.c
528
while ((rtg = xfs_rtgroup_next(sc->mp, rtg))) {
fs/xfs/scrub/rmap_repair.c
677
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/scrub/rmap_repair.c
690
xfs_inobt_btrec_to_irec(mp, rec, &irec);
fs/xfs/scrub/rmap_repair.c
698
agbno = XFS_AGINO_TO_AGBNO(mp, agino);
fs/xfs/scrub/rmap_repair.c
700
XFS_INODES_PER_CHUNK / mp->m_sb.sb_inopblock);
fs/xfs/scrub/rmap_repair.c
706
iperhole = max_t(xfs_agino_t, mp->m_sb.sb_inopblock,
fs/xfs/scrub/rmap_repair.c
708
aglen = iperhole / mp->m_sb.sb_inopblock;
fs/xfs/scrub/rmap_repair.c
717
agbno = XFS_AGINO_TO_AGBNO(mp, agino);
fs/xfs/scrub/rmap_repair.c
762
if (xfs_has_finobt(sc->mp)) {
fs/xfs/scrub/rmap_repair.c
820
if (!xfs_has_reflink(sc->mp))
fs/xfs/scrub/rmap_repair.c
858
return xrep_rmap_stash(rr, XFS_SB_BLOCK(sc->mp),
fs/xfs/scrub/rmap_repair.c
859
XFS_AGFL_BLOCK(sc->mp) - XFS_SB_BLOCK(sc->mp) + 1,
fs/xfs/scrub/rmap_repair.c
870
if (!xfs_ag_contains_log(sc->mp, pag_agno(sc->sa.pag)))
fs/xfs/scrub/rmap_repair.c
874
XFS_FSB_TO_AGBNO(sc->mp, sc->mp->m_sb.sb_logstart),
fs/xfs/scrub/rmap_repair.c
875
sc->mp->m_sb.sb_logblocks, XFS_RMAP_OWN_LOG, 0, 0);
fs/xfs/scrub/rmap_repair.c
994
sc->sa.bno_cur = xfs_bnobt_init_cursor(sc->mp, sc->tp, sc->sa.agf_bp,
fs/xfs/scrub/rtbitmap.c
101
xfs_rgblock_t rgbno = xfs_rtb_to_rgbno(sc->mp, startblock);
fs/xfs/scrub/rtbitmap.c
152
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap.c
159
endoff = XFS_B_TO_FSB(mp, ip->i_disk_size);
fs/xfs/scrub/rtbitmap.c
194
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap.c
202
if (mp->m_sb.sb_rextents != rtb->rextents) {
fs/xfs/scrub/rtbitmap.c
208
if (mp->m_sb.sb_rextslog != rtb->rextslog) {
fs/xfs/scrub/rtbitmap.c
221
if (mp->m_sb.sb_rbmblocks != rtb->rbmblocks) {
fs/xfs/scrub/rtbitmap.c
227
if (rbmip->i_disk_size & mp->m_blockmask) {
fs/xfs/scrub/rtbitmap.c
237
if (rbmip->i_disk_size < XFS_FSB_TO_B(mp, rtb->rbmblocks)) {
fs/xfs/scrub/rtbitmap.c
261
last_rgbno = rtg->rtg_extents * mp->m_sb.sb_rextsize - 1;
fs/xfs/scrub/rtbitmap.c
284
if (xfs_has_zoned(sc->mp)) {
fs/xfs/scrub/rtbitmap.c
286
xfs_rtb_to_rgbno(sc->mp, rtbno) + len - 1))
fs/xfs/scrub/rtbitmap.c
291
startext = xfs_rtb_to_rtx(sc->mp, rtbno);
fs/xfs/scrub/rtbitmap.c
292
endext = xfs_rtb_to_rtx(sc->mp, rtbno + len - 1);
fs/xfs/scrub/rtbitmap.c
37
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap.c
82
if (mp->m_sb.sb_rblocks) {
fs/xfs/scrub/rtbitmap.c
83
rtb->rextents = xfs_blen_to_rtbxlen(mp, mp->m_sb.sb_rblocks);
fs/xfs/scrub/rtbitmap.c
85
rtb->rbmblocks = xfs_rtbitmap_blockcount(mp);
fs/xfs/scrub/rtbitmap.h
69
return sc->mp->m_sb.sb_blocksize >> XFS_WORDLOG;
fs/xfs/scrub/rtbitmap_repair.c
106
ASSERT(xfs_has_rtgroups(rtb->sc->mp));
fs/xfs/scrub/rtbitmap_repair.c
126
ASSERT(xfs_has_rtgroups(rtb->sc->mp));
fs/xfs/scrub/rtbitmap_repair.c
170
trace_xrep_rtbitmap_or(rtb->sc->mp, wordoff, mask, word);
fs/xfs/scrub/rtbitmap_repair.c
184
struct xfs_mount *mp = rtb->sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
204
startrtx = xfs_rgbno_to_rtx(mp, rtb->next_rgbno);
fs/xfs/scrub/rtbitmap_repair.c
205
mod = xfs_rgbno_to_rtxoff(mp, rtb->next_rgbno);
fs/xfs/scrub/rtbitmap_repair.c
209
nextrtx = xfs_rgbno_to_rtx(mp, rgbno - 1) + 1;
fs/xfs/scrub/rtbitmap_repair.c
210
mod = xfs_rgbno_to_rtxoff(mp, rgbno - 1);
fs/xfs/scrub/rtbitmap_repair.c
211
if (mod != mp->m_sb.sb_rextsize - 1)
fs/xfs/scrub/rtbitmap_repair.c
233
trace_xrep_rtbitmap_record_free(mp, startrtx, nextrtx - 1);
fs/xfs/scrub/rtbitmap_repair.c
244
error = xrep_rtbitmap_or(rtb, rtx_to_wordoff(mp, startrtx),
fs/xfs/scrub/rtbitmap_repair.c
256
error = xrep_rtbitmap_or(rtb, rtx_to_wordoff(mp, nextrtx),
fs/xfs/scrub/rtbitmap_repair.c
263
trace_xrep_rtbitmap_record_free_bulk(mp, startrtx, nextrtx - 1);
fs/xfs/scrub/rtbitmap_repair.c
266
wordoff = rtx_to_wordoff(mp, startrtx);
fs/xfs/scrub/rtbitmap_repair.c
267
nextwordoff = rtx_to_wordoff(mp, nextrtx);
fs/xfs/scrub/rtbitmap_repair.c
268
bufwsize = mp->m_sb.sb_blocksize >> XFS_WORDLOG;
fs/xfs/scrub/rtbitmap_repair.c
329
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
335
memset(rtb->words, 0xFF, mp->m_sb.sb_blocksize);
fs/xfs/scrub/rtbitmap_repair.c
347
blockcount = rtg->rtg_extents * mp->m_sb.sb_rextsize;
fs/xfs/scrub/rtbitmap_repair.c
366
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
370
rtb->args.mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
377
mp->m_blockwsize);
fs/xfs/scrub/rtbitmap_repair.c
381
if (xfs_has_rtgroups(sc->mp)) {
fs/xfs/scrub/rtbitmap_repair.c
388
uuid_copy(&hdr->rt_uuid, &sc->mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/rtbitmap_repair.c
394
rtb->prep_wordoff += mp->m_blockwsize;
fs/xfs/scrub/rtbitmap_repair.c
45
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
46
unsigned long long blocks = mp->m_sb.sb_rbmblocks;
fs/xfs/scrub/rtbitmap_repair.c
480
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
484
if (mp->m_sb.sb_rextents != rtb->rextents)
fs/xfs/scrub/rtbitmap_repair.c
486
rtb->rextents - mp->m_sb.sb_rextents);
fs/xfs/scrub/rtbitmap_repair.c
488
if (mp->m_sb.sb_rbmblocks != rtb->rbmblocks)
fs/xfs/scrub/rtbitmap_repair.c
490
rtb->rbmblocks - mp->m_sb.sb_rbmblocks);
fs/xfs/scrub/rtbitmap_repair.c
492
if (mp->m_sb.sb_rextslog != rtb->rextslog)
fs/xfs/scrub/rtbitmap_repair.c
494
rtb->rextslog - mp->m_sb.sb_rextslog);
fs/xfs/scrub/rtbitmap_repair.c
498
mp->m_sb.sb_blocksize);
fs/xfs/scrub/rtbitmap_repair.c
500
if (sc->ip->i_disk_size < XFS_FSB_TO_B(mp, rtb->rbmblocks))
fs/xfs/scrub/rtbitmap_repair.c
501
sc->ip->i_disk_size = XFS_FSB_TO_B(mp, rtb->rbmblocks);
fs/xfs/scrub/rtbitmap_repair.c
513
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtbitmap_repair.c
520
if (!xfs_has_rtrmapbt(sc->mp))
fs/xfs/scrub/rtbitmap_repair.c
523
if (!xfs_has_exchange_range(sc->mp))
fs/xfs/scrub/rtbitmap_repair.c
535
blocks = xfs_bmbt_calc_size(mp, rtb->rbmblocks);
fs/xfs/scrub/rtbitmap_repair.c
55
blocks * mp->m_sb.sb_blocksize, &sc->xfile);
fs/xfs/scrub/rtbitmap_repair.c
614
XFS_FSB_TO_B(sc->mp, sc->mp->m_sb.sb_rbmblocks));
fs/xfs/scrub/rtbitmap_repair.c
72
blocks += xfs_bmbt_calc_size(mp, blocks) * 2;
fs/xfs/scrub/rtbitmap_repair.c
82
struct xfs_mount *mp,
fs/xfs/scrub/rtrefcount.c
461
struct xfs_mount *mp = bs->cur->bc_mp;
fs/xfs/scrub/rtrefcount.c
474
mod = xfs_rgbno_to_rtxoff(mp, irec.rc_startblock);
fs/xfs/scrub/rtrefcount.c
477
mod = xfs_extlen_to_rtxmod(mp, irec.rc_blockcount);
fs/xfs/scrub/rtrefcount.c
569
xchk_rtrefcountbt_xref_gaps(sc, &rrc, sc->mp->m_sb.sb_rblocks);
fs/xfs/scrub/rtrefcount_repair.c
145
if (xfs_rgbno_to_rtxoff(sc->mp, rec->rc_startblock) != 0)
fs/xfs/scrub/rtrefcount_repair.c
149
if (xfs_rgbno_to_rtxoff(sc->mp, last) != sc->mp->m_sb.sb_rextsize - 1)
fs/xfs/scrub/rtrefcount_repair.c
222
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/scrub/rtrefcount_repair.c
247
if (XFS_IS_CORRUPT(mp, !have_gt)) {
fs/xfs/scrub/rtrefcount_repair.c
257
} else if (xfs_is_sb_inum(mp, rmap->rm_owner) ||
fs/xfs/scrub/rtrefcount_repair.c
406
if (XFS_IS_CORRUPT(sc->mp, !have_gt)) {
fs/xfs/scrub/rtrefcount_repair.c
448
while ((pag = xfs_perag_next(sc->mp, pag))) {
fs/xfs/scrub/rtrefcount_repair.c
463
error = rcbag_init(sc->mp, sc->xmbtp, &rcstack);
fs/xfs/scrub/rtrefcount_repair.c
700
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtrefcount_repair.c
704
if (!xfs_has_rtrmapbt(mp))
fs/xfs/scrub/rtrefcount_repair.c
719
mp->m_sb.sb_rextents, sizeof(struct xfs_refcount_irec),
fs/xfs/scrub/rtrmap.c
86
if (!xfs_has_rtreflink(sc->mp))
fs/xfs/scrub/rtrmap_repair.c
164
trace_xrep_rtrmap_found(sc->mp, &rmap);
fs/xfs/scrub/rtrmap_repair.c
224
struct xfs_mount *mp = rf->rr->sc->mp;
fs/xfs/scrub/rtrmap_repair.c
229
if (xfs_rtb_to_rgno(mp, rec->br_startblock) !=
fs/xfs/scrub/rtrmap_repair.c
237
rgbno = xfs_rtb_to_rgbno(mp, rec->br_startblock);
fs/xfs/scrub/rtrmap_repair.c
285
cur = xfs_bmbt_init_cursor(rr->sc->mp, rr->sc->tp, ip, XFS_DATA_FORK);
fs/xfs/scrub/rtrmap_repair.c
498
if (!xfs_has_rtreflink(sc->mp))
fs/xfs/scrub/rtrmap_repair.c
549
if (xfs_has_rtsb(sc->mp) && rtg_rgno(rr->sc->sr.rtg) == 0) {
fs/xfs/scrub/rtrmap_repair.c
550
error = xrep_rtrmap_stash(rr, 0, sc->mp->m_sb.sb_rextsize,
fs/xfs/scrub/rtrmap_repair.c
615
while ((pag = xfs_perag_next(sc->mp, pag))) {
fs/xfs/scrub/rtrmap_repair.c
841
struct xfs_mount *mp;
fs/xfs/scrub/rtrmap_repair.c
847
mp = rr->sc->mp;
fs/xfs/scrub/rtrmap_repair.c
854
tp = xfs_trans_alloc_empty(mp);
fs/xfs/scrub/rtrmap_repair.c
893
error = xfs_rtrmapbt_mem_init(sc->mp, &rr->rtrmap_btree, sc->xmbtp,
fs/xfs/scrub/rtsummary.c
102
if (mp->m_sb.sb_rblocks) {
fs/xfs/scrub/rtsummary.c
103
rts->rextents = xfs_blen_to_rtbxlen(mp, mp->m_sb.sb_rblocks);
fs/xfs/scrub/rtsummary.c
104
rts->rbmblocks = xfs_rtbitmap_blockcount(mp);
fs/xfs/scrub/rtsummary.c
106
xfs_rtsummary_blockcount(mp, &rts->rsumlevels);
fs/xfs/scrub/rtsummary.c
149
struct xfs_mount *mp,
fs/xfs/scrub/rtsummary.c
152
if (xfs_has_rtgroups(mp)) {
fs/xfs/scrub/rtsummary.c
169
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/scrub/rtsummary.c
184
rbmoff = xfs_rtx_to_rbmblock(mp, rec->ar_startext);
fs/xfs/scrub/rtsummary.c
186
offs = xfs_rtsumoffs(mp, lenlog, rbmoff);
fs/xfs/scrub/rtsummary.c
189
rtlen = xfs_rtxlen_to_extlen(mp, rec->ar_extcount);
fs/xfs/scrub/rtsummary.c
191
if (!xfs_verify_rtbext(mp, rtbno, rtlen)) {
fs/xfs/scrub/rtsummary.c
201
value = xchk_rtsum_inc(sc->mp, &v);
fs/xfs/scrub/rtsummary.c
202
trace_xchk_rtsum_record_free(mp, rec->ar_startext, rec->ar_extcount,
fs/xfs/scrub/rtsummary.c
213
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary.c
217
if (XFS_FSB_TO_B(mp, xfs_rtbitmap_blockcount(mp)) !=
fs/xfs/scrub/rtsummary.c
232
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary.c
240
rts->args.mp = mp;
fs/xfs/scrub/rtsummary.c
245
endoff = XFS_B_TO_FSB(mp, ip->i_disk_size);
fs/xfs/scrub/rtsummary.c
282
error = xfsum_copyout(sc, sumoff, rts->words, mp->m_blockwsize);
fs/xfs/scrub/rtsummary.c
290
mp->m_blockwsize << XFS_WORDLOG) != 0) {
fs/xfs/scrub/rtsummary.c
297
sumoff += mp->m_blockwsize;
fs/xfs/scrub/rtsummary.c
308
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary.c
316
if (mp->m_sb.sb_rextents != rts->rextents) {
fs/xfs/scrub/rtsummary.c
322
if (mp->m_rsumlevels != rts->rsumlevels) {
fs/xfs/scrub/rtsummary.c
328
if (mp->m_rsumblocks != rts->rsumblocks) {
fs/xfs/scrub/rtsummary.c
334
if (rsumip->i_disk_size & mp->m_blockmask) {
fs/xfs/scrub/rtsummary.c
344
if (rsumip->i_disk_size < XFS_FSB_TO_B(mp, rts->rsumblocks)) {
fs/xfs/scrub/rtsummary.c
45
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary.c
52
rts = kvzalloc_flex(*rts, words, mp->m_blockwsize, XCHK_GFP_FLAGS);
fs/xfs/scrub/rtsummary.c
72
XFS_FSB_TO_B(mp, mp->m_rsumblocks), &sc->xfile);
fs/xfs/scrub/rtsummary_repair.c
103
rts->prep_wordoff += mp->m_blockwsize;
fs/xfs/scrub/rtsummary_repair.c
114
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary_repair.c
118
if (!xfs_has_rmapbt(mp))
fs/xfs/scrub/rtsummary_repair.c
121
if (!xfs_has_exchange_range(mp))
fs/xfs/scrub/rtsummary_repair.c
125
if (rts->rbmblocks != mp->m_sb.sb_rbmblocks)
fs/xfs/scrub/rtsummary_repair.c
160
error = xrep_tempfile_set_isize(sc, XFS_FSB_TO_B(mp, rts->rsumblocks));
fs/xfs/scrub/rtsummary_repair.c
179
memset(sc->sr.rtg->rtg_rsum_cache, 0xFF, mp->m_sb.sb_rbmblocks);
fs/xfs/scrub/rtsummary_repair.c
181
mp->m_rsumlevels = rts->rsumlevels;
fs/xfs/scrub/rtsummary_repair.c
182
mp->m_rsumblocks = rts->rsumblocks;
fs/xfs/scrub/rtsummary_repair.c
38
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary_repair.c
59
blocks = mp->m_rsumblocks;
fs/xfs/scrub/rtsummary_repair.c
60
blocks += xfs_bmbt_calc_size(mp, blocks) * 2;
fs/xfs/scrub/rtsummary_repair.c
75
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/rtsummary_repair.c
79
rts->args.mp = mp;
fs/xfs/scrub/rtsummary_repair.c
86
error = xfsum_copyout(sc, rts->prep_wordoff, ondisk, mp->m_blockwsize);
fs/xfs/scrub/rtsummary_repair.c
90
if (xfs_has_rtgroups(sc->mp)) {
fs/xfs/scrub/rtsummary_repair.c
97
uuid_copy(&hdr->rt_uuid, &sc->mp->m_sb.sb_meta_uuid);
fs/xfs/scrub/scrub.c
495
struct xfs_mount *mp,
fs/xfs/scrub/scrub.c
518
if (ops->has && !ops->has(mp))
fs/xfs/scrub/scrub.c
531
sm->sm_agno >= mp->m_sb.sb_agcount)
fs/xfs/scrub/scrub.c
543
if (xfs_has_rtgroups(mp)) {
fs/xfs/scrub/scrub.c
552
if (sm->sm_agno >= mp->m_sb.sb_rgcount) {
fs/xfs/scrub/scrub.c
584
if (!xfs_has_crc(mp))
fs/xfs/scrub/scrub.c
588
if (xfs_is_readonly(mp))
fs/xfs/scrub/scrub.c
608
xrep_failure(sc->mp);
fs/xfs/scrub/scrub.c
619
xfs_alert_ratelimited(sc->mp,
fs/xfs/scrub/scrub.c
662
struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount;
fs/xfs/scrub/scrub.c
673
if (xfs_is_shutdown(mp))
fs/xfs/scrub/scrub.c
676
if (xfs_has_norecovery(mp))
fs/xfs/scrub/scrub.c
679
error = xchk_validate_inputs(mp, sm);
fs/xfs/scrub/scrub.c
689
sc->mp = mp;
fs/xfs/scrub/scrub.c
756
xrep_failure(mp);
fs/xfs/scrub/scrub.c
769
xchk_stats_merge(mp, sm, &run);
fs/xfs/scrub/scrub.c
827
struct xfs_mount *mp,
fs/xfs/scrub/scrub.c
872
struct xfs_mount *mp,
fs/xfs/scrub/scrub.c
879
tp = xfs_trans_alloc_empty(mp);
fs/xfs/scrub/scrub.c
880
error = xfs_iget(mp, tp, head->svh_ino, XCHK_IGET_FLAGS, 0, &ip);
fs/xfs/scrub/scrub.c
904
struct xfs_mount *mp = ip_in->i_mount;
fs/xfs/scrub/scrub.c
947
trace_xchk_scrubv_item(mp, &head, i, v);
fs/xfs/scrub/scrub.c
959
handle_ip = xchk_scrubv_open_by_handle(mp, &head);
fs/xfs/scrub/scrub.c
972
v->sv_ret = xfs_scrubv_check_barrier(mp, vectors, v);
fs/xfs/scrub/scrub.c
974
trace_xchk_scrubv_barrier_fail(mp, &head, i, v);
fs/xfs/scrub/scrub.c
984
trace_xchk_scrubv_outcome(mp, &head, i, v);
fs/xfs/scrub/scrub.h
137
struct xfs_mount *mp;
fs/xfs/scrub/stats.c
231
struct xfs_mount *mp,
fs/xfs/scrub/stats.c
236
xchk_stats_merge_one(mp->m_scrub_stats, sm, run);
fs/xfs/scrub/stats.c
315
struct xfs_mount *mp)
fs/xfs/scrub/stats.c
387
struct xfs_mount *mp)
fs/xfs/scrub/stats.c
396
error = xchk_stats_init(cs, mp);
fs/xfs/scrub/stats.c
400
mp->m_scrub_stats = cs;
fs/xfs/scrub/stats.c
410
struct xfs_mount *mp)
fs/xfs/scrub/stats.c
412
xchk_stats_teardown(mp->m_scrub_stats);
fs/xfs/scrub/stats.c
413
kvfree(mp->m_scrub_stats);
fs/xfs/scrub/stats.c
414
mp->m_scrub_stats = NULL;
fs/xfs/scrub/stats.h
23
int xchk_mount_stats_alloc(struct xfs_mount *mp);
fs/xfs/scrub/stats.h
24
void xchk_mount_stats_free(struct xfs_mount *mp);
fs/xfs/scrub/stats.h
29
void xchk_stats_merge(struct xfs_mount *mp, const struct xfs_scrub_metadata *sm,
fs/xfs/scrub/stats.h
50
# define xchk_mount_stats_alloc(mp) (0)
fs/xfs/scrub/stats.h
51
# define xchk_mount_stats_free(mp) ((void)0)
fs/xfs/scrub/stats.h
56
# define xchk_stats_merge(mp, sm, run) ((void)0)
fs/xfs/scrub/symlink_repair.c
115
fsblocks = xfs_symlink_blocks(sc->mp, len);
fs/xfs/scrub/symlink_repair.c
123
d = XFS_FSB_TO_DADDR(sc->mp, mval[n].br_startblock);
fs/xfs/scrub/symlink_repair.c
126
error = xfs_trans_read_buf(sc->mp, sc->tp, sc->mp->m_ddev_targp,
fs/xfs/scrub/symlink_repair.c
127
d, XFS_FSB_TO_BB(sc->mp, mval[n].br_blockcount),
fs/xfs/scrub/symlink_repair.c
134
byte_cnt = XFS_FSB_TO_B(sc->mp, mval[n].br_blockcount);
fs/xfs/scrub/symlink_repair.c
135
byte_cnt = XFS_SYMLINK_BUF_SPACE(sc->mp, byte_cnt);
fs/xfs/scrub/symlink_repair.c
259
if (!xfs_has_crc(sc->mp))
fs/xfs/scrub/symlink_repair.c
424
fs_blocks = xfs_symlink_blocks(sc->mp, target_len);
fs/xfs/scrub/symlink_repair.c
425
resblks = xfs_symlink_space_res(sc->mp, target_len, fs_blocks);
fs/xfs/scrub/symlink_repair.c
492
if (!xfs_has_rmapbt(sc->mp))
fs/xfs/scrub/symlink_repair.c
495
if (!xfs_has_exchange_range(sc->mp))
fs/xfs/scrub/symlink_repair.c
59
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/symlink_repair.c
80
blocks = xfs_symlink_blocks(sc->mp, XFS_SYMLINK_MAXLEN);
fs/xfs/scrub/symlink_repair.c
81
blocks += xfs_bmbt_calc_size(mp, blocks) * 2;
fs/xfs/scrub/tempfile.c
423
ASSERT(!XFS_NOT_DQATTACHED(sc->mp, sc->tempip));
fs/xfs/scrub/tempfile.c
45
.pip = sc->mp->m_rootip,
fs/xfs/scrub/tempfile.c
488
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/tempfile.c
49
struct xfs_mount *mp = sc->mp;
fs/xfs/scrub/tempfile.c
492
loff_t pos = XFS_FSB_TO_B(mp, off);
fs/xfs/scrub/tempfile.c
498
flush_mask = XFS_B_TO_FSBT(mp, (1U << 19)) - 1;
fs/xfs/scrub/tempfile.c
500
for (; off < end; off++, pos += mp->m_sb.sb_blocksize) {
fs/xfs/scrub/tempfile.c
514
error = xfs_trans_get_buf(sc->tp, mp->m_ddev_targp,
fs/xfs/scrub/tempfile.c
515
XFS_FSB_TO_DADDR(mp, map.br_startblock),
fs/xfs/scrub/tempfile.c
516
mp->m_bsize, 0, &bp);
fs/xfs/scrub/tempfile.c
55
struct xfs_inode *dp = mp->m_rootip;
fs/xfs/scrub/tempfile.c
61
if (xfs_is_shutdown(mp))
fs/xfs/scrub/tempfile.c
63
if (xfs_is_readonly(mp))
fs/xfs/scrub/tempfile.c
79
resblks = xfs_mkdir_space_res(mp, 0);
fs/xfs/scrub/tempfile.c
80
tres = &M_RES(mp)->tr_mkdir;
fs/xfs/scrub/tempfile.c
82
resblks = XFS_IALLOC_SPACE_RES(mp);
fs/xfs/scrub/tempfile.c
83
tres = &M_RES(mp)->tr_create_tmpfile;
fs/xfs/scrub/tempfile.c
848
ASSERT(xfs_has_exchange_range(sc->mp));
fs/xfs/scrub/tempfile.c
859
if (xfs_has_lazysbcount(sc->mp))
fs/xfs/scrub/tempfile.c
86
error = xfs_trans_alloc_icreate(mp, tres, udqp, gdqp, pdqp, resblks,
fs/xfs/scrub/tempfile.c
862
error = xfs_trans_alloc(sc->mp, &M_RES(sc->mp)->tr_itruncate,
fs/xfs/scrub/tempfile.c
886
ASSERT(xfs_has_exchange_range(sc->mp));
fs/xfs/scrub/tempfile.c
964
struct xfs_mount *mp = ip->i_mount;
fs/xfs/scrub/tempfile.c
973
if (xfs_has_metadir(mp) && (ip->i_diflags2 & XFS_DIFLAG2_METADATA))
fs/xfs/scrub/trace.h
1245
TP_PROTO(struct xfs_mount *mp, xfs_rtxnum_t start,
fs/xfs/scrub/trace.h
1248
TP_ARGS(mp, start, len, log, pos, value),
fs/xfs/scrub/trace.h
1259
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1260
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
1287
__entry->dev = iscan->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1316
__entry->dev = iscan->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1346
__entry->dev = iscan->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1359
TP_PROTO(struct xfs_mount *mp, struct xchk_iscan *iscan,
fs/xfs/scrub/trace.h
1361
TP_ARGS(mp, iscan, nr, avail),
fs/xfs/scrub/trace.h
1373
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1405
__entry->dev = iscan->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1428
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *dp,
fs/xfs/scrub/trace.h
1430
TP_ARGS(mp, dp, ino, name),
fs/xfs/scrub/trace.h
1439
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1454
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *dp,
fs/xfs/scrub/trace.h
1457
TP_ARGS(mp, dp, name, pptr),
fs/xfs/scrub/trace.h
1466
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1481
TP_PROTO(struct xfs_mount *mp, xfs_ino_t ino),
fs/xfs/scrub/trace.h
1482
TP_ARGS(mp, ino),
fs/xfs/scrub/trace.h
1488
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1497
TP_PROTO(struct xfs_mount *mp, const struct xfs_inode *dp,
fs/xfs/scrub/trace.h
1500
TP_ARGS(mp, dp, action, ino, delta, name, namelen),
fs/xfs/scrub/trace.h
1511
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1529
TP_PROTO(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/scrub/trace.h
1531
TP_ARGS(mp, ino, live),
fs/xfs/scrub/trace.h
1540
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1555
TP_PROTO(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/scrub/trace.h
1558
TP_ARGS(mp, ino, live, parents_delta, backrefs_delta, children_delta),
fs/xfs/scrub/trace.h
1570
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1591
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *ip,
fs/xfs/scrub/trace.h
1593
TP_ARGS(mp, ip, live),
fs/xfs/scrub/trace.h
1604
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1623
TP_PROTO(struct xfs_mount *mp, struct xfs_inode *ip, \
fs/xfs/scrub/trace.h
1625
TP_ARGS(mp, ip, live))
fs/xfs/scrub/trace.h
1681
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1727
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1797
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1832
__entry->dev = dl->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1873
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1906
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
1935
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
206
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2345
TP_PROTO(struct xfs_mount *mp, xfs_rgnumber_t rgno,
fs/xfs/scrub/trace.h
2347
TP_ARGS(mp, rgno, usedlen, rmapbt_sz),
fs/xfs/scrub/trace.h
2355
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2369
TP_PROTO(struct xfs_mount *mp, struct xchk_fscounters *fsc),
fs/xfs/scrub/trace.h
2370
TP_ARGS(mp, fsc),
fs/xfs/scrub/trace.h
2379
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2452
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2525
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2579
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2610
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2634
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
265
TP_PROTO(struct xfs_mount *mp, struct xfs_scrub_vec_head *vhead,
fs/xfs/scrub/trace.h
267
TP_ARGS(mp, vhead, vec_nr, v),
fs/xfs/scrub/trace.h
2734
TP_PROTO(struct xfs_mount *mp, uint8_t type, uint32_t id),
fs/xfs/scrub/trace.h
2735
TP_ARGS(mp, type, id),
fs/xfs/scrub/trace.h
2742
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2754
TP_PROTO(struct xfs_mount *mp, uint8_t type, uint32_t id), \
fs/xfs/scrub/trace.h
2755
TP_ARGS(mp, type, id))
fs/xfs/scrub/trace.h
276
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2819
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2852
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2892
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2897
__entry->agno = XFS_FSB_TO_AGNO(sc->mp, irec->br_startblock);
fs/xfs/scrub/trace.h
2898
__entry->agbno = XFS_FSB_TO_AGBNO(sc->mp, irec->br_startblock);
fs/xfs/scrub/trace.h
291
TP_PROTO(struct xfs_mount *mp, struct xfs_scrub_vec_head *vhead, \
fs/xfs/scrub/trace.h
2924
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
2926
__entry->agno = XFS_FSB_TO_AGNO(sc->mp, irec->br_startblock);
fs/xfs/scrub/trace.h
2927
__entry->agbno = XFS_FSB_TO_AGBNO(sc->mp, irec->br_startblock);
fs/xfs/scrub/trace.h
293
TP_ARGS(mp, vhead, vec_nr, v))
fs/xfs/scrub/trace.h
312
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3318
TP_PROTO(struct xfs_mount *mp, xfs_ino_t ino,
fs/xfs/scrub/trace.h
3320
TP_ARGS(mp, ino, obs),
fs/xfs/scrub/trace.h
3329
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3344
TP_PROTO(struct xfs_mount *mp, const struct dentry *dentry),
fs/xfs/scrub/trace.h
3345
TP_ARGS(mp, dentry),
fs/xfs/scrub/trace.h
3356
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3378
TP_PROTO(struct xfs_mount *mp, const struct dentry *dentry), \
fs/xfs/scrub/trace.h
3379
TP_ARGS(mp, dentry))
fs/xfs/scrub/trace.h
3697
TP_PROTO(struct xfs_mount *mp, xfs_rtxnum_t start, xfs_rtxnum_t end),
fs/xfs/scrub/trace.h
3698
TP_ARGS(mp, start, end),
fs/xfs/scrub/trace.h
3706
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3707
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
371
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3719
TP_PROTO(struct xfs_mount *mp, xfs_rtxnum_t start, \
fs/xfs/scrub/trace.h
3721
TP_ARGS(mp, start, end))
fs/xfs/scrub/trace.h
3726
TP_PROTO(struct xfs_mount *mp, unsigned long long wordoff,
fs/xfs/scrub/trace.h
3728
TP_ARGS(mp, wordoff, mask, word),
fs/xfs/scrub/trace.h
373
__entry->agno = xfs_daddr_to_agno(sc->mp, daddr);
fs/xfs/scrub/trace.h
3737
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3738
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
374
__entry->agbno = xfs_daddr_to_agbno(sc->mp, daddr);
fs/xfs/scrub/trace.h
3781
TP_PROTO(struct xfs_mount *mp, xfs_fileoff_t rbmoff,
fs/xfs/scrub/trace.h
3783
TP_ARGS(mp, rbmoff, wordoff, wordcnt),
fs/xfs/scrub/trace.h
3792
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3793
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
3807
TP_PROTO(struct xfs_mount *mp, unsigned long long wordoff,
fs/xfs/scrub/trace.h
3810
TP_ARGS(mp, wordoff, bit, ondisk_word, xfile_word, word_mask),
fs/xfs/scrub/trace.h
3821
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3822
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
3844
TP_PROTO(struct xfs_mount *mp, const struct xfs_rmap_irec *rec),
fs/xfs/scrub/trace.h
3845
TP_ARGS(mp, rec),
fs/xfs/scrub/trace.h
3856
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
3857
__entry->rtdev = mp->m_rtdev_targp->bt_dev;
fs/xfs/scrub/trace.h
405
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
482
__entry->dev = cursor->sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
524
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
546
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
574
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
615
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
657
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
695
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
736
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
774
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/scrub/trace.h
854
TP_PROTO(struct xfs_mount *mp, uint64_t icount, uint64_t ifree,
fs/xfs/scrub/trace.h
856
TP_ARGS(mp, icount, ifree, fdblocks, delalloc),
fs/xfs/scrub/trace.h
868
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
869
__entry->icount_sb = mp->m_sb.sb_icount;
fs/xfs/scrub/trace.h
871
__entry->ifree_sb = mp->m_sb.sb_ifree;
fs/xfs/scrub/trace.h
873
__entry->fdblocks_sb = mp->m_sb.sb_fdblocks;
fs/xfs/scrub/trace.h
889
TP_PROTO(struct xfs_mount *mp, uint64_t expected, int64_t curr_value,
fs/xfs/scrub/trace.h
891
TP_ARGS(mp, expected, curr_value, old_value),
fs/xfs/scrub/trace.h
899
__entry->dev = mp->m_super->s_dev;
fs/xfs/scrub/trace.h
920
__entry->dev = sc->mp->m_super->s_dev;
fs/xfs/xfs_acl.c
132
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_acl.c
137
.valuelen = XFS_ACL_MAX_SIZE(mp),
fs/xfs/xfs_acl.c
164
acl = xfs_acl_from_disk(mp, args.value, args.valuelen,
fs/xfs/xfs_acl.c
165
XFS_ACL_MAX_ENTRIES(mp));
fs/xfs/xfs_acl.c
226
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_acl.c
230
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
fs/xfs/xfs_acl.c
240
if (xfs_has_wsync(mp))
fs/xfs/xfs_acl.c
32
struct xfs_mount *mp,
fs/xfs/xfs_acl.c
43
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, aclp,
fs/xfs/xfs_acl.c
50
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, aclp,
fs/xfs/xfs_aops.c
110
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
127
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_aops.c
220
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
234
WARN_ON_ONCE(!queue_work(mp->m_unwritten_workqueue,
fs/xfs/xfs_aops.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
265
if (xfs_is_shutdown(mp))
fs/xfs/xfs_aops.c
268
xfs_alert_ratelimited(mp,
fs/xfs/xfs_aops.c
330
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
332
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_aops.c
333
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + count);
fs/xfs/xfs_aops.c
342
if (xfs_is_shutdown(mp))
fs/xfs/xfs_aops.c
345
XFS_ERRORTAG_DELAY(mp, XFS_ERRTAG_WB_DELAY_MS);
fs/xfs/xfs_aops.c
471
loff_t cow_offset = XFS_FSB_TO_B(mp, cow_fsb);
fs/xfs/xfs_aops.c
56
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
575
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_aops.c
576
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_aops.c
577
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + len);
fs/xfs/xfs_aops.c
582
if (xfs_is_shutdown(mp))
fs/xfs/xfs_aops.c
585
XFS_ERRORTAG_DELAY(mp, XFS_ERRTAG_WB_DELAY_MS);
fs/xfs/xfs_aops.c
61
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_fsyncts, 0, 0, 0, &tp);
fs/xfs/xfs_aops.c
624
wpc->iomap.bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_aops.c
626
wpc->iomap.length = XFS_FSB_TO_B(mp, count_fsb);
fs/xfs/xfs_attr_inactive.c
137
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_attr_inactive.c
155
xfs_da3_node_hdr_from_disk(mp, &ichdr, bp->b_addr);
fs/xfs/xfs_attr_inactive.c
212
error = xfs_trans_get_buf(*trans, mp->m_ddev_targp,
fs/xfs/xfs_attr_inactive.c
214
XFS_FSB_TO_BB(mp, mp->m_attr_geo->fsbcount), 0,
fs/xfs/xfs_attr_inactive.c
231
xfs_da3_node_hdr_from_disk(mp, &ichdr, bp->b_addr);
fs/xfs/xfs_attr_inactive.c
331
struct xfs_mount *mp;
fs/xfs/xfs_attr_inactive.c
336
mp = dp->i_mount;
fs/xfs/xfs_attr_inactive.c
345
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_attrinval, 0, 0, 0, &trans);
fs/xfs/xfs_attr_inactive.c
376
XFS_FSB_TO_B(mp, mp->m_attr_geo->fsbcount));
fs/xfs/xfs_attr_inactive.c
89
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_attr_inactive.c
96
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr, leaf);
fs/xfs/xfs_attr_item.c
1003
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_attr_item.c
1021
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1027
if (!xfs_attri_validate(mp, attri_formatp)) {
fs/xfs/xfs_attr_item.c
1028
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1040
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1051
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1061
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1073
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1082
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1089
attr_name = xfs_attri_validate_name_iovec(mp, attri_formatp,
fs/xfs/xfs_attr_item.c
1097
attr_new_name = xfs_attri_validate_name_iovec(mp,
fs/xfs/xfs_attr_item.c
1107
attr_value = xfs_attri_validate_value_iovec(mp, attri_formatp,
fs/xfs/xfs_attr_item.c
1116
attr_new_value = xfs_attri_validate_value_iovec(mp,
fs/xfs/xfs_attr_item.c
1129
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
1144
attrip = xfs_attri_init(mp, nv);
fs/xfs/xfs_attr_item.c
272
struct xfs_mount *mp,
fs/xfs/xfs_attr_item.c
286
xfs_log_item_init(mp, &attrip->attri_item, XFS_LI_ATTRI,
fs/xfs/xfs_attr_item.c
415
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_attr_item.c
450
attrip = xfs_attri_init(mp, attr->xattri_nameval);
fs/xfs/xfs_attr_item.c
547
struct xfs_mount *mp,
fs/xfs/xfs_attr_item.c
565
if (!xfs_has_parent(mp))
fs/xfs/xfs_attr_item.c
576
if (!xfs_is_using_logged_xattrs(mp))
fs/xfs/xfs_attr_item.c
584
if (!xfs_is_using_logged_xattrs(mp))
fs/xfs/xfs_attr_item.c
592
if (!xfs_has_parent(mp))
fs/xfs/xfs_attr_item.c
607
return xfs_verify_ino(mp, attrp->alfi_ino);
fs/xfs/xfs_attr_item.c
628
struct xfs_mount *mp,
fs/xfs/xfs_attr_item.c
648
error = xlog_recover_iget_handle(mp, attrp->alfi_ino,
fs/xfs/xfs_attr_item.c
652
error = xlog_recover_iget(mp, attrp->alfi_ino, &ip);
fs/xfs/xfs_attr_item.c
656
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, attrp,
fs/xfs/xfs_attr_item.c
685
args->geo = mp->m_attr_geo;
fs/xfs/xfs_attr_item.c
735
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_attr_item.c
750
if (!xfs_attri_validate(mp, attrp) ||
fs/xfs/xfs_attr_item.c
755
attr = xfs_attri_recover_work(mp, dfp, attrp, &ip, nv);
fs/xfs/xfs_attr_item.c
770
resv = M_RES(mp)->tr_attrrm;
fs/xfs/xfs_attr_item.c
771
total = XFS_ATTRRM_SPACE_RES(mp);
fs/xfs/xfs_attr_item.c
775
error = xfs_trans_alloc(mp, &resv, total, 0, XFS_TRANS_RESERVE, &tp);
fs/xfs/xfs_attr_item.c
785
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
948
struct xfs_mount *mp,
fs/xfs/xfs_attr_item.c
954
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
961
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
963
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
973
struct xfs_mount *mp,
fs/xfs/xfs_attr_item.c
979
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
985
!xfs_parent_valuecheck(mp, iovec->iov_base, value_len)) {
fs/xfs/xfs_attr_item.c
986
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_item.c
988
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_list.c
223
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_attr_list.c
246
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_attr_list.c
255
xfs_da3_node_hdr_from_disk(mp, &nodehdr, node);
fs/xfs/xfs_attr_list.c
284
if (XFS_IS_CORRUPT(mp, cursor->blkno == 0)) {
fs/xfs/xfs_attr_list.c
320
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_attr_list.c
370
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo,
fs/xfs/xfs_attr_list.c
415
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &leafhdr, leaf);
fs/xfs/xfs_attr_list.c
443
struct xfs_mount *mp = context->dp->i_mount;
fs/xfs/xfs_attr_list.c
448
xfs_attr3_leaf_hdr_from_disk(mp->m_attr_geo, &ichdr, leaf);
fs/xfs/xfs_bmap_item.c
144
struct xfs_mount *mp)
fs/xfs/xfs_bmap_item.c
151
xfs_log_item_init(mp, &buip->bui_item, XFS_LI_BUI, &xfs_bui_item_ops);
fs/xfs/xfs_bmap_item.c
297
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_bmap_item.c
298
struct xfs_bui_log_item *buip = xfs_bui_init(mp);
fs/xfs/xfs_bmap_item.c
304
list_sort(mp, items, xfs_bmap_update_diff_items);
fs/xfs/xfs_bmap_item.c
332
struct xfs_mount *mp,
fs/xfs/xfs_bmap_item.c
347
bi->bi_group = xfs_group_intent_get(mp, bi->bi_bmap.br_startblock,
fs/xfs/xfs_bmap_item.c
420
struct xfs_mount *mp,
fs/xfs/xfs_bmap_item.c
442
if (!xfs_verify_ino(mp, map->me_owner))
fs/xfs/xfs_bmap_item.c
445
if (!xfs_verify_fileext(mp, map->me_startoff, map->me_len))
fs/xfs/xfs_bmap_item.c
449
return xfs_verify_rtbext(mp, map->me_startblock, map->me_len);
fs/xfs/xfs_bmap_item.c
451
return xfs_verify_fsbext(mp, map->me_startblock, map->me_len);
fs/xfs/xfs_bmap_item.c
456
struct xfs_mount *mp,
fs/xfs/xfs_bmap_item.c
464
error = xlog_recover_iget(mp, map->me_owner, ipp);
fs/xfs/xfs_bmap_item.c
479
xfs_bmap_update_get_group(mp, bi);
fs/xfs/xfs_bmap_item.c
502
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_bmap_item.c
508
if (!xfs_bui_validate(mp, buip)) {
fs/xfs/xfs_bmap_item.c
509
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_bmap_item.c
515
work = xfs_bui_recover_work(mp, dfp, &ip, map);
fs/xfs/xfs_bmap_item.c
520
resv = xlog_recover_resv(&M_RES(mp)->tr_itruncate);
fs/xfs/xfs_bmap_item.c
521
error = xfs_trans_alloc(mp, &resv,
fs/xfs/xfs_bmap_item.c
522
XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK), 0, 0, &tp);
fs/xfs/xfs_bmap_item.c
546
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_bmap_item.c
650
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_bmap_item.c
658
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_bmap_item.c
664
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_bmap_item.c
671
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_bmap_item.c
676
buip = xfs_bui_init(mp);
fs/xfs/xfs_bmap_util.c
1010
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, 0, 0, 0, &tp);
fs/xfs/xfs_bmap_util.c
104
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
1059
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
1062
xfs_fileoff_t stop_fsb = XFS_B_TO_FSB(mp, offset);
fs/xfs/xfs_bmap_util.c
1064
xfs_fileoff_t shift_fsb = XFS_B_TO_FSB(mp, len);
fs/xfs/xfs_bmap_util.c
1079
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write,
fs/xfs/xfs_bmap_util.c
1080
XFS_DIOSTRAT_SPACE_RES(mp, 0), 0, 0, &tp);
fs/xfs/xfs_bmap_util.c
122
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
fs/xfs/xfs_bmap_util.c
1512
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
1582
if (xfs_has_rmapbt(mp)) {
fs/xfs/xfs_bmap_util.c
1592
resblks = XFS_SWAP_RMAP_SPACE_RES(mp, ipnext, w);
fs/xfs/xfs_bmap_util.c
1593
resblks += XFS_SWAP_RMAP_SPACE_RES(mp, tipnext, w);
fs/xfs/xfs_bmap_util.c
1604
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, resblks, 0, flags,
fs/xfs/xfs_bmap_util.c
1632
xfs_notice(mp,
fs/xfs/xfs_bmap_util.c
1665
if (xfs_has_rmapbt(mp))
fs/xfs/xfs_bmap_util.c
1684
if (xfs_has_reflink(mp)) {
fs/xfs/xfs_bmap_util.c
1726
if (xfs_has_wsync(mp))
fs/xfs/xfs_bmap_util.c
254
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
305
max_len = mp->m_super->s_maxbytes;
fs/xfs/xfs_bmap_util.c
328
max_len = mp->m_super->s_maxbytes;
fs/xfs/xfs_bmap_util.c
351
max_len = XFS_FSB_TO_BB(mp, XFS_B_TO_FSB(mp, max_len));
fs/xfs/xfs_bmap_util.c
357
first_bno = bno = XFS_BB_TO_FSBT(mp, bmv->bmv_offset);
fs/xfs/xfs_bmap_util.c
358
len = XFS_BB_TO_FSB(mp, bmv->bmv_length);
fs/xfs/xfs_bmap_util.c
371
XFS_B_TO_FSB(mp, XFS_ISIZE(ip)));
fs/xfs/xfs_bmap_util.c
404
xfs_fileoff_t end = XFS_B_TO_FSB(mp, XFS_ISIZE(ip));
fs/xfs/xfs_bmap_util.c
443
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
445
xfs_fileoff_t start_fsb = XFS_B_TO_FSBT(mp, start_byte);
fs/xfs/xfs_bmap_util.c
446
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, end_byte);
fs/xfs/xfs_bmap_util.c
506
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
549
end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)XFS_ISIZE(ip));
fs/xfs/xfs_bmap_util.c
551
end_fsb = xfs_fileoff_roundup_rtx(mp, end_fsb);
fs/xfs/xfs_bmap_util.c
552
last_fsb = XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes);
fs/xfs/xfs_bmap_util.c
580
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
600
round_up(XFS_ISIZE(ip), mp->m_sb.sb_blocksize),
fs/xfs/xfs_bmap_util.c
607
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, 0, 0, 0, &tp);
fs/xfs/xfs_bmap_util.c
609
ASSERT(xfs_is_shutdown(mp));
fs/xfs/xfs_bmap_util.c
651
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
667
if (xfs_is_shutdown(mp))
fs/xfs/xfs_bmap_util.c
682
startoffset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_bmap_util.c
683
endoffset_fsb = XFS_B_TO_FSB(mp, offset + count);
fs/xfs/xfs_bmap_util.c
724
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_bmap_util.c
727
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, resblks);
fs/xfs/xfs_bmap_util.c
731
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_bmap_util.c
784
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
786
uint resblks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_bmap_util.c
789
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, resblks, 0,
fs/xfs/xfs_bmap_util.c
847
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
869
startoffset_fsb = XFS_B_TO_FSB(mp, offset);
fs/xfs/xfs_bmap_util.c
870
endoffset_fsb = XFS_B_TO_FSBT(mp, offset + len);
fs/xfs/xfs_bmap_util.c
874
startoffset_fsb = xfs_fileoff_roundup_rtx(mp, startoffset_fsb);
fs/xfs/xfs_bmap_util.c
875
endoffset_fsb = xfs_fileoff_rounddown_rtx(mp, endoffset_fsb);
fs/xfs/xfs_bmap_util.c
991
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_bmap_util.c
994
xfs_fileoff_t next_fsb = XFS_B_TO_FSB(mp, offset + len);
fs/xfs/xfs_bmap_util.c
995
xfs_fileoff_t shift_fsb = XFS_B_TO_FSB(mp, len);
fs/xfs/xfs_bmap_util.h
48
int xfs_bmap_extsize_align(struct xfs_mount *mp, struct xfs_bmbt_irec *gotp,
fs/xfs/xfs_buf.c
1052
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_buf.c
1062
if (xfs_is_unmounting(mp) && mp->m_fail_unmount)
fs/xfs/xfs_buf.c
1089
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_buf.c
1097
if (xlog_is_shutdown(mp->m_log))
fs/xfs/xfs_buf.c
1107
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_buf.c
1119
cfg = xfs_error_get_cfg(mp, XFS_ERR_METADATA, bp->b_error);
fs/xfs/xfs_buf.c
1134
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_buf.c
1695
struct xfs_mount *mp = btp->bt_mount;
fs/xfs/xfs_buf.c
1706
min_bytes > mp->m_sb.sb_blocksize ||
fs/xfs/xfs_buf.c
1707
max_bytes < mp->m_sb.sb_blocksize) {
fs/xfs/xfs_buf.c
1723
struct xfs_mount *mp = btp->bt_mount;
fs/xfs/xfs_buf.c
1730
xfs_warn(mp,
fs/xfs/xfs_buf.c
1743
btp->bt_nr_sectors = nr_blocks << (mp->m_sb.sb_blocklog - BBSHIFT);
fs/xfs/xfs_buf.c
1791
struct xfs_mount *mp,
fs/xfs/xfs_buf.c
1804
btp->bt_mount = mp;
fs/xfs/xfs_buf.c
1809
mp, ops);
fs/xfs/xfs_buf.c
1827
mp->m_super->s_id);
fs/xfs/xfs_buf.c
2107
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_buf.c
2110
idx = xfs_has_crc(mp);
fs/xfs/xfs_buf.c
2125
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_buf.c
2128
idx = xfs_has_crc(mp);
fs/xfs/xfs_buf.c
545
struct xfs_mount *mp = btp->bt_mount;
fs/xfs/xfs_buf.c
549
return xfs_perag_get(mp, xfs_daddr_to_agno(mp, map->bm_bn));
fs/xfs/xfs_buf.h
372
struct xfs_buftarg *xfs_alloc_buftarg(struct xfs_mount *mp,
fs/xfs/xfs_buf_item.c
811
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_buf_item.c
823
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_buf_item.c
853
struct xfs_mount *mp)
fs/xfs/xfs_buf_item.c
865
ASSERT(bp->b_mount == mp);
fs/xfs/xfs_buf_item.c
874
xfs_log_item_init(mp, &bip->bli_item, XFS_LI_BUF, &xfs_buf_item_ops);
fs/xfs/xfs_buf_item.c
896
xfs_err(mp,
fs/xfs/xfs_buf_item_recover.c
1016
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_buf_item_recover.c
1037
error = xfs_buf_read(mp->m_ddev_targp, buf_f->blf_blkno, buf_f->blf_len,
fs/xfs/xfs_buf_item_recover.c
1061
lsn = xlog_recover_get_buf_lsn(mp, bp, buf_f);
fs/xfs/xfs_buf_item_recover.c
1064
xlog_recover_validate_buf_type(mp, bp, buf_f, NULLCOMMITLSN);
fs/xfs/xfs_buf_item_recover.c
1079
error = xlog_recover_do_inode_buffer(mp, item, bp, buf_f);
fs/xfs/xfs_buf_item_recover.c
1086
dirty = xlog_recover_do_dquot_buffer(mp, log, item, bp, buf_f);
fs/xfs/xfs_buf_item_recover.c
1091
error = xlog_recover_do_primary_sb_buffer(mp, item, bp, buf_f,
fs/xfs/xfs_buf_item_recover.c
1097
if (xfs_has_rtsb(mp) && mp->m_rtsb_bp) {
fs/xfs/xfs_buf_item_recover.c
1098
struct xfs_buf *rtsb_bp = mp->m_rtsb_bp;
fs/xfs/xfs_buf_item_recover.c
1108
xlog_recover_do_reg_buffer(mp, item, bp, buf_f, current_lsn);
fs/xfs/xfs_buf_item_recover.c
1141
ASSERT(bp->b_mount == mp);
fs/xfs/xfs_buf_item_recover.c
217
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
236
if (!xfs_has_crc(mp))
fs/xfs/xfs_buf_item_recover.c
313
xfs_alert(mp,
fs/xfs/xfs_buf_item_recover.c
404
if (xfs_has_rtgroups(mp) && magic32 != XFS_RTBITMAP_MAGIC) {
fs/xfs/xfs_buf_item_recover.c
408
bp->b_ops = xfs_rtblock_ops(mp, XFS_RTGI_BITMAP);
fs/xfs/xfs_buf_item_recover.c
411
if (xfs_has_rtgroups(mp) && magic32 != XFS_RTSUMMARY_MAGIC) {
fs/xfs/xfs_buf_item_recover.c
415
bp->b_ops = xfs_rtblock_ops(mp, XFS_RTGI_SUMMARY);
fs/xfs/xfs_buf_item_recover.c
419
xfs_warn(mp, "Unknown buffer type %d!",
fs/xfs/xfs_buf_item_recover.c
433
xfs_warn(mp, warnmsg);
fs/xfs/xfs_buf_item_recover.c
452
xfs_buf_item_init(bp, mp);
fs/xfs/xfs_buf_item_recover.c
466
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
478
trace_xfs_log_recover_buf_reg_buf(mp->m_log, buf_f);
fs/xfs/xfs_buf_item_recover.c
515
xfs_alert(mp,
fs/xfs/xfs_buf_item_recover.c
520
xfs_alert(mp,
fs/xfs/xfs_buf_item_recover.c
525
fa = xfs_dquot_verify(mp, item->ri_buf[i].iov_base, -1);
fs/xfs/xfs_buf_item_recover.c
527
xfs_alert(mp,
fs/xfs/xfs_buf_item_recover.c
546
xlog_recover_validate_buf_type(mp, bp, buf_f, current_lsn);
fs/xfs/xfs_buf_item_recover.c
560
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
573
if (!mp->m_qflags)
fs/xfs/xfs_buf_item_recover.c
589
xlog_recover_do_reg_buffer(mp, item, bp, buf_f, NULLCOMMITLSN);
fs/xfs/xfs_buf_item_recover.c
607
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
623
trace_xfs_log_recover_buf_inode_buf(mp->m_log, buf_f);
fs/xfs/xfs_buf_item_recover.c
629
if (xfs_has_crc(mp))
fs/xfs/xfs_buf_item_recover.c
632
inodes_per_buf = BBTOB(bp->b_length) >> mp->m_sb.sb_inodelog;
fs/xfs/xfs_buf_item_recover.c
634
next_unlinked_offset = (i * mp->m_sb.sb_inodesize) +
fs/xfs/xfs_buf_item_recover.c
683
if (XFS_IS_CORRUPT(mp, *logged_nextp == 0)) {
fs/xfs/xfs_buf_item_recover.c
684
xfs_alert(mp,
fs/xfs/xfs_buf_item_recover.c
699
xfs_dinode_calc_crc(mp,
fs/xfs/xfs_buf_item_recover.c
700
xfs_buf_offset(bp, i * mp->m_sb.sb_inodesize));
fs/xfs/xfs_buf_item_recover.c
716
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
723
xfs_agnumber_t orig_agcount = mp->m_sb.sb_agcount;
fs/xfs/xfs_buf_item_recover.c
724
xfs_rgnumber_t orig_rgcount = mp->m_sb.sb_rgcount;
fs/xfs/xfs_buf_item_recover.c
727
xlog_recover_do_reg_buffer(mp, item, bp, buf_f, current_lsn);
fs/xfs/xfs_buf_item_recover.c
730
xfs_alert(mp, "Trying to grow file system without AGs");
fs/xfs/xfs_buf_item_recover.c
737
xfs_sb_from_disk(&mp->m_sb, dsb);
fs/xfs/xfs_buf_item_recover.c
742
mp->m_ddev_targp->bt_nr_sectors =
fs/xfs/xfs_buf_item_recover.c
743
XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks);
fs/xfs/xfs_buf_item_recover.c
744
if (mp->m_rtdev_targp && mp->m_rtdev_targp != mp->m_ddev_targp) {
fs/xfs/xfs_buf_item_recover.c
745
mp->m_rtdev_targp->bt_nr_sectors =
fs/xfs/xfs_buf_item_recover.c
746
XFS_FSB_TO_BB(mp, mp->m_sb.sb_rblocks);
fs/xfs/xfs_buf_item_recover.c
749
if (mp->m_sb.sb_agcount < orig_agcount) {
fs/xfs/xfs_buf_item_recover.c
750
xfs_alert(mp, "Shrinking AG count in log recovery not supported");
fs/xfs/xfs_buf_item_recover.c
753
if (mp->m_sb.sb_rgcount < orig_rgcount) {
fs/xfs/xfs_buf_item_recover.c
754
xfs_warn(mp,
fs/xfs/xfs_buf_item_recover.c
763
error = xfs_update_last_ag_size(mp, orig_agcount);
fs/xfs/xfs_buf_item_recover.c
773
error = xfs_update_last_rtgroup_size(mp, orig_rgcount);
fs/xfs/xfs_buf_item_recover.c
784
error = xfs_initialize_perag(mp, orig_agcount, mp->m_sb.sb_agcount,
fs/xfs/xfs_buf_item_recover.c
785
mp->m_sb.sb_dblocks, &mp->m_maxagi);
fs/xfs/xfs_buf_item_recover.c
787
xfs_warn(mp, "Failed recovery per-ag init: %d", error);
fs/xfs/xfs_buf_item_recover.c
790
mp->m_alloc_set_aside = xfs_alloc_set_aside(mp);
fs/xfs/xfs_buf_item_recover.c
792
error = xfs_initialize_rtgroups(mp, orig_rgcount, mp->m_sb.sb_rgcount,
fs/xfs/xfs_buf_item_recover.c
793
mp->m_sb.sb_rextents);
fs/xfs/xfs_buf_item_recover.c
795
xfs_warn(mp, "Failed recovery rtgroup init: %d", error);
fs/xfs/xfs_buf_item_recover.c
823
struct xfs_mount *mp,
fs/xfs/xfs_buf_item_recover.c
836
if (!xfs_has_crc(mp))
fs/xfs/xfs_buf_item_recover.c
844
if (!xfs_has_rtgroups(mp) && (blft == XFS_BLFT_RTBITMAP_BUF ||
fs/xfs/xfs_buf_item_recover.c
924
if (xfs_has_metauuid(mp))
fs/xfs/xfs_buf_item_recover.c
934
if (!uuid_equal(&mp->m_sb.sb_meta_uuid, uuid))
fs/xfs/xfs_buf_item_recover.c
953
if (!uuid_equal(&mp->m_sb.sb_meta_uuid, uuid))
fs/xfs/xfs_buf_mem.c
52
struct xfs_mount *mp,
fs/xfs/xfs_buf_mem.c
89
btp->bt_mount = mp;
fs/xfs/xfs_buf_mem.h
18
int xmbuf_alloc(struct xfs_mount *mp, const char *descr,
fs/xfs/xfs_dir2_readdir.c
112
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/xfs_dir2_readdir.c
116
ino = xfs_dir2_sf_get_ino(mp, sfp, sfep);
fs/xfs/xfs_dir2_readdir.c
117
filetype = xfs_dir2_sf_get_ftype(mp, sfep);
fs/xfs/xfs_dir2_readdir.c
126
xfs_dir3_get_dtype(mp, filetype)))
fs/xfs/xfs_dir2_readdir.c
128
sfep = xfs_dir2_sf_nextentry(mp, sfp, sfep);
fs/xfs/xfs_dir2_readdir.c
33
struct xfs_mount *mp,
fs/xfs/xfs_dir2_readdir.c
355
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_dir2_readdir.c
36
if (!xfs_has_ftype(mp))
fs/xfs/xfs_dir2_readdir.c
433
length = xfs_dir2_data_entsize(mp,
fs/xfs/xfs_dir2_readdir.c
465
length = xfs_dir2_data_entsize(mp, dep->namelen);
fs/xfs/xfs_dir2_readdir.c
466
filetype = xfs_dir2_data_get_ftype(mp, dep);
fs/xfs/xfs_dir2_readdir.c
52
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_dir2_readdir.c
80
xfs_dir2_data_entsize(mp, sizeof(".") - 1));
fs/xfs/xfs_discard.c
113
struct xfs_mount *mp,
fs/xfs/xfs_discard.c
130
XFS_FSB_TO_BB(mp, busyp->length),
fs/xfs/xfs_discard.c
166
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_discard.c
179
xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_discard.c
181
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_discard.c
197
cur = xfs_bnobt_init_cursor(mp, tp, agbp, pag);
fs/xfs/xfs_discard.c
203
cur = xfs_cntbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/xfs_discard.c
207
cur = xfs_cntbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/xfs_discard.c
229
if (XFS_IS_CORRUPT(mp, i != 1)) {
fs/xfs/xfs_discard.c
390
struct xfs_mount *mp,
fs/xfs/xfs_discard.c
402
XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks) - 1);
fs/xfs/xfs_discard.c
404
start_agno = xfs_daddr_to_agno(mp, start);
fs/xfs/xfs_discard.c
405
start_agbno = xfs_daddr_to_agbno(mp, start);
fs/xfs/xfs_discard.c
406
end_agno = xfs_daddr_to_agno(mp, ddev_end);
fs/xfs/xfs_discard.c
407
end_agbno = xfs_daddr_to_agbno(mp, ddev_end);
fs/xfs/xfs_discard.c
409
while ((pag = xfs_perag_next_range(mp, pag, start_agno, end_agno))) {
fs/xfs/xfs_discard.c
468
struct xfs_mount *mp,
fs/xfs/xfs_discard.c
471
struct block_device *bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_discard.c
484
trace_xfs_discard_rtextent(mp, busyp->bno, busyp->length);
fs/xfs/xfs_discard.c
487
xfs_rtb_to_daddr(mp, busyp->bno),
fs/xfs/xfs_discard.c
488
XFS_FSB_TO_BB(mp, busyp->length),
fs/xfs/xfs_discard.c
498
xfs_info(mp,
fs/xfs/xfs_discard.c
561
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_discard.c
563
.minlen_fsb = XFS_BB_TO_FSB(mp, minlen),
fs/xfs/xfs_discard.c
569
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_discard.c
576
tr.stop_rtx = low + xfs_rtbitmap_rtx_per_rbmblock(mp);
fs/xfs/xfs_discard.c
594
error = xfs_discard_rtdev_extents(mp, &tr);
fs/xfs/xfs_discard.c
678
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_discard.c
680
.minlen_fsb = XFS_BB_TO_FSB(mp, minlen),
fs/xfs/xfs_discard.c
685
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_discard.c
740
struct xfs_mount *mp,
fs/xfs/xfs_discard.c
753
daddr_offset = XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks);
fs/xfs/xfs_discard.c
758
start_rtbno = xfs_daddr_to_rtb(mp, start);
fs/xfs/xfs_discard.c
759
start_rtx = xfs_rtb_to_rtx(mp, start_rtbno);
fs/xfs/xfs_discard.c
760
start_rgno = xfs_rtb_to_rgno(mp, start_rtbno);
fs/xfs/xfs_discard.c
766
end_rtbno = xfs_daddr_to_rtb(mp, end);
fs/xfs/xfs_discard.c
767
end_rtx = xfs_rtb_to_rtx(mp, end_rtbno + mp->m_sb.sb_rextsize - 1);
fs/xfs/xfs_discard.c
768
end_rgno = xfs_rtb_to_rgno(mp, end_rtbno);
fs/xfs/xfs_discard.c
770
while ((rtg = xfs_rtgroup_next_range(mp, rtg, start_rgno, end_rgno))) {
fs/xfs/xfs_discard.c
776
if (xfs_has_rtgroups(mp))
fs/xfs/xfs_discard.c
812
struct xfs_mount *mp,
fs/xfs/xfs_discard.c
816
bdev_discard_granularity(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_discard.c
827
if (mp->m_rtdev_targp && !xfs_has_zoned(mp) &&
fs/xfs/xfs_discard.c
828
bdev_max_discard_sectors(mp->m_rtdev_targp->bt_bdev))
fs/xfs/xfs_discard.c
829
rt_bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_discard.c
830
if (!bdev_max_discard_sectors(mp->m_ddev_targp->bt_bdev) && !rt_bdev)
fs/xfs/xfs_discard.c
841
if (xfs_has_norecovery(mp))
fs/xfs/xfs_discard.c
848
minlen = XFS_B_TO_FSB(mp, range.minlen);
fs/xfs/xfs_discard.c
857
max_blocks = mp->m_sb.sb_dblocks + mp->m_sb.sb_rblocks;
fs/xfs/xfs_discard.c
858
if (range.start >= XFS_FSB_TO_B(mp, max_blocks) ||
fs/xfs/xfs_discard.c
859
range.minlen > XFS_FSB_TO_B(mp, mp->m_ag_max_usable) ||
fs/xfs/xfs_discard.c
860
range.len < mp->m_sb.sb_blocksize)
fs/xfs/xfs_discard.c
866
if (bdev_max_discard_sectors(mp->m_ddev_targp->bt_bdev)) {
fs/xfs/xfs_discard.c
867
error = xfs_trim_datadev_extents(mp, start, end, minlen);
fs/xfs/xfs_discard.c
873
error = xfs_trim_rtdev_extents(mp, start, end, minlen);
fs/xfs/xfs_discard.c
882
XFS_FSB_TO_B(mp, max_blocks) - range.start);
fs/xfs/xfs_discard.h
10
int xfs_ioc_trim(struct xfs_mount *mp, struct fstrim_range __user *fstrim);
fs/xfs/xfs_discard.h
9
void xfs_discard_extents(struct xfs_mount *mp, struct xfs_busy_extents *busy);
fs/xfs/xfs_dquot.c
1006
error = xfs_qm_dqget_checks(mp, type);
fs/xfs/xfs_dquot.c
1017
dqp = xfs_qm_dqget_cache_lookup(mp, id, type);
fs/xfs/xfs_dquot.c
1029
error = xfs_qm_dqread(mp, id, type, can_alloc, &dqp);
fs/xfs/xfs_dquot.c
1038
if (xfs_this_quota_on(mp, type)) {
fs/xfs/xfs_dquot.c
1053
error = xfs_qm_dqget_cache_insert(mp, id, type, dqp);
fs/xfs/xfs_dquot.c
1061
XFS_STATS_INC(mp, xs_qm_dquot_dups);
fs/xfs/xfs_dquot.c
1082
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
1091
for (; !error; error = xfs_dq_get_next_id(mp, type, &id)) {
fs/xfs/xfs_dquot.c
1092
error = xfs_qm_dqget(mp, id, type, false, &dqp);
fs/xfs/xfs_dquot.c
122
struct xfs_mount *mp = dq->q_mount;
fs/xfs/xfs_dquot.c
123
struct xfs_quotainfo *q = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
1257
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_dquot.c
1261
error = xfs_trans_read_buf(mp, tp, mp->m_ddev_targp, dqp->q_blkno,
fs/xfs/xfs_dquot.c
1262
mp->m_quotainfo->qi_dqchunklen, 0,
fs/xfs/xfs_dquot.c
1275
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_dquot.c
1367
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_dquot.c
1381
xfs_alert(mp, "corrupt dquot ID 0x%x in memory at %pS",
fs/xfs/xfs_dquot.c
1405
xfs_trans_ail_copy_lsn(mp->m_ail, &qlip->qli_flush_lsn, &lip->li_lsn);
fs/xfs/xfs_dquot.c
1416
if (xfs_has_crc(mp)) {
fs/xfs/xfs_dquot.c
1435
xfs_log_force(mp, 0);
fs/xfs/xfs_dquot.c
1448
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_dquot.c
154
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
157
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
177
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
186
res->timer = xfs_dquot_set_timeout(mp,
fs/xfs/xfs_dquot.c
210
struct xfs_mount *mp = dq->q_mount;
fs/xfs/xfs_dquot.c
211
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
232
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_dquot.c
233
struct xfs_quotainfo *q = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
273
if (curid > 0 && xfs_has_bigtime(mp))
fs/xfs/xfs_dquot.c
275
if (xfs_has_crc(mp)) {
fs/xfs/xfs_dquot.c
276
uuid_copy(&d->dd_uuid, &mp->m_sb.sb_meta_uuid);
fs/xfs/xfs_dquot.c
300
if (!(mp->m_qflags & qflag))
fs/xfs/xfs_dquot.c
349
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_dquot.c
352
struct xfs_inode *quotip = xfs_quota_inode(mp, qtype);
fs/xfs/xfs_dquot.c
358
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_qm_dqalloc,
fs/xfs/xfs_dquot.c
359
XFS_QM_DQALLOC_SPACE_RES(mp), 0, 0, &tp);
fs/xfs/xfs_dquot.c
394
dqp->q_blkno = XFS_FSB_TO_DADDR(mp, map.br_startblock);
fs/xfs/xfs_dquot.c
397
error = xfs_trans_get_buf(tp, mp->m_ddev_targp, dqp->q_blkno,
fs/xfs/xfs_dquot.c
398
mp->m_quotainfo->qi_dqchunklen, 0, &bp);
fs/xfs/xfs_dquot.c
457
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
464
struct xfs_inode *quotip = xfs_quota_inode(mp, qtype);
fs/xfs/xfs_dquot.c
470
if (!xfs_this_quota_on(mp, qtype)) {
fs/xfs/xfs_dquot.c
500
dqp->q_blkno = XFS_FSB_TO_DADDR(mp, map.br_startblock);
fs/xfs/xfs_dquot.c
502
error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp, dqp->q_blkno,
fs/xfs/xfs_dquot.c
503
mp->m_quotainfo->qi_dqchunklen, 0, &bp,
fs/xfs/xfs_dquot.c
522
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
53
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_dquot.c
532
dqp->q_mount = mp;
fs/xfs/xfs_dquot.c
536
dqp->q_fileoffset = (xfs_fileoff_t)id / mp->m_quotainfo->qi_dqperchunk;
fs/xfs/xfs_dquot.c
540
dqp->q_bufoffset = (id % mp->m_quotainfo->qi_dqperchunk) *
fs/xfs/xfs_dquot.c
57
xfs_fs_mark_sick(mp, XFS_SICK_FS_UQUOTA);
fs/xfs/xfs_dquot.c
572
XFS_STATS_INC(mp, xs_qm_dquot);
fs/xfs/xfs_dquot.c
60
xfs_fs_mark_sick(mp, XFS_SICK_FS_GQUOTA);
fs/xfs/xfs_dquot.c
63
xfs_fs_mark_sick(mp, XFS_SICK_FS_PQUOTA);
fs/xfs/xfs_dquot.c
704
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
714
dqp = xfs_dquot_alloc(mp, id, type);
fs/xfs/xfs_dquot.c
718
error = xfs_dquot_disk_read(mp, dqp, &bp);
fs/xfs/xfs_dquot.c
753
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
757
struct xfs_inode *quotip = xfs_quota_inode(mp, type);
fs/xfs/xfs_dquot.c
770
if (next_id % mp->m_quotainfo->qi_dqperchunk) {
fs/xfs/xfs_dquot.c
776
start = (xfs_fsblock_t)next_id / mp->m_quotainfo->qi_dqperchunk;
fs/xfs/xfs_dquot.c
787
*id = got.br_startoff * mp->m_quotainfo->qi_dqperchunk;
fs/xfs/xfs_dquot.c
803
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
807
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
816
XFS_STATS_INC(mp, xs_qm_dqcachemisses);
fs/xfs/xfs_dquot.c
829
XFS_STATS_INC(mp, xs_qm_dqcachehits);
fs/xfs/xfs_dquot.c
846
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
851
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_dquot.c
876
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
881
if (!XFS_IS_UQUOTA_ON(mp))
fs/xfs/xfs_dquot.c
885
if (!XFS_IS_GQUOTA_ON(mp))
fs/xfs/xfs_dquot.c
889
if (!XFS_IS_PQUOTA_ON(mp))
fs/xfs/xfs_dquot.c
904
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
913
error = xfs_qm_dqget_checks(mp, type);
fs/xfs/xfs_dquot.c
918
dqp = xfs_qm_dqget_cache_lookup(mp, id, type);
fs/xfs/xfs_dquot.c
922
error = xfs_qm_dqread(mp, id, type, can_alloc, &dqp);
fs/xfs/xfs_dquot.c
926
error = xfs_qm_dqget_cache_insert(mp, id, type, dqp);
fs/xfs/xfs_dquot.c
934
XFS_STATS_INC(mp, xs_qm_dquot_dups);
fs/xfs/xfs_dquot.c
954
struct xfs_mount *mp,
fs/xfs/xfs_dquot.c
961
error = xfs_qm_dqget_checks(mp, type);
fs/xfs/xfs_dquot.c
965
return xfs_qm_dqread(mp, id, type, 0, dqpp);
fs/xfs/xfs_dquot.c
998
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_dquot.h
130
static inline int xfs_this_quota_on(struct xfs_mount *mp, xfs_dqtype_t type)
fs/xfs/xfs_dquot.h
134
return XFS_IS_UQUOTA_ON(mp);
fs/xfs/xfs_dquot.h
136
return XFS_IS_GQUOTA_ON(mp);
fs/xfs/xfs_dquot.h
138
return XFS_IS_PQUOTA_ON(mp);
fs/xfs/xfs_dquot.h
211
int xfs_qm_dqget(struct xfs_mount *mp, xfs_dqid_t id,
fs/xfs/xfs_dquot.h
216
int xfs_qm_dqget_next(struct xfs_mount *mp, xfs_dqid_t id,
fs/xfs/xfs_dquot.h
218
int xfs_qm_dqget_uncached(struct xfs_mount *mp,
fs/xfs/xfs_dquot.h
237
time64_t xfs_dquot_set_timeout(struct xfs_mount *mp, time64_t timeout);
fs/xfs/xfs_dquot_item.c
218
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_dquot_item.c
223
fa = xfs_dquot_verify(mp, &ddq, dqp->q_id);
fs/xfs/xfs_dquot_item.c
226
XFS_ERRLEVEL_LOW, mp, &ddq, sizeof(ddq));
fs/xfs/xfs_dquot_item.c
227
xfs_alert(mp,
fs/xfs/xfs_dquot_item.c
230
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_dquot_item_recover.c
113
fa = xfs_dquot_verify(mp, recddq, dq_f->qlf_id);
fs/xfs/xfs_dquot_item_recover.c
115
xfs_alert(mp, "corrupt dquot ID 0x%x in log at %pS",
fs/xfs/xfs_dquot_item_recover.c
128
error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp, dq_f->qlf_blkno,
fs/xfs/xfs_dquot_item_recover.c
129
XFS_FSB_TO_BB(mp, dq_f->qlf_len), 0, &bp,
fs/xfs/xfs_dquot_item_recover.c
142
if (xfs_has_crc(mp)) {
fs/xfs/xfs_dquot_item_recover.c
151
if (xfs_has_crc(mp)) {
fs/xfs/xfs_dquot_item_recover.c
160
XFS_ERRLEVEL_LOW, mp, dqb,
fs/xfs/xfs_dquot_item_recover.c
162
xfs_alert(mp,
fs/xfs/xfs_dquot_item_recover.c
170
ASSERT(bp->b_mount == mp);
fs/xfs/xfs_dquot_item_recover.c
29
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_dquot_item_recover.c
34
if (mp->m_qflags == 0)
fs/xfs/xfs_dquot_item_recover.c
53
XFS_FSB_TO_BB(mp, dq_f->qlf_len),
fs/xfs/xfs_dquot_item_recover.c
67
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_dquot_item_recover.c
79
if (mp->m_qflags == 0)
fs/xfs/xfs_drain.c
105
struct xfs_mount *mp,
fs/xfs/xfs_drain.c
111
xg = xfs_group_get_by_fsb(mp, fsbno, type);
fs/xfs/xfs_drain.h
68
struct xfs_group *xfs_group_intent_get(struct xfs_mount *mp,
fs/xfs/xfs_error.c
122
struct xfs_mount *mp)
fs/xfs/xfs_error.c
124
return xfs_sysfs_init(&mp->m_errortag_kobj, &xfs_errortag_ktype,
fs/xfs/xfs_error.c
125
&mp->m_kobj, "errortag");
fs/xfs/xfs_error.c
130
struct xfs_mount *mp)
fs/xfs/xfs_error.c
132
xfs_sysfs_del(&mp->m_errortag_kobj);
fs/xfs/xfs_error.c
137
struct xfs_mount *mp,
fs/xfs/xfs_error.c
144
randfactor = READ_ONCE(mp->m_errortag[error_tag]);
fs/xfs/xfs_error.c
148
xfs_warn_ratelimited(mp,
fs/xfs/xfs_error.c
150
file, line, mp->m_super->s_id);
fs/xfs/xfs_error.c
156
struct xfs_mount *mp,
fs/xfs/xfs_error.c
161
unsigned int delay = READ_ONCE(mp->m_errortag[error_tag]);
fs/xfs/xfs_error.c
168
xfs_warn_ratelimited(mp,
fs/xfs/xfs_error.c
171
mp->m_super->s_id);
fs/xfs/xfs_error.c
177
struct xfs_mount *mp,
fs/xfs/xfs_error.c
193
WRITE_ONCE(mp->m_errortag[error_tag],
fs/xfs/xfs_error.c
200
struct xfs_mount *mp,
fs/xfs/xfs_error.c
208
return xfs_errortag_add(mp, i);
fs/xfs/xfs_error.c
230
struct xfs_mount *mp)
fs/xfs/xfs_error.c
235
WRITE_ONCE(mp->m_errortag[i], 0);
fs/xfs/xfs_error.c
244
struct xfs_mount *mp,
fs/xfs/xfs_error.c
250
xfs_alert_tag(mp, XFS_PTAG_ERROR_REPORT,
fs/xfs/xfs_error.c
262
struct xfs_mount *mp,
fs/xfs/xfs_error.c
271
xfs_error_report(tag, level, mp, filename, linenum, failaddr);
fs/xfs/xfs_error.c
272
xfs_alert(mp, "Corruption detected. Unmount and run xfs_repair");
fs/xfs/xfs_error.c
287
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_error.c
289
xfs_alert_tag(mp, XFS_PTAG_VERIFIER_ERROR,
fs/xfs/xfs_error.c
293
xfs_alert(mp, "Unmount and run xfs_repair");
fs/xfs/xfs_error.c
312
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_error.c
319
xfs_alert_tag(mp, XFS_PTAG_VERIFIER_ERROR,
fs/xfs/xfs_error.c
324
xfs_alert(mp, "Unmount and run xfs_repair");
fs/xfs/xfs_error.c
328
xfs_alert(mp, "First %d bytes of corrupted metadata buffer:",
fs/xfs/xfs_error.c
364
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_error.c
370
xfs_alert(mp, "Metadata %s detected at %pS, inode 0x%llx %s",
fs/xfs/xfs_error.c
374
xfs_alert(mp, "Unmount and run xfs_repair");
fs/xfs/xfs_error.c
378
xfs_alert(mp, "First %d bytes of corrupted metadata buffer:",
fs/xfs/xfs_error.c
57
struct xfs_mount *mp = to_mp(kobject);
fs/xfs/xfs_error.c
70
WRITE_ONCE(mp->m_errortag[error_tag], val);
fs/xfs/xfs_error.c
80
struct xfs_mount *mp = to_mp(kobject);
fs/xfs/xfs_error.c
83
READ_ONCE(mp->m_errortag[to_attr(attr)->tag]));
fs/xfs/xfs_error.h
11
void xfs_error_report(const char *tag, int level, struct xfs_mount *mp,
fs/xfs/xfs_error.h
13
void xfs_corruption_error(const char *tag, int level, struct xfs_mount *mp,
fs/xfs/xfs_error.h
23
#define XFS_ERROR_REPORT(e, lvl, mp) \
fs/xfs/xfs_error.h
24
xfs_error_report(e, lvl, mp, __FILE__, __LINE__, __return_address)
fs/xfs/xfs_error.h
25
#define XFS_CORRUPTION_ERROR(e, lvl, mp, buf, bufsize) \
fs/xfs/xfs_error.h
26
xfs_corruption_error(e, lvl, mp, buf, bufsize, \
fs/xfs/xfs_error.h
37
int xfs_errortag_init(struct xfs_mount *mp);
fs/xfs/xfs_error.h
38
void xfs_errortag_del(struct xfs_mount *mp);
fs/xfs/xfs_error.h
39
bool xfs_errortag_test(struct xfs_mount *mp, const char *file, int line,
fs/xfs/xfs_error.h
41
#define XFS_TEST_ERROR(mp, tag) \
fs/xfs/xfs_error.h
42
xfs_errortag_test((mp), __FILE__, __LINE__, (tag))
fs/xfs/xfs_error.h
43
void xfs_errortag_delay(struct xfs_mount *mp, const char *file, int line,
fs/xfs/xfs_error.h
45
#define XFS_ERRORTAG_DELAY(mp, tag) \
fs/xfs/xfs_error.h
46
xfs_errortag_delay((mp), __FILE__, __LINE__, (tag))
fs/xfs/xfs_error.h
47
int xfs_errortag_add(struct xfs_mount *mp, unsigned int error_tag);
fs/xfs/xfs_error.h
48
int xfs_errortag_add_name(struct xfs_mount *mp, const char *tag_name);
fs/xfs/xfs_error.h
50
int xfs_errortag_clearall(struct xfs_mount *mp);
fs/xfs/xfs_error.h
52
#define xfs_errortag_init(mp) (0)
fs/xfs/xfs_error.h
53
#define xfs_errortag_del(mp)
fs/xfs/xfs_error.h
54
#define XFS_TEST_ERROR(mp, tag) (false)
fs/xfs/xfs_error.h
55
#define XFS_ERRORTAG_DELAY(mp, tag) ((void)0)
fs/xfs/xfs_error.h
56
#define xfs_errortag_add(mp, tag) (-ENOSYS)
fs/xfs/xfs_error.h
58
#define xfs_errortag_add_name(mp, tag_name) (-ENOSYS)
fs/xfs/xfs_error.h
59
#define xfs_errortag_clearall(mp) (-ENOSYS)
fs/xfs/xfs_exchmaps_item.c
130
struct xfs_mount *mp)
fs/xfs/xfs_exchmaps_item.c
137
xfs_log_item_init(mp, &xmi_lip->xmi_item, XFS_LI_XMI, &xfs_xmi_item_ops);
fs/xfs/xfs_exchmaps_item.c
326
struct xfs_mount *mp,
fs/xfs/xfs_exchmaps_item.c
331
if (!xfs_has_exchange_range(mp))
fs/xfs/xfs_exchmaps_item.c
340
if (!xfs_verify_ino(mp, xlf->xmi_inode1) ||
fs/xfs/xfs_exchmaps_item.c
341
!xfs_verify_ino(mp, xlf->xmi_inode2))
fs/xfs/xfs_exchmaps_item.c
344
if (!xfs_verify_fileext(mp, xlf->xmi_startoff1, xlf->xmi_blockcount))
fs/xfs/xfs_exchmaps_item.c
347
return xfs_verify_fileext(mp, xlf->xmi_startoff2, xlf->xmi_blockcount);
fs/xfs/xfs_exchmaps_item.c
356
struct xfs_mount *mp,
fs/xfs/xfs_exchmaps_item.c
374
error = xlog_recover_iget_handle(mp, xlf->xmi_inode1, xlf->xmi_igen1,
fs/xfs/xfs_exchmaps_item.c
377
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, xlf,
fs/xfs/xfs_exchmaps_item.c
382
error = xlog_recover_iget_handle(mp, xlf->xmi_inode2, xlf->xmi_igen2,
fs/xfs/xfs_exchmaps_item.c
385
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp, xlf,
fs/xfs/xfs_exchmaps_item.c
428
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_exchmaps_item.c
433
if (!xfs_xmi_validate(mp, xmi_lip)) {
fs/xfs/xfs_exchmaps_item.c
434
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_exchmaps_item.c
440
xmi = xfs_xmi_item_recover_intent(mp, dfp, &xmi_lip->xmi_format, &req,
fs/xfs/xfs_exchmaps_item.c
445
trace_xfs_exchmaps_recover(mp, xmi);
fs/xfs/xfs_exchmaps_item.c
447
resv = xlog_recover_resv(&M_RES(mp)->tr_write);
fs/xfs/xfs_exchmaps_item.c
448
error = xfs_trans_alloc(mp, &resv, req.resblks, 0, 0, &tp);
fs/xfs/xfs_exchmaps_item.c
458
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_exchmaps_item.c
552
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_exchmaps_item.c
569
xmi_lip = xfs_xmi_init(mp);
fs/xfs/xfs_exchrange.c
197
struct xfs_mount *mp = ip1->i_mount;
fs/xfs/xfs_exchrange.c
201
.startoff1 = XFS_B_TO_FSBT(mp, fxr->file1_offset),
fs/xfs/xfs_exchrange.c
202
.startoff2 = XFS_B_TO_FSBT(mp, fxr->file2_offset),
fs/xfs/xfs_exchrange.c
203
.blockcount = XFS_B_TO_FSB(mp, fxr->length),
fs/xfs/xfs_exchrange.c
223
req.blockcount = xfs_blen_roundup_rtx(mp, req.blockcount);
fs/xfs/xfs_exchrange.c
231
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, req.resblks, 0,
fs/xfs/xfs_exchrange.c
241
error = xfs_exchmaps_check_forks(mp, &req);
fs/xfs/xfs_exchrange.c
281
if (xfs_has_wsync(mp) || (fxr->flags & XFS_EXCHANGE_RANGE_DSYNC))
fs/xfs/xfs_exchrange.c
592
struct xfs_mount *mp = ip2->i_mount;
fs/xfs/xfs_exchrange.c
612
alloc_unit = mp->m_sb.sb_blocksize;
fs/xfs/xfs_exchrange.c
677
struct xfs_mount *mp = ip1->i_mount;
fs/xfs/xfs_exchrange.c
680
if (!xfs_has_exchange_range(mp))
fs/xfs/xfs_exchrange.c
687
if (xfs_is_shutdown(mp))
fs/xfs/xfs_exchrange.c
893
struct xfs_mount *mp = ip2->i_mount;
fs/xfs/xfs_exchrange.c
903
if (memcmp(&kern_f->fsid, mp->m_fixedfsid, sizeof(xfs_fsid_t)))
fs/xfs/xfs_export.c
111
xfs_mount_t *mp = XFS_M(sb);
fs/xfs/xfs_export.c
126
error = xfs_iget(mp, NULL, ino, XFS_IGET_UNTRUSTED, 0, &ip);
fs/xfs/xfs_export.c
157
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_export.c
47
struct xfs_mount *mp = XFS_M(inode->i_sb);
fs/xfs/xfs_export.c
67
if (!xfs_has_small_inums(mp) || xfs_is_inode32(mp))
fs/xfs/xfs_extent_busy.c
665
struct xfs_mount *mp)
fs/xfs/xfs_extent_busy.c
670
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/xfs_extent_busy.c
673
if (xfs_has_rtgroups(mp) && !xfs_has_zoned(mp))
fs/xfs/xfs_extent_busy.c
674
while ((rtg = xfs_rtgroup_next(mp, rtg)))
fs/xfs/xfs_extent_busy.h
60
void xfs_extent_busy_wait_all(struct xfs_mount *mp);
fs/xfs/xfs_extent_busy.h
76
#define xfs_group_has_extent_busy(mp, type) \
fs/xfs/xfs_extent_busy.h
77
((type) == XG_TYPE_AG || !xfs_has_zoned((mp)))
fs/xfs/xfs_extfree_item.c
150
struct xfs_mount *mp,
fs/xfs/xfs_extfree_item.c
167
xfs_log_item_init(mp, &efip->efi_item, item_type, &xfs_efi_item_ops);
fs/xfs/xfs_extfree_item.c
423
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_extfree_item.c
429
efip = xfs_efi_init(mp, item_type, count);
fs/xfs/xfs_extfree_item.c
431
list_sort(mp, items, xfs_extent_free_diff_items);
fs/xfs/xfs_extfree_item.c
501
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_extfree_item.c
503
xefi->xefi_group = xfs_group_intent_get(mp, xefi->xefi_startblock,
fs/xfs/xfs_extfree_item.c
506
trace_xfs_extent_free_defer(mp, xefi);
fs/xfs/xfs_extfree_item.c
532
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_extfree_item.c
536
agbno = XFS_FSB_TO_AGBNO(mp, xefi->xefi_startblock);
fs/xfs/xfs_extfree_item.c
544
trace_xfs_extent_free_deferred(mp, xefi);
fs/xfs/xfs_extfree_item.c
586
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_extfree_item.c
594
agbno = XFS_FSB_TO_AGBNO(mp, xefi->xefi_startblock);
fs/xfs/xfs_extfree_item.c
597
trace_xfs_agfl_free_deferred(mp, xefi);
fs/xfs/xfs_extfree_item.c
612
struct xfs_mount *mp,
fs/xfs/xfs_extfree_item.c
617
return xfs_verify_rtbext(mp, extp->ext_start, extp->ext_len);
fs/xfs/xfs_extfree_item.c
619
return xfs_verify_fsbext(mp, extp->ext_start, extp->ext_len);
fs/xfs/xfs_extfree_item.c
624
struct xfs_mount *mp,
fs/xfs/xfs_extfree_item.c
637
xefi->xefi_group = xfs_group_intent_get(mp, extp->ext_start,
fs/xfs/xfs_extfree_item.c
657
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_extfree_item.c
669
if (!xfs_efi_validate_ext(mp, isrt,
fs/xfs/xfs_extfree_item.c
671
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_extfree_item.c
677
xfs_efi_recover_work(mp, dfp, isrt,
fs/xfs/xfs_extfree_item.c
681
resv = xlog_recover_resv(&M_RES(mp)->tr_itruncate);
fs/xfs/xfs_extfree_item.c
682
error = xfs_trans_alloc(mp, &resv, 0, 0, 0, &tp);
fs/xfs/xfs_extfree_item.c
688
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_extfree_item.c
774
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_extfree_item.c
780
trace_xfs_extent_free_deferred(mp, xefi);
fs/xfs/xfs_extfree_item.c
788
if (xfs_has_zoned(mp))
fs/xfs/xfs_extfree_item.c
798
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_extfree_item.c
864
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_extfree_item.c
872
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_extfree_item.c
877
efip = xfs_efi_init(mp, ITEM_TYPE(item), efi_formatp->efi_nextents);
fs/xfs/xfs_extfree_item.c
903
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_extfree_item.c
911
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_extfree_item.c
916
efip = xfs_efi_init(mp, ITEM_TYPE(item), efi_formatp->efi_nextents);
fs/xfs/xfs_file.c
1062
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
1068
ret = xfs_zoned_write_space_reserve(mp, iocb, from, XFS_ZR_GREEDY, &ac);
fs/xfs/xfs_file.c
1087
XFS_FSB_TO_B(mp, ac.reserved_blocks) -
fs/xfs/xfs_file.c
1088
(iocb->ki_pos & mp->m_blockmask));
fs/xfs/xfs_file.c
1102
xfs_flush_inodes(mp);
fs/xfs/xfs_file.c
1112
XFS_STATS_ADD(mp, xs_write_bytes, ret);
fs/xfs/xfs_file.c
132
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
142
if (xfs_is_shutdown(mp))
fs/xfs/xfs_file.c
1487
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
1499
XFS_TEST_ERROR(mp, XFS_ERRTAG_FORCE_ZERO_RANGE))
fs/xfs/xfs_file.c
1500
count_fsb += XFS_B_TO_FSB(mp, len) + 1;
fs/xfs/xfs_file.c
1502
error = xfs_zoned_space_reserve(mp, count_fsb, XFS_ZR_RESERVED, &ac);
fs/xfs/xfs_file.c
1506
xfs_zoned_space_unreserve(mp, &ac);
fs/xfs/xfs_file.c
153
if (XFS_IS_REALTIME_INODE(ip) && mp->m_rtdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_file.c
154
error = blkdev_issue_flush(mp->m_rtdev_targp->bt_bdev);
fs/xfs/xfs_file.c
155
else if (mp->m_logdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_file.c
156
error = blkdev_issue_flush(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_file.c
1577
struct xfs_mount *mp = src->i_mount;
fs/xfs/xfs_file.c
1585
if (!xfs_has_reflink(mp))
fs/xfs/xfs_file.c
1588
if (xfs_is_shutdown(mp))
fs/xfs/xfs_file.c
1687
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
1693
if (xfs_is_readonly(mp) || xfs_is_shutdown(mp))
fs/xfs/xfs_file.c
176
mp->m_logdev_targp == mp->m_ddev_targp) {
fs/xfs/xfs_file.c
177
err2 = blkdev_issue_flush(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_file.c
327
struct xfs_mount *mp = XFS_I(inode)->i_mount;
fs/xfs/xfs_file.c
330
XFS_STATS_INC(mp, xs_read_calls);
fs/xfs/xfs_file.c
332
if (xfs_is_shutdown(mp))
fs/xfs/xfs_file.c
343
XFS_STATS_ADD(mp, xs_read_bytes, ret);
fs/xfs/xfs_file.c
357
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_file.c
360
XFS_STATS_INC(mp, xs_read_calls);
fs/xfs/xfs_file.c
362
if (xfs_is_shutdown(mp))
fs/xfs/xfs_file.c
371
XFS_STATS_ADD(mp, xs_read_bytes, ret);
fs/xfs/xfs_file.c
523
struct xfs_mount *mp,
fs/xfs/xfs_file.c
559
return xfs_zoned_space_reserve(mp, XFS_B_TO_FSB(mp, count) + 1 + 2,
fs/xfs/xfs_file.c
666
struct xfs_mount *mp = XFS_I(iter->inode)->i_mount;
fs/xfs/xfs_file.c
671
count_fsb = XFS_B_TO_FSB(mp, bio->bi_iter.bi_size);
fs/xfs/xfs_file.c
673
xfs_err(mp,
fs/xfs/xfs_file.c
676
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_filestream.c
151
for_each_perag_wrap(args->mp, 0, start_agno, pag) {
fs/xfs/xfs_filestream.c
209
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_filestream.c
215
mru = xfs_mru_cache_lookup(mp->m_filestream, pino);
fs/xfs/xfs_filestream.c
226
xfs_mru_cache_done(mp->m_filestream);
fs/xfs/xfs_filestream.c
265
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_filestream.c
268
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, pino);
fs/xfs/xfs_filestream.c
273
mru = xfs_mru_cache_remove(mp->m_filestream, pino);
fs/xfs/xfs_filestream.c
278
agno = (pag_agno(item->pag) + 1) % mp->m_sb.sb_agcount;
fs/xfs/xfs_filestream.c
279
xfs_fstrm_free_func(mp, mru);
fs/xfs/xfs_filestream.c
280
} else if (xfs_is_inode32(mp)) {
fs/xfs/xfs_filestream.c
283
agno = (mp->m_agfrotor / rotorstep) % mp->m_sb.sb_agcount;
fs/xfs/xfs_filestream.c
284
mp->m_agfrotor = (mp->m_agfrotor + 1) %
fs/xfs/xfs_filestream.c
285
(mp->m_sb.sb_agcount * rotorstep);
fs/xfs/xfs_filestream.c
288
ap->blkno = XFS_AGB_TO_FSB(args->mp, agno, 0);
fs/xfs/xfs_filestream.c
317
xfs_mru_cache_insert(mp->m_filestream, pino, &item->mru);
fs/xfs/xfs_filestream.c
378
xfs_mount_t *mp)
fs/xfs/xfs_filestream.c
387
return xfs_mru_cache_create(&mp->m_filestream, mp,
fs/xfs/xfs_filestream.c
393
xfs_mount_t *mp)
fs/xfs/xfs_filestream.c
395
xfs_mru_cache_destroy(mp->m_filestream);
fs/xfs/xfs_filestream.c
63
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_filestream.c
72
minfree = mp->m_sb.sb_agblocks / 50;
fs/xfs/xfs_filestream.c
75
for_each_perag_wrap(mp, start_agno, agno, pag) {
fs/xfs/xfs_filestream.h
14
int xfs_filestream_mount(struct xfs_mount *mp);
fs/xfs/xfs_filestream.h
15
void xfs_filestream_unmount(struct xfs_mount *mp);
fs/xfs/xfs_fsmap.c
1046
struct xfs_mount *mp,
fs/xfs/xfs_fsmap.c
1049
if (mp->m_sb.sb_rtstart)
fs/xfs/xfs_fsmap.c
1054
return new_encode_dev(mp->m_ddev_targp->bt_dev);
fs/xfs/xfs_fsmap.c
1056
return new_encode_dev(mp->m_logdev_targp->bt_dev);
fs/xfs/xfs_fsmap.c
1058
if (!mp->m_rtdev_targp)
fs/xfs/xfs_fsmap.c
1060
return new_encode_dev(mp->m_rtdev_targp->bt_dev);
fs/xfs/xfs_fsmap.c
1069
struct xfs_mount *mp,
fs/xfs/xfs_fsmap.c
1074
fm->fmr_device == xfs_getfsmap_device(mp, XFS_DEV_DATA) ||
fs/xfs/xfs_fsmap.c
1075
fm->fmr_device == xfs_getfsmap_device(mp, XFS_DEV_LOG) ||
fs/xfs/xfs_fsmap.c
1076
(mp->m_rtdev_targp &&
fs/xfs/xfs_fsmap.c
1077
fm->fmr_device == xfs_getfsmap_device(mp, XFS_DEV_RT));
fs/xfs/xfs_fsmap.c
1154
struct xfs_mount *mp,
fs/xfs/xfs_fsmap.c
1171
if (!xfs_getfsmap_is_valid_device(mp, &head->fmh_keys[0]) ||
fs/xfs/xfs_fsmap.c
1172
!xfs_getfsmap_is_valid_device(mp, &head->fmh_keys[1]))
fs/xfs/xfs_fsmap.c
1177
use_rmap = xfs_has_rmapbt(mp) &&
fs/xfs/xfs_fsmap.c
1183
handlers[0].nr_sectors = XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks);
fs/xfs/xfs_fsmap.c
1184
handlers[0].dev = xfs_getfsmap_device(mp, XFS_DEV_DATA);
fs/xfs/xfs_fsmap.c
1189
if (mp->m_logdev_targp != mp->m_ddev_targp) {
fs/xfs/xfs_fsmap.c
1190
handlers[1].nr_sectors = XFS_FSB_TO_BB(mp,
fs/xfs/xfs_fsmap.c
1191
mp->m_sb.sb_logblocks);
fs/xfs/xfs_fsmap.c
1192
handlers[1].dev = xfs_getfsmap_device(mp, XFS_DEV_LOG);
fs/xfs/xfs_fsmap.c
1200
if (mp->m_rtdev_targp && (use_rmap || !xfs_has_zoned(mp))) {
fs/xfs/xfs_fsmap.c
1201
handlers[2].nr_sectors = XFS_FSB_TO_BB(mp, mp->m_sb.sb_rblocks);
fs/xfs/xfs_fsmap.c
1202
handlers[2].dev = xfs_getfsmap_device(mp, XFS_DEV_RT);
fs/xfs/xfs_fsmap.c
1273
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_fsmap.c
1295
if (!mp->m_sb.sb_rtstart)
fs/xfs/xfs_fsmap.c
213
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
220
if (!xfs_has_reflink(mp) || !info->group)
fs/xfs/xfs_fsmap.c
226
cur = xfs_refcountbt_init_cursor(mp, tp, info->agf_bp,
fs/xfs/xfs_fsmap.c
231
XFS_BB_TO_FSBT(mp, frec->len_daddr), &fbno, &flen,
fs/xfs/xfs_fsmap.c
244
struct xfs_mount *mp,
fs/xfs/xfs_fsmap.c
250
trace_xfs_getfsmap_mapping(mp, xfm);
fs/xfs/xfs_fsmap.c
288
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
332
xfs_getfsmap_format(mp, &fmr, info);
fs/xfs/xfs_fsmap.c
342
trace_xfs_fsmap_mapping(mp, info->dev,
fs/xfs/xfs_fsmap.c
351
fmr.fmr_offset = XFS_FSB_TO_BB(mp, frec->offset);
fs/xfs/xfs_fsmap.c
367
xfs_getfsmap_format(mp, &fmr, info);
fs/xfs/xfs_fsmap.c
453
rmap_not_shareable(struct xfs_mount *mp, const struct xfs_rmap_irec *r)
fs/xfs/xfs_fsmap.c
455
if (!xfs_has_reflink(mp))
fs/xfs/xfs_fsmap.c
477
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
486
eofs = XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks);
fs/xfs/xfs_fsmap.c
489
start_fsb = XFS_DADDR_TO_FSB(mp, keys[0].fmr_physical);
fs/xfs/xfs_fsmap.c
490
end_fsb = XFS_DADDR_TO_FSB(mp, min(eofs - 1, keys[1].fmr_physical));
fs/xfs/xfs_fsmap.c
497
info->low.rm_offset = XFS_BB_TO_FSBT(mp, keys[0].fmr_offset);
fs/xfs/xfs_fsmap.c
501
info->low.rm_blockcount = XFS_BB_TO_FSBT(mp, keys[0].fmr_length);
fs/xfs/xfs_fsmap.c
507
} else if (rmap_not_shareable(mp, &info->low)) {
fs/xfs/xfs_fsmap.c
513
if (XFS_FSB_TO_DADDR(mp, start_fsb) >= eofs)
fs/xfs/xfs_fsmap.c
519
info->low.rm_startblock = XFS_FSB_TO_AGBNO(mp, start_fsb);
fs/xfs/xfs_fsmap.c
527
start_ag = XFS_FSB_TO_AGNO(mp, start_fsb);
fs/xfs/xfs_fsmap.c
528
end_ag = XFS_FSB_TO_AGNO(mp, end_fsb);
fs/xfs/xfs_fsmap.c
530
while ((pag = xfs_perag_next_range(mp, pag, start_ag, end_ag))) {
fs/xfs/xfs_fsmap.c
537
info->high.rm_startblock = XFS_FSB_TO_AGBNO(mp,
fs/xfs/xfs_fsmap.c
539
info->high.rm_offset = XFS_BB_TO_FSBT(mp,
fs/xfs/xfs_fsmap.c
558
trace_xfs_fsmap_low_group_key(mp, info->dev, pag_agno(pag),
fs/xfs/xfs_fsmap.c
560
trace_xfs_fsmap_high_group_key(mp, info->dev, pag_agno(pag),
fs/xfs/xfs_fsmap.c
687
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
691
eofs = XFS_FSB_TO_BB(mp, mp->m_sb.sb_logblocks);
fs/xfs/xfs_fsmap.c
694
start_fsb = XFS_BB_TO_FSBT(mp,
fs/xfs/xfs_fsmap.c
696
end_fsb = XFS_BB_TO_FSB(mp, min(eofs - 1, keys[1].fmr_physical));
fs/xfs/xfs_fsmap.c
700
info->low_daddr = XFS_FSB_TO_BB(mp, start_fsb);
fs/xfs/xfs_fsmap.c
702
trace_xfs_fsmap_low_linear_key(mp, info->dev, start_fsb);
fs/xfs/xfs_fsmap.c
703
trace_xfs_fsmap_high_linear_key(mp, info->dev, end_fsb);
fs/xfs/xfs_fsmap.c
709
frec.len_daddr = XFS_FSB_TO_BB(mp, mp->m_sb.sb_logblocks);
fs/xfs/xfs_fsmap.c
725
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_fsmap.c
730
xfs_rtbxlen_to_blen(mp, rec->ar_extcount);
fs/xfs/xfs_fsmap.c
744
frec.start_daddr = xfs_rtb_to_daddr(mp, start_rtb);
fs/xfs/xfs_fsmap.c
746
frec.len_daddr = XFS_FSB_TO_BB(mp, rtbcount);
fs/xfs/xfs_fsmap.c
757
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
765
eofs = XFS_FSB_TO_BB(mp, mp->m_sb.sb_rblocks);
fs/xfs/xfs_fsmap.c
772
start_rtbno = xfs_daddr_to_rtb(mp,
fs/xfs/xfs_fsmap.c
775
info->low_daddr = xfs_rtb_to_daddr(mp, start_rtbno);
fs/xfs/xfs_fsmap.c
779
start_rtx = xfs_rtb_to_rtx(mp, start_rtbno);
fs/xfs/xfs_fsmap.c
780
start_rgno = xfs_rtb_to_rgno(mp, start_rtbno);
fs/xfs/xfs_fsmap.c
782
end_rtbno = xfs_daddr_to_rtb(mp, min(eofs - 1, keys[1].fmr_physical));
fs/xfs/xfs_fsmap.c
783
end_rgno = xfs_rtb_to_rgno(mp, end_rtbno);
fs/xfs/xfs_fsmap.c
785
trace_xfs_fsmap_low_linear_key(mp, info->dev, start_rtbno);
fs/xfs/xfs_fsmap.c
786
trace_xfs_fsmap_high_linear_key(mp, info->dev, end_rtbno);
fs/xfs/xfs_fsmap.c
790
while ((rtg = xfs_rtgroup_next_range(mp, rtg, start_rgno, end_rgno))) {
fs/xfs/xfs_fsmap.c
792
end_rtx = xfs_rtb_to_rtx(mp,
fs/xfs/xfs_fsmap.c
793
end_rtbno + mp->m_sb.sb_rextsize - 1);
fs/xfs/xfs_fsmap.c
880
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsmap.c
890
eofs = XFS_FSB_TO_BB(mp, mp->m_sb.sb_rtstart + mp->m_sb.sb_rblocks);
fs/xfs/xfs_fsmap.c
902
rtstart_daddr = XFS_FSB_TO_BB(mp, mp->m_sb.sb_rtstart);
fs/xfs/xfs_fsmap.c
903
if (xfs_has_zoned(mp) && key0.fmr_physical < rtstart_daddr) {
fs/xfs/xfs_fsmap.c
929
start_rtb = xfs_daddr_to_rtb(mp, key0.fmr_physical);
fs/xfs/xfs_fsmap.c
930
end_rtb = xfs_daddr_to_rtb(mp, min(eofs - 1, keys[1].fmr_physical));
fs/xfs/xfs_fsmap.c
938
info->low.rm_offset = XFS_BB_TO_FSBT(mp, key0.fmr_offset);
fs/xfs/xfs_fsmap.c
942
info->low.rm_blockcount = XFS_BB_TO_FSBT(mp, key0.fmr_length);
fs/xfs/xfs_fsmap.c
948
} else if (rmap_not_shareable(mp, &info->low)) {
fs/xfs/xfs_fsmap.c
954
if (xfs_rtb_to_daddr(mp, start_rtb) >= eofs)
fs/xfs/xfs_fsmap.c
960
info->low.rm_startblock = xfs_rtb_to_rgbno(mp, start_rtb);
fs/xfs/xfs_fsmap.c
968
start_rg = xfs_rtb_to_rgno(mp, start_rtb);
fs/xfs/xfs_fsmap.c
969
end_rg = xfs_rtb_to_rgno(mp, end_rtb);
fs/xfs/xfs_fsmap.c
971
while ((rtg = xfs_rtgroup_next_range(mp, rtg, start_rg, end_rg))) {
fs/xfs/xfs_fsmap.c
979
xfs_rtb_to_rgbno(mp, end_rtb);
fs/xfs/xfs_fsmap.c
981
XFS_BB_TO_FSBT(mp, keys[1].fmr_offset);
fs/xfs/xfs_fsmap.c
996
trace_xfs_fsmap_low_group_key(mp, info->dev, rtg_rgno(rtg),
fs/xfs/xfs_fsmap.c
998
trace_xfs_fsmap_high_group_key(mp, info->dev, rtg_rgno(rtg),
fs/xfs/xfs_fsops.c
109
error = xfs_sb_validate_fsb_count(&mp->m_sb, nb);
fs/xfs/xfs_fsops.c
113
if (nb > mp->m_sb.sb_dblocks) {
fs/xfs/xfs_fsops.c
114
error = xfs_buf_read_uncached(mp->m_ddev_targp,
fs/xfs/xfs_fsops.c
115
XFS_FSB_TO_BB(mp, nb) - XFS_FSS_TO_BB(mp, 1),
fs/xfs/xfs_fsops.c
116
XFS_FSS_TO_BB(mp, 1), &bp, NULL);
fs/xfs/xfs_fsops.c
123
error = xfs_growfs_check_rtgeom(mp, nb, mp->m_sb.sb_rblocks,
fs/xfs/xfs_fsops.c
124
mp->m_sb.sb_rextsize);
fs/xfs/xfs_fsops.c
127
xfs_growfs_compute_deltas(mp, nb, &delta, &nagcount);
fs/xfs/xfs_fsops.c
146
error = xfs_initialize_perag(mp, oagcount, nagcount, nb, &nagimax);
fs/xfs/xfs_fsops.c
151
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_growdata,
fs/xfs/xfs_fsops.c
152
XFS_GROWFS_SPACE_RES(mp), 0, XFS_TRANS_RESERVE,
fs/xfs/xfs_fsops.c
155
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_growdata, -delta, 0,
fs/xfs/xfs_fsops.c
160
last_pag = xfs_perag_get(mp, oagcount - 1);
fs/xfs/xfs_fsops.c
165
xfs_warn_experimental(mp, XFS_EXPERIMENTAL_SHRINK);
fs/xfs/xfs_fsops.c
189
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_fsops.c
199
mp->m_maxagi = nagimax;
fs/xfs/xfs_fsops.c
200
xfs_set_low_space_thresholds(mp);
fs/xfs/xfs_fsops.c
201
mp->m_alloc_set_aside = xfs_alloc_set_aside(mp);
fs/xfs/xfs_fsops.c
211
pag = xfs_perag_get(mp, id.agno);
fs/xfs/xfs_fsops.c
220
error = xfs_fs_reserve_ag_blocks(mp);
fs/xfs/xfs_fsops.c
225
xfs_rtrmapbt_compute_maxlevels(mp);
fs/xfs/xfs_fsops.c
226
xfs_rtrefcountbt_compute_maxlevels(mp);
fs/xfs/xfs_fsops.c
235
xfs_free_perag_range(mp, oagcount, nagcount);
fs/xfs/xfs_fsops.c
241
struct xfs_mount *mp, /* mount point for filesystem */
fs/xfs/xfs_fsops.c
247
if (nb < XFS_MIN_LOG_BLOCKS || nb < XFS_B_TO_FSB(mp, XFS_MIN_LOG_BYTES))
fs/xfs/xfs_fsops.c
249
if (nb == mp->m_sb.sb_logblocks &&
fs/xfs/xfs_fsops.c
250
in->isint == (mp->m_sb.sb_logstart != 0))
fs/xfs/xfs_fsops.c
263
struct xfs_mount *mp,
fs/xfs/xfs_fsops.c
273
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_growdata,
fs/xfs/xfs_fsops.c
274
XFS_GROWFS_SPACE_RES(mp), 0, XFS_TRANS_RESERVE, &tp);
fs/xfs/xfs_fsops.c
278
dpct = imaxpct - mp->m_sb.sb_imax_pct;
fs/xfs/xfs_fsops.c
291
struct xfs_mount *mp,
fs/xfs/xfs_fsops.c
298
if (!mutex_trylock(&mp->m_growlock))
fs/xfs/xfs_fsops.c
302
if (in->newblocks != mp->m_sb.sb_dblocks && mp->m_sb.sb_rtstart) {
fs/xfs/xfs_fsops.c
308
if (in->imaxpct != mp->m_sb.sb_imax_pct) {
fs/xfs/xfs_fsops.c
309
error = xfs_growfs_imaxpct(mp, in->imaxpct);
fs/xfs/xfs_fsops.c
314
if (in->newblocks != mp->m_sb.sb_dblocks) {
fs/xfs/xfs_fsops.c
315
error = xfs_growfs_data_private(mp, in);
fs/xfs/xfs_fsops.c
321
if (mp->m_sb.sb_imax_pct) {
fs/xfs/xfs_fsops.c
322
uint64_t icount = mp->m_sb.sb_dblocks * mp->m_sb.sb_imax_pct;
fs/xfs/xfs_fsops.c
324
M_IGEO(mp)->maxicount = XFS_FSB_TO_INO(mp, icount);
fs/xfs/xfs_fsops.c
326
M_IGEO(mp)->maxicount = 0;
fs/xfs/xfs_fsops.c
329
error = xfs_update_secondary_sbs(mp);
fs/xfs/xfs_fsops.c
335
mp->m_generation++;
fs/xfs/xfs_fsops.c
337
mutex_unlock(&mp->m_growlock);
fs/xfs/xfs_fsops.c
343
xfs_mount_t *mp,
fs/xfs/xfs_fsops.c
350
if (!mutex_trylock(&mp->m_growlock))
fs/xfs/xfs_fsops.c
352
error = xfs_growfs_log_private(mp, in);
fs/xfs/xfs_fsops.c
353
mutex_unlock(&mp->m_growlock);
fs/xfs/xfs_fsops.c
364
struct xfs_mount *mp,
fs/xfs/xfs_fsops.c
385
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
393
if (mp->m_free[ctr].res_total > request) {
fs/xfs/xfs_fsops.c
394
lcounter = mp->m_free[ctr].res_avail - request;
fs/xfs/xfs_fsops.c
397
mp->m_free[ctr].res_avail -= lcounter;
fs/xfs/xfs_fsops.c
399
mp->m_free[ctr].res_total = request;
fs/xfs/xfs_fsops.c
401
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
402
xfs_add_freecounter(mp, ctr, fdblks_delta);
fs/xfs/xfs_fsops.c
403
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
421
free = xfs_sum_freecounter_raw(mp, ctr) -
fs/xfs/xfs_fsops.c
422
xfs_freecounter_unavailable(mp, ctr);
fs/xfs/xfs_fsops.c
423
delta = request - mp->m_free[ctr].res_total;
fs/xfs/xfs_fsops.c
424
mp->m_free[ctr].res_total = request;
fs/xfs/xfs_fsops.c
437
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
438
error = xfs_dec_freecounter(mp, ctr, fdblks_delta, 0);
fs/xfs/xfs_fsops.c
440
xfs_add_freecounter(mp, ctr, fdblks_delta);
fs/xfs/xfs_fsops.c
441
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
444
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_fsops.c
450
xfs_mount_t *mp,
fs/xfs/xfs_fsops.c
455
if (!bdev_freeze(mp->m_super->s_bdev)) {
fs/xfs/xfs_fsops.c
456
xfs_force_shutdown(mp, SHUTDOWN_FORCE_UMOUNT);
fs/xfs/xfs_fsops.c
457
bdev_thaw(mp->m_super->s_bdev);
fs/xfs/xfs_fsops.c
462
xfs_force_shutdown(mp, SHUTDOWN_FORCE_UMOUNT);
fs/xfs/xfs_fsops.c
465
xfs_force_shutdown(mp,
fs/xfs/xfs_fsops.c
488
struct xfs_mount *mp,
fs/xfs/xfs_fsops.c
497
if (xfs_set_shutdown(mp)) {
fs/xfs/xfs_fsops.c
498
xlog_shutdown_wait(mp->m_log);
fs/xfs/xfs_fsops.c
501
if (mp->m_sb_bp)
fs/xfs/xfs_fsops.c
502
mp->m_sb_bp->b_flags |= XBF_DONE;
fs/xfs/xfs_fsops.c
505
xfs_alert(mp, "User initiated shutdown received.");
fs/xfs/xfs_fsops.c
507
if (xlog_force_shutdown(mp->m_log, flags)) {
fs/xfs/xfs_fsops.c
524
trace_xfs_force_shutdown(mp, tag, flags, fname, lnnum);
fs/xfs/xfs_fsops.c
526
xfs_alert_tag(mp, tag,
fs/xfs/xfs_fsops.c
529
xfs_alert(mp,
fs/xfs/xfs_fsops.c
534
fserror_report_shutdown(mp->m_super, GFP_KERNEL);
fs/xfs/xfs_fsops.c
535
xfs_healthmon_report_shutdown(mp, flags);
fs/xfs/xfs_fsops.c
54
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_fsops.c
543
struct xfs_mount *mp)
fs/xfs/xfs_fsops.c
549
mp->m_finobt_nores = false;
fs/xfs/xfs_fsops.c
55
xfs_rfsblock_t nb = mp->m_sb.sb_dblocks + delta;
fs/xfs/xfs_fsops.c
550
while ((pag = xfs_perag_next(mp, pag))) {
fs/xfs/xfs_fsops.c
557
xfs_warn(mp,
fs/xfs/xfs_fsops.c
559
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_fsops.c
563
err2 = xfs_metafile_resv_init(mp);
fs/xfs/xfs_fsops.c
565
xfs_warn(mp,
fs/xfs/xfs_fsops.c
567
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_fsops.c
581
struct xfs_mount *mp)
fs/xfs/xfs_fsops.c
585
xfs_metafile_resv_free(mp);
fs/xfs/xfs_fsops.c
586
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/xfs_fsops.c
67
(xfs_rfsblock_t)mp->m_sb.sb_agblocks);
fs/xfs/xfs_fsops.c
69
id->agsize = mp->m_sb.sb_agblocks;
fs/xfs/xfs_fsops.c
71
error = xfs_ag_init_headers(mp, id);
fs/xfs/xfs_fsops.c
94
struct xfs_mount *mp, /* mount point for filesystem */
fs/xfs/xfs_fsops.c
97
xfs_agnumber_t oagcount = mp->m_sb.sb_agcount;
fs/xfs/xfs_fsops.h
10
int xfs_growfs_log(struct xfs_mount *mp, struct xfs_growfs_log *in);
fs/xfs/xfs_fsops.h
11
int xfs_reserve_blocks(struct xfs_mount *mp, enum xfs_free_counter cnt,
fs/xfs/xfs_fsops.h
13
int xfs_fs_goingdown(struct xfs_mount *mp, uint32_t inflags);
fs/xfs/xfs_fsops.h
15
int xfs_fs_reserve_ag_blocks(struct xfs_mount *mp);
fs/xfs/xfs_fsops.h
16
void xfs_fs_unreserve_ag_blocks(struct xfs_mount *mp);
fs/xfs/xfs_fsops.h
9
int xfs_growfs_data(struct xfs_mount *mp, struct xfs_growfs_data *in);
fs/xfs/xfs_handle.c
186
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
195
inode = xfs_nfs_get_inode(mp->m_super, handle->ha_fid.fid_ino,
fs/xfs/xfs_handle.c
43
struct xfs_mount *mp,
fs/xfs/xfs_handle.c
48
memcpy(&handle->ha_fsid, mp->m_fixedfsid, sizeof(struct xfs_fsid));
fs/xfs/xfs_handle.c
60
struct xfs_mount *mp,
fs/xfs/xfs_handle.c
63
memcpy(&handle->ha_fsid, mp->m_fixedfsid, sizeof(struct xfs_fsid));
fs/xfs/xfs_handle.c
692
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
704
error = xfs_parent_from_attr(mp, flags, name, namelen, value, valuelen,
fs/xfs/xfs_handle.c
724
xfs_filehandle_init(mp, ino, gen, &gpr->gpr_parent);
fs/xfs/xfs_handle.c
758
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
811
if (ip->i_ino == mp->m_sb.sb_rootino)
fs/xfs/xfs_handle.c
855
struct xfs_mount *mp = gpx.ip->i_mount;
fs/xfs/xfs_handle.c
860
if (!xfs_has_parent(mp))
fs/xfs/xfs_handle.c
883
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_handle.c
890
if (!xfs_has_parent(mp))
fs/xfs/xfs_health.c
101
xfs_fs_mark_healthy(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/xfs_health.c
108
struct xfs_mount *mp,
fs/xfs/xfs_health.c
114
trace_xfs_fs_mark_sick(mp, mask);
fs/xfs/xfs_health.c
116
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
117
old_mask = mp->m_fs_sick;
fs/xfs/xfs_health.c
118
mp->m_fs_sick |= mask;
fs/xfs/xfs_health.c
119
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
121
fserror_report_metadata(mp->m_super, -EFSCORRUPTED, GFP_NOFS);
fs/xfs/xfs_health.c
123
xfs_healthmon_report_fs(mp, XFS_HEALTHMON_SICK, old_mask, mask);
fs/xfs/xfs_health.c
129
struct xfs_mount *mp,
fs/xfs/xfs_health.c
135
trace_xfs_fs_mark_corrupt(mp, mask);
fs/xfs/xfs_health.c
137
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
138
old_mask = mp->m_fs_sick;
fs/xfs/xfs_health.c
139
mp->m_fs_sick |= mask;
fs/xfs/xfs_health.c
140
mp->m_fs_checked |= mask;
fs/xfs/xfs_health.c
141
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
143
fserror_report_metadata(mp->m_super, -EFSCORRUPTED, GFP_NOFS);
fs/xfs/xfs_health.c
145
xfs_healthmon_report_fs(mp, XFS_HEALTHMON_CORRUPT, old_mask,
fs/xfs/xfs_health.c
152
struct xfs_mount *mp,
fs/xfs/xfs_health.c
158
trace_xfs_fs_mark_healthy(mp, mask);
fs/xfs/xfs_health.c
160
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
161
old_mask = mp->m_fs_sick;
fs/xfs/xfs_health.c
162
mp->m_fs_sick &= ~mask;
fs/xfs/xfs_health.c
163
if (!(mp->m_fs_sick & XFS_SICK_FS_PRIMARY))
fs/xfs/xfs_health.c
164
mp->m_fs_sick &= ~XFS_SICK_FS_SECONDARY;
fs/xfs/xfs_health.c
165
mp->m_fs_checked |= mask;
fs/xfs/xfs_health.c
166
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
169
xfs_healthmon_report_fs(mp, XFS_HEALTHMON_HEALTHY, old_mask,
fs/xfs/xfs_health.c
176
struct xfs_mount *mp,
fs/xfs/xfs_health.c
180
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
181
*sick = mp->m_fs_sick;
fs/xfs/xfs_health.c
182
*checked = mp->m_fs_checked;
fs/xfs/xfs_health.c
183
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_health.c
189
struct xfs_mount *mp,
fs/xfs/xfs_health.c
193
struct xfs_perag *pag = xfs_perag_get(mp, agno);
fs/xfs/xfs_health.c
303
struct xfs_mount *mp,
fs/xfs/xfs_health.c
307
struct xfs_rtgroup *rtg = xfs_rtgroup_get(mp, rgno);
fs/xfs/xfs_health.c
475
struct xfs_mount *mp,
fs/xfs/xfs_health.c
48
struct xfs_mount *mp)
fs/xfs/xfs_health.c
486
xfs_fs_measure_sickness(mp, &sick, &checked);
fs/xfs/xfs_health.c
490
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/xfs_health.c
56
if (xfs_is_shutdown(mp))
fs/xfs/xfs_health.c
60
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/xfs_health.c
64
while ((rtg = xfs_rtgroup_next(mp, rtg)))
fs/xfs/xfs_health.c
71
xfs_fs_measure_sickness(mp, &sick, &checked);
fs/xfs/xfs_health.c
73
trace_xfs_fs_unfixed_corruption(mp, sick);
fs/xfs/xfs_health.c
78
xfs_warn(mp,
fs/xfs/xfs_healthmon.c
109
struct xfs_mount *mp,
fs/xfs/xfs_healthmon.c
113
if (rcu_access_pointer(mp->m_healthmon) != NULL) {
fs/xfs/xfs_healthmon.c
1174
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_healthmon.c
1185
if (ip->i_ino != mp->m_sb.sb_rootino)
fs/xfs/xfs_healthmon.c
119
rcu_assign_pointer(mp->m_healthmon, hm);
fs/xfs/xfs_healthmon.c
1199
hm->dev = mp->m_super->s_dev;
fs/xfs/xfs_healthmon.c
120
hm->mount_cookie = (uintptr_t)mp->m_super;
fs/xfs/xfs_healthmon.c
1235
ret = xfs_healthmon_attach(mp, hm);
fs/xfs/xfs_healthmon.c
1250
trace_xfs_healthmon_create(mp->m_super->s_dev, hmo.flags, hmo.format);
fs/xfs/xfs_healthmon.c
131
struct xfs_mount *mp;
fs/xfs/xfs_healthmon.c
139
mp = XFS_M((struct super_block *)hm->mount_cookie);
fs/xfs/xfs_healthmon.c
140
rcu_assign_pointer(mp->m_healthmon, NULL);
fs/xfs/xfs_healthmon.c
403
struct xfs_mount *mp)
fs/xfs/xfs_healthmon.c
405
struct xfs_healthmon *hm = xfs_healthmon_get(mp);
fs/xfs/xfs_healthmon.c
458
struct xfs_mount *mp,
fs/xfs/xfs_healthmon.c
467
struct xfs_healthmon *hm = xfs_healthmon_get(mp);
fs/xfs/xfs_healthmon.c
557
struct xfs_mount *mp,
fs/xfs/xfs_healthmon.c
565
struct xfs_healthmon *hm = xfs_healthmon_get(mp);
fs/xfs/xfs_healthmon.c
596
struct xfs_mount *mp,
fs/xfs/xfs_healthmon.c
607
struct xfs_healthmon *hm = xfs_healthmon_get(mp);
fs/xfs/xfs_healthmon.c
67
struct xfs_mount *mp)
fs/xfs/xfs_healthmon.c
72
hm = rcu_dereference(mp->m_healthmon);
fs/xfs/xfs_healthmon.h
163
void xfs_healthmon_report_fs(struct xfs_mount *mp,
fs/xfs/xfs_healthmon.h
173
void xfs_healthmon_report_shutdown(struct xfs_mount *mp, uint32_t flags);
fs/xfs/xfs_healthmon.h
175
void xfs_healthmon_report_media(struct xfs_mount *mp, enum xfs_device fdev,
fs/xfs/xfs_healthmon.h
68
void xfs_healthmon_unmount(struct xfs_mount *mp);
fs/xfs/xfs_icache.c
1063
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1065
return xfs_is_unmounting(mp) || xfs_has_norecovery(mp) ||
fs/xfs/xfs_icache.c
1066
xfs_is_shutdown(mp);
fs/xfs/xfs_icache.c
107
M_IGEO(mp)->min_folio_order);
fs/xfs/xfs_icache.c
1071
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1077
if (xfs_want_reclaim_sick(mp))
fs/xfs/xfs_icache.c
1080
while (xfs_group_marked(mp, XG_TYPE_AG, XFS_PERAG_RECLAIM_MARK)) {
fs/xfs/xfs_icache.c
1081
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/xfs_icache.c
1082
xfs_icwalk(mp, XFS_ICWALK_RECLAIM, &icw);
fs/xfs/xfs_icache.c
109
XFS_STATS_INC(mp, xs_inodes_active);
fs/xfs/xfs_icache.c
1095
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
1103
if (xfs_want_reclaim_sick(mp))
fs/xfs/xfs_icache.c
1107
xfs_reclaim_work_queue(mp);
fs/xfs/xfs_icache.c
1108
xfs_ail_push_all(mp->m_ail);
fs/xfs/xfs_icache.c
1110
xfs_icwalk(mp, XFS_ICWALK_RECLAIM, &icw);
fs/xfs/xfs_icache.c
1120
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1122
XA_STATE (xas, &mp->m_groups[XG_TYPE_AG].xa, 0);
fs/xfs/xfs_icache.c
115
ip->i_mount = mp;
fs/xfs/xfs_icache.c
1220
struct xfs_mount *mp = container_of(to_delayed_work(work),
fs/xfs/xfs_icache.c
1223
xfs_icwalk(mp, XFS_ICWALK_RECLAIM, NULL);
fs/xfs/xfs_icache.c
1224
xfs_reclaim_work_queue(mp);
fs/xfs/xfs_icache.c
123
ip->i_diflags2 = mp->m_ino_geo.new_diflags2;
fs/xfs/xfs_icache.c
1275
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
1290
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1293
xfs_perag_set_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
1313
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
1327
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1330
xfs_perag_clear_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
1465
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1469
if (!xfs_clear_blockgc_enabled(mp))
fs/xfs/xfs_icache.c
1472
while ((pag = xfs_perag_next(mp, pag)))
fs/xfs/xfs_icache.c
1474
trace_xfs_blockgc_stop(mp, __return_address);
fs/xfs/xfs_icache.c
1480
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1484
if (xfs_set_blockgc_enabled(mp))
fs/xfs/xfs_icache.c
1487
trace_xfs_blockgc_start(mp, __return_address);
fs/xfs/xfs_icache.c
1488
while ((pag = xfs_perag_grab_next_tag(mp, pag, XFS_ICI_BLOCKGC_TAG)))
fs/xfs/xfs_icache.c
1564
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_icache.c
1567
trace_xfs_blockgc_worker(mp, __return_address);
fs/xfs/xfs_icache.c
1571
xfs_info(mp, "AG %u preallocation gc worker failed, err=%d",
fs/xfs/xfs_icache.c
1582
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
1587
trace_xfs_blockgc_free_space(mp, icw, _RET_IP_);
fs/xfs/xfs_icache.c
1589
error = xfs_icwalk(mp, XFS_ICWALK_BLOCKGC, icw);
fs/xfs/xfs_icache.c
1593
return xfs_inodegc_flush(mp);
fs/xfs/xfs_icache.c
1602
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
1606
trace_xfs_blockgc_flush_all(mp, __return_address);
fs/xfs/xfs_icache.c
1612
while ((pag = xfs_perag_grab_next_tag(mp, pag, XFS_ICI_BLOCKGC_TAG)))
fs/xfs/xfs_icache.c
1613
mod_delayed_work(mp->m_blockgc_wq, &pag->pag_blockgc_work, 0);
fs/xfs/xfs_icache.c
1615
while ((pag = xfs_perag_grab_next_tag(mp, pag, XFS_ICI_BLOCKGC_TAG)))
fs/xfs/xfs_icache.c
1618
return xfs_inodegc_flush(mp);
fs/xfs/xfs_icache.c
1633
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
1651
if (XFS_IS_UQUOTA_ENFORCED(mp) && udqp && xfs_dquot_lowsp(udqp)) {
fs/xfs/xfs_icache.c
1652
icw.icw_uid = make_kuid(mp->m_super->s_user_ns, udqp->q_id);
fs/xfs/xfs_icache.c
1657
if (XFS_IS_UQUOTA_ENFORCED(mp) && gdqp && xfs_dquot_lowsp(gdqp)) {
fs/xfs/xfs_icache.c
1658
icw.icw_gid = make_kgid(mp->m_super->s_user_ns, gdqp->q_id);
fs/xfs/xfs_icache.c
1663
if (XFS_IS_PQUOTA_ENFORCED(mp) && pdqp && xfs_dquot_lowsp(pdqp)) {
fs/xfs/xfs_icache.c
1672
return xfs_blockgc_free_space(mp, &icw);
fs/xfs/xfs_icache.c
1752
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_icache.c
1805
if (XFS_INO_TO_AGNO(mp, ip->i_ino) != pag_agno(pag))
fs/xfs/xfs_icache.c
1807
first_index = XFS_INO_TO_AGINO(mp, ip->i_ino + 1);
fs/xfs/xfs_icache.c
1808
if (first_index < XFS_INO_TO_AGINO(mp, ip->i_ino))
fs/xfs/xfs_icache.c
1857
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
1865
while ((pag = xfs_perag_grab_next_tag(mp, pag, goal))) {
fs/xfs/xfs_icache.c
1910
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
1913
if (!xfs_is_shutdown(mp) && ip->i_delayed_blks) {
fs/xfs/xfs_icache.c
1919
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_icache.c
1926
xfs_perag_set_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
1960
struct xfs_mount *mp = gc->mp;
fs/xfs/xfs_icache.c
1969
cpumask_clear_cpu(gc->cpu, &mp->m_inodegc_cpumask);
fs/xfs/xfs_icache.c
1985
trace_xfs_inodegc_worker(mp, READ_ONCE(gc->shrinker_hits));
fs/xfs/xfs_icache.c
2006
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2008
if (!xfs_is_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
2010
trace_xfs_inodegc_push(mp, __return_address);
fs/xfs/xfs_icache.c
2011
xfs_inodegc_queue_all(mp);
fs/xfs/xfs_icache.c
2020
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2022
xfs_inodegc_push(mp);
fs/xfs/xfs_icache.c
2023
trace_xfs_inodegc_flush(mp, __return_address);
fs/xfs/xfs_icache.c
2024
return xfs_inodegc_wait_all(mp);
fs/xfs/xfs_icache.c
2034
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2038
if (!xfs_clear_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
2051
xfs_inodegc_queue_all(mp);
fs/xfs/xfs_icache.c
2053
flush_workqueue(mp->m_inodegc_wq);
fs/xfs/xfs_icache.c
2054
rerun = xfs_inodegc_queue_all(mp);
fs/xfs/xfs_icache.c
2057
trace_xfs_inodegc_stop(mp, __return_address);
fs/xfs/xfs_icache.c
2067
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2069
if (xfs_set_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
2072
trace_xfs_inodegc_start(mp, __return_address);
fs/xfs/xfs_icache.c
2073
xfs_inodegc_queue_all(mp);
fs/xfs/xfs_icache.c
208
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2081
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2083
if (!XFS_IS_REALTIME_INODE(ip) || xfs_has_zoned(mp))
fs/xfs/xfs_icache.c
2086
if (xfs_compare_freecounter(mp, XC_FREE_RTEXTENTS,
fs/xfs/xfs_icache.c
2087
mp->m_low_rtexts[XFS_LOWSP_5_PCNT],
fs/xfs/xfs_icache.c
2109
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2111
if (items > mp->m_ino_geo.inodes_per_cluster)
fs/xfs/xfs_icache.c
2114
if (xfs_compare_freecounter(mp, XC_FREE_BLOCKS,
fs/xfs/xfs_icache.c
2115
mp->m_low_space[XFS_LOWSP_5_PCNT],
fs/xfs/xfs_icache.c
212
if (xfs_group_marked(mp, XG_TYPE_AG, XFS_PERAG_RECLAIM_MARK)) {
fs/xfs/xfs_icache.c
213
queue_delayed_work(mp->m_reclaim_workqueue, &mp->m_reclaim_work,
fs/xfs/xfs_icache.c
2178
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2191
gc = this_cpu_ptr(mp->m_inodegc);
fs/xfs/xfs_icache.c
2203
if (!cpumask_test_cpu(cpu_nr, &mp->m_inodegc_cpumask))
fs/xfs/xfs_icache.c
2204
cpumask_test_and_set_cpu(cpu_nr, &mp->m_inodegc_cpumask);
fs/xfs/xfs_icache.c
2210
if (!xfs_is_inodegc_enabled(mp)) {
fs/xfs/xfs_icache.c
2218
trace_xfs_inodegc_queue(mp, __return_address);
fs/xfs/xfs_icache.c
2219
mod_delayed_work_on(current_cpu(), mp->m_inodegc_wq, &gc->work,
fs/xfs/xfs_icache.c
2224
trace_xfs_inodegc_throttle(mp, __return_address);
fs/xfs/xfs_icache.c
2243
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
2246
XFS_STATS_INC(mp, xs_inode_mark_reclaimable);
fs/xfs/xfs_icache.c
227
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_icache.c
2282
struct xfs_mount *mp = shrink->private_data;
fs/xfs/xfs_icache.c
2286
if (!xfs_is_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
2289
for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
fs/xfs/xfs_icache.c
229
if (!xfs_is_blockgc_enabled(mp))
fs/xfs/xfs_icache.c
2290
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
2303
struct xfs_mount *mp = shrink->private_data;
fs/xfs/xfs_icache.c
2308
if (!xfs_is_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
2311
trace_xfs_inodegc_shrinker_scan(mp, sc, __return_address);
fs/xfs/xfs_icache.c
2313
for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
fs/xfs/xfs_icache.c
2314
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
2319
mod_delayed_work_on(cpu, mp->m_inodegc_wq, &gc->work, 0);
fs/xfs/xfs_icache.c
2337
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
2339
mp->m_inodegc_shrinker = shrinker_alloc(SHRINKER_NONSLAB,
fs/xfs/xfs_icache.c
234
queue_delayed_work(mp->m_blockgc_wq, &pag->pag_blockgc_work,
fs/xfs/xfs_icache.c
2341
mp->m_super->s_id);
fs/xfs/xfs_icache.c
2342
if (!mp->m_inodegc_shrinker)
fs/xfs/xfs_icache.c
2345
mp->m_inodegc_shrinker->count_objects = xfs_inodegc_shrinker_count;
fs/xfs/xfs_icache.c
2346
mp->m_inodegc_shrinker->scan_objects = xfs_inodegc_shrinker_scan;
fs/xfs/xfs_icache.c
2347
mp->m_inodegc_shrinker->seeks = 0;
fs/xfs/xfs_icache.c
2348
mp->m_inodegc_shrinker->batch = XFS_INODEGC_SHRINKER_BATCH;
fs/xfs/xfs_icache.c
2349
mp->m_inodegc_shrinker->private_data = mp;
fs/xfs/xfs_icache.c
2351
shrinker_register(mp->m_inodegc_shrinker);
fs/xfs/xfs_icache.c
309
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
313
return to_perag(xfs_group_grab_next_mark(mp,
fs/xfs/xfs_icache.c
328
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
341
error = inode_init_always(mp->m_super, inode);
fs/xfs/xfs_icache.c
352
M_IGEO(mp)->min_folio_order);
fs/xfs/xfs_icache.c
365
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
372
error = xfs_reinit_inode(mp, inode);
fs/xfs/xfs_icache.c
400
xfs_perag_clear_inode_tag(pag, XFS_INO_TO_AGINO(mp, ip->i_ino),
fs/xfs/xfs_icache.c
458
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
464
for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
fs/xfs/xfs_icache.c
465
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
467
mod_delayed_work_on(cpu, mp->m_inodegc_wq, &gc->work, 0);
fs/xfs/xfs_icache.c
478
struct xfs_mount *mp)
fs/xfs/xfs_icache.c
483
flush_workqueue(mp->m_inodegc_wq);
fs/xfs/xfs_icache.c
484
for_each_cpu(cpu, &mp->m_inodegc_cpumask) {
fs/xfs/xfs_icache.c
487
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_icache.c
508
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_icache.c
51
static int xfs_icwalk(struct xfs_mount *mp,
fs/xfs/xfs_icache.c
598
XFS_STATS_INC(mp, xs_ig_found);
fs/xfs/xfs_icache.c
604
XFS_STATS_INC(mp, xs_ig_frecycle);
fs/xfs/xfs_icache.c
618
if (xfs_is_inodegc_enabled(mp))
fs/xfs/xfs_icache.c
619
xfs_inodegc_queue_all(mp);
fs/xfs/xfs_icache.c
625
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
635
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ino);
fs/xfs/xfs_icache.c
637
ip = xfs_inode_alloc(mp, ino);
fs/xfs/xfs_icache.c
662
if (xfs_has_v3inodes(mp) && (flags & XFS_IGET_CREATE)) {
fs/xfs/xfs_icache.c
667
error = xfs_imap_to_bp(mp, tp, &ip->i_imap, &bp);
fs/xfs/xfs_icache.c
732
XFS_STATS_INC(mp, xs_ig_dup);
fs/xfs/xfs_icache.c
767
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
782
if (!xfs_verify_ino(mp, ino))
fs/xfs/xfs_icache.c
785
XFS_STATS_INC(mp, xs_ig_attempts);
fs/xfs/xfs_icache.c
788
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ino));
fs/xfs/xfs_icache.c
789
agino = XFS_INO_TO_AGINO(mp, ino);
fs/xfs/xfs_icache.c
806
XFS_STATS_INC(mp, xs_ig_missed);
fs/xfs/xfs_icache.c
808
error = xfs_iget_cache_miss(mp, pag, tp, ino, &ip,
fs/xfs/xfs_icache.c
849
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_icache.c
854
error = xfs_iget(mp, tp, ino, 0, 0, &ip);
fs/xfs/xfs_icache.c
869
if (xfs_has_metadir(mp)) {
fs/xfs/xfs_icache.c
88
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
881
xfs_err(mp, "metadata inode 0x%llx type %u is corrupt", ino,
fs/xfs/xfs_icache.c
883
xfs_fs_mark_sick(mp, XFS_SICK_FS_METADIR);
fs/xfs/xfs_icache.c
890
struct xfs_mount *mp,
fs/xfs/xfs_icache.c
898
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_icache.c
97
ip = alloc_inode_sb(mp->m_super, xfs_inode_cache, GFP_KERNEL | __GFP_NOFAIL);
fs/xfs/xfs_icache.c
99
if (inode_init_always(mp->m_super, VFS_I(ip))) {
fs/xfs/xfs_icache.h
45
int xfs_iget(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t ino,
fs/xfs/xfs_icache.h
49
struct xfs_inode * xfs_inode_alloc(struct xfs_mount *mp, xfs_ino_t ino);
fs/xfs/xfs_icache.h
54
void xfs_reclaim_inodes(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
55
long xfs_reclaim_inodes_count(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
56
long xfs_reclaim_inodes_nr(struct xfs_mount *mp, unsigned long nr_to_scan);
fs/xfs/xfs_icache.h
60
int xfs_blockgc_free_dquots(struct xfs_mount *mp, struct xfs_dquot *udqp,
fs/xfs/xfs_icache.h
64
int xfs_blockgc_free_space(struct xfs_mount *mp, struct xfs_icwalk *icm);
fs/xfs/xfs_icache.h
65
int xfs_blockgc_flush_all(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
74
void xfs_blockgc_stop(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
75
void xfs_blockgc_start(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
78
void xfs_inodegc_push(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
79
int xfs_inodegc_flush(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
80
void xfs_inodegc_stop(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
81
void xfs_inodegc_start(struct xfs_mount *mp);
fs/xfs/xfs_icache.h
82
int xfs_inodegc_register_shrinker(struct xfs_mount *mp);
fs/xfs/xfs_icreate_item.c
146
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_icreate_item.c
148
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/xfs_icreate_item.c
171
if (agno >= mp->m_sb.sb_agcount) {
fs/xfs/xfs_icreate_item.c
176
if (!agbno || agbno == NULLAGBLOCK || agbno >= mp->m_sb.sb_agblocks) {
fs/xfs/xfs_icreate_item.c
181
if (isize != mp->m_sb.sb_inodesize) {
fs/xfs/xfs_icreate_item.c
191
if (!length || length >= mp->m_sb.sb_agblocks) {
fs/xfs/xfs_icreate_item.c
208
if ((count >> mp->m_sb.sb_inopblog) != length) {
fs/xfs/xfs_icreate_item.c
221
bb_per_cluster = XFS_FSB_TO_BB(mp, igeo->blocks_per_cluster);
fs/xfs/xfs_icreate_item.c
226
daddr = XFS_AGB_TO_DADDR(mp, agno,
fs/xfs/xfs_icreate_item.c
245
xfs_warn(mp,
fs/xfs/xfs_icreate_item.c
252
return xfs_ialloc_inode_init(mp, NULL, buffer_list, count, agno, agbno,
fs/xfs/xfs_inode.c
1043
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1056
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
fs/xfs/xfs_inode.c
1072
first_unmap_block = XFS_B_TO_FSB(mp, (xfs_ufsize_t)new_size);
fs/xfs/xfs_inode.c
1073
if (!xfs_verify_fileoff(mp, first_unmap_block)) {
fs/xfs/xfs_inode.c
1118
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_inode.c
1119
struct xfs_da_geometry *geo = mp->m_dir_geo;
fs/xfs/xfs_inode.c
1138
error = xfs_buf_incore(mp->m_ddev_targp,
fs/xfs/xfs_inode.c
1139
XFS_FSB_TO_DADDR(mp, fsbno),
fs/xfs/xfs_inode.c
1140
XFS_FSB_TO_BB(mp, geo->fsbcount),
fs/xfs/xfs_inode.c
1160
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1164
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, 0, 0, 0, &tp);
fs/xfs/xfs_inode.c
1166
ASSERT(xfs_is_shutdown(mp));
fs/xfs/xfs_inode.c
1209
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1224
if (unlikely(mp->m_finobt_nores)) {
fs/xfs/xfs_inode.c
1225
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ifree,
fs/xfs/xfs_inode.c
1226
XFS_IFREE_SPACE_RES(mp), 0, XFS_TRANS_RESERVE,
fs/xfs/xfs_inode.c
1229
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ifree, 0, 0, 0, &tp);
fs/xfs/xfs_inode.c
1233
xfs_warn_ratelimited(mp,
fs/xfs/xfs_inode.c
1237
ASSERT(xfs_is_shutdown(mp));
fs/xfs/xfs_inode.c
1273
if (!xfs_is_shutdown(mp)) {
fs/xfs/xfs_inode.c
1274
xfs_notice(mp, "%s: xfs_ifree returned error %d",
fs/xfs/xfs_inode.c
1276
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_inode.c
1300
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1314
if (xfs_is_readonly(mp) && !xlog_recovery_needed(mp->m_log))
fs/xfs/xfs_inode.c
1318
if (xfs_is_shutdown(mp) || xfs_has_norecovery(mp))
fs/xfs/xfs_inode.c
1352
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1366
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_inode.c
1389
struct xfs_mount *mp;
fs/xfs/xfs_inode.c
1402
mp = ip->i_mount;
fs/xfs/xfs_inode.c
1411
if (xfs_is_readonly(mp) && !xlog_recovery_needed(mp->m_log))
fs/xfs/xfs_inode.c
1544
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_inode.c
1557
xfs_info_ratelimited(mp,
fs/xfs/xfs_inode.c
1567
error = xfs_iget(mp, tp, xfs_agino_to_ino(pag, next_agino),
fs/xfs/xfs_inode.c
1585
if (xfs_is_quotacheck_running(mp) && next_ip)
fs/xfs/xfs_inode.c
1602
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_inode.c
1608
ip = radix_tree_lookup(&pag->pag_ici_root, XFS_INO_TO_AGINO(mp, inum));
fs/xfs/xfs_inode.c
1650
ASSERT(iip->ili_last_fields || xlog_is_shutdown(mp->m_log));
fs/xfs/xfs_inode.c
1698
struct xfs_mount *mp = free_ip->i_mount;
fs/xfs/xfs_inode.c
1699
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/xfs_inode.c
1722
blkno = XFS_AGB_TO_DADDR(mp, XFS_INO_TO_AGNO(mp, inum),
fs/xfs/xfs_inode.c
1723
XFS_INO_TO_AGBNO(mp, inum));
fs/xfs/xfs_inode.c
1734
error = xfs_trans_get_buf(tp, mp->m_ddev_targp, blkno,
fs/xfs/xfs_inode.c
1735
mp->m_bsize * igeo->blocks_per_cluster, 0, &bp);
fs/xfs/xfs_inode.c
1793
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
1805
pag = xfs_perag_get(mp, XFS_INO_TO_AGNO(mp, ip->i_ino));
fs/xfs/xfs_inode.c
1915
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_inode.c
1924
if (xfs_is_shutdown(mp))
fs/xfs/xfs_inode.c
1937
error = xfs_parent_start(mp, &du.ppargs);
fs/xfs/xfs_inode.c
1952
resblks = xfs_remove_space_res(mp, name->len);
fs/xfs/xfs_inode.c
1953
error = xfs_trans_alloc_dir(dp, &M_RES(mp)->tr_remove, ip, &resblks,
fs/xfs/xfs_inode.c
1969
if (xfs_has_wsync(mp) || xfs_has_dirsync(mp))
fs/xfs/xfs_inode.c
1981
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.c
1990
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.c
2146
struct xfs_mount *mp = src_dp->i_mount;
fs/xfs/xfs_inode.c
2180
error = xfs_parent_start(mp, &du_src.ppargs);
fs/xfs/xfs_inode.c
2185
error = xfs_parent_start(mp, &du_wip.ppargs);
fs/xfs/xfs_inode.c
2191
error = xfs_parent_start(mp, &du_tgt.ppargs);
fs/xfs/xfs_inode.c
2198
spaceres = xfs_rename_space_res(mp, src_name->len, target_ip != NULL,
fs/xfs/xfs_inode.c
2200
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_rename, spaceres, 0, 0, &tp);
fs/xfs/xfs_inode.c
2204
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_rename, 0, 0, 0,
fs/xfs/xfs_inode.c
2314
pag = xfs_perag_get(mp,
fs/xfs/xfs_inode.c
2315
XFS_INO_TO_AGNO(mp, inodes[i]->i_ino));
fs/xfs/xfs_inode.c
2354
xfs_parent_finish(mp, du_tgt.ppargs);
fs/xfs/xfs_inode.c
2356
xfs_parent_finish(mp, du_wip.ppargs);
fs/xfs/xfs_inode.c
2358
xfs_parent_finish(mp, du_src.ppargs);
fs/xfs/xfs_inode.c
2374
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.c
2393
XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_1)) {
fs/xfs/xfs_inode.c
2394
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2402
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2411
XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_3)) {
fs/xfs/xfs_inode.c
2412
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2421
XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_4)) {
fs/xfs/xfs_inode.c
2422
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2429
ip->i_nblocks || XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_5)) {
fs/xfs/xfs_inode.c
2430
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2438
if (ip->i_forkoff > mp->m_sb.sb_inodesize ||
fs/xfs/xfs_inode.c
2439
XFS_TEST_ERROR(mp, XFS_ERRTAG_IFLUSH_6)) {
fs/xfs/xfs_inode.c
2440
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2448
xfs_alert_tag(mp, XFS_PTAG_IFLUSH,
fs/xfs/xfs_inode.c
2462
if (!xfs_has_v3inodes(mp))
fs/xfs/xfs_inode.c
2485
if (!xfs_has_v3inodes(mp)) {
fs/xfs/xfs_inode.c
2522
xfs_trans_ail_copy_lsn(mp->m_ail, &iip->ili_flush_lsn,
fs/xfs/xfs_inode.c
2526
xfs_dinode_calc_crc(mp, dip);
fs/xfs/xfs_inode.c
2549
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_inode.c
2606
if (xlog_is_shutdown(mp->m_log)) {
fs/xfs/xfs_inode.c
2642
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_inode.c
2651
XFS_STATS_INC(mp, xs_icluster_flushcnt);
fs/xfs/xfs_inode.c
2652
XFS_STATS_ADD(mp, xs_icluster_flushinode, clcount);
fs/xfs/xfs_inode.c
2858
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_inode.c
2862
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, ip->i_ino);
fs/xfs/xfs_inode.c
2863
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, ip->i_ino);
fs/xfs/xfs_inode.c
2870
pag = xfs_perag_get(mp, agno);
fs/xfs/xfs_inode.c
2891
xfs_info_ratelimited(mp,
fs/xfs/xfs_inode.c
592
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_inode.c
600
error = xfs_iget(mp, tp, ino, XFS_IGET_CREATE, XFS_ILOCK_EXCL, &ip);
fs/xfs/xfs_inode.c
658
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_inode.c
672
if (xfs_is_shutdown(mp))
fs/xfs/xfs_inode.c
683
resblks = xfs_mkdir_space_res(mp, name->len);
fs/xfs/xfs_inode.c
684
tres = &M_RES(mp)->tr_mkdir;
fs/xfs/xfs_inode.c
686
resblks = xfs_create_space_res(mp, name->len);
fs/xfs/xfs_inode.c
687
tres = &M_RES(mp)->tr_create;
fs/xfs/xfs_inode.c
690
error = xfs_parent_start(mp, &du.ppargs);
fs/xfs/xfs_inode.c
700
error = xfs_trans_alloc_icreate(mp, tres, udqp, gdqp, pdqp, resblks,
fs/xfs/xfs_inode.c
704
xfs_flush_inodes(mp);
fs/xfs/xfs_inode.c
705
error = xfs_trans_alloc_icreate(mp, tres, udqp, gdqp, pdqp,
fs/xfs/xfs_inode.c
743
if (xfs_has_wsync(mp) || xfs_has_dirsync(mp))
fs/xfs/xfs_inode.c
764
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.c
781
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.c
798
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_inode.c
811
if (xfs_is_shutdown(mp))
fs/xfs/xfs_inode.c
819
resblks = XFS_IALLOC_SPACE_RES(mp);
fs/xfs/xfs_inode.c
820
tres = &M_RES(mp)->tr_create_tmpfile;
fs/xfs/xfs_inode.c
822
error = xfs_trans_alloc_icreate(mp, tres, udqp, gdqp, pdqp, resblks,
fs/xfs/xfs_inode.c
833
if (xfs_has_wsync(mp))
fs/xfs/xfs_inode.c
920
struct xfs_mount *mp = tdp->i_mount;
fs/xfs/xfs_inode.c
929
if (xfs_is_shutdown(mp))
fs/xfs/xfs_inode.c
942
error = xfs_parent_start(mp, &du.ppargs);
fs/xfs/xfs_inode.c
946
resblks = xfs_link_space_res(mp, target_name->len);
fs/xfs/xfs_inode.c
947
error = xfs_trans_alloc_dir(tdp, &M_RES(mp)->tr_link, sip, &resblks,
fs/xfs/xfs_inode.c
975
if (xfs_has_wsync(mp) || xfs_has_dirsync(mp))
fs/xfs/xfs_inode.c
981
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.c
989
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_inode.h
292
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode.h
295
if (xfs_has_metadir(mp))
fs/xfs/xfs_inode.h
302
return ip->i_ino == mp->m_sb.sb_rbmino ||
fs/xfs/xfs_inode.h
303
ip->i_ino == mp->m_sb.sb_rsumino ||
fs/xfs/xfs_inode.h
304
xfs_is_quota_inode(&mp->m_sb, ip->i_ino);
fs/xfs/xfs_inode_item.c
45
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_inode_item.c
49
dip = kzalloc(mp->m_sb.sb_inodesize, GFP_KERNEL | GFP_NOFS);
fs/xfs/xfs_inode_item.c
56
xfs_dinode_calc_crc(mp, dip);
fs/xfs/xfs_inode_item.c
57
fa = xfs_dinode_verify(mp, ip->i_ino, dip);
fs/xfs/xfs_inode_item.c
61
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_inode_item.c
926
struct xfs_mount *mp)
fs/xfs/xfs_inode_item.c
936
xfs_log_item_init(mp, &iip->ili_item, XFS_LI_INODE,
fs/xfs/xfs_inode_item_recover.c
223
struct xfs_mount *mp,
fs/xfs/xfs_inode_item_recover.c
230
if (!xfs_has_large_extent_counts(mp) ||
fs/xfs/xfs_inode_item_recover.c
234
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
235
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
237
ldip->di_ino, xfs_has_large_extent_counts(mp),
fs/xfs/xfs_inode_item_recover.c
248
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
249
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
261
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
262
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
264
ldip->di_ino, xfs_has_large_extent_counts(mp), nextents,
fs/xfs/xfs_inode_item_recover.c
274
struct xfs_mount *mp,
fs/xfs/xfs_inode_item_recover.c
280
unsigned int dsize = XFS_DFORK_DSIZE(dip, mp);
fs/xfs/xfs_inode_item_recover.c
284
xfs_bmbt_to_bmdr(mp, src, len, dfork, dsize);
fs/xfs/xfs_inode_item_recover.c
289
xfs_rtrmapbt_to_disk(mp, src, len, dfork, dsize);
fs/xfs/xfs_inode_item_recover.c
292
xfs_rtrefcountbt_to_disk(mp, src, len, dfork, dsize);
fs/xfs/xfs_inode_item_recover.c
315
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_inode_item_recover.c
351
error = xfs_buf_read(mp->m_ddev_targp, in_f->ilf_blkno, in_f->ilf_len,
fs/xfs/xfs_inode_item_recover.c
362
if (XFS_IS_CORRUPT(mp, !xfs_verify_magic16(bp, dip->di_magic))) {
fs/xfs/xfs_inode_item_recover.c
363
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
370
if (XFS_IS_CORRUPT(mp, ldip->di_magic != XFS_DINODE_MAGIC)) {
fs/xfs/xfs_inode_item_recover.c
371
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
412
if (!xfs_has_v3inodes(mp)) {
fs/xfs/xfs_inode_item_recover.c
439
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
440
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
452
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
453
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
461
error = xlog_dinode_verify_extent_counts(mp, ldip);
fs/xfs/xfs_inode_item_recover.c
465
if (unlikely(ldip->di_forkoff > mp->m_sb.sb_inodesize)) {
fs/xfs/xfs_inode_item_recover.c
467
XFS_ERRLEVEL_LOW, mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
468
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
474
isize = xfs_log_dinode_size(mp);
fs/xfs/xfs_inode_item_recover.c
477
mp, ldip, sizeof(*ldip));
fs/xfs/xfs_inode_item_recover.c
478
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
517
error = xlog_recover_inode_dbroot(mp, src, len, dip);
fs/xfs/xfs_inode_item_recover.c
549
ASSERT(len <= XFS_DFORK_ASIZE(dip, mp));
fs/xfs/xfs_inode_item_recover.c
555
xfs_bmbt_to_bmdr(mp, (struct xfs_btree_block *)src,
fs/xfs/xfs_inode_item_recover.c
557
XFS_DFORK_ASIZE(dip, mp));
fs/xfs/xfs_inode_item_recover.c
572
error = xfs_recover_inode_owner_change(mp, dip, in_f,
fs/xfs/xfs_inode_item_recover.c
580
XFS_ERRLEVEL_LOW, mp, dip, sizeof(*dip));
fs/xfs/xfs_inode_item_recover.c
581
xfs_alert(mp,
fs/xfs/xfs_inode_item_recover.c
588
ASSERT(bp->b_mount == mp);
fs/xfs/xfs_inode_item_recover.c
78
struct xfs_mount *mp,
fs/xfs/xfs_inode_item_recover.c
88
ip = xfs_inode_alloc(mp, in_f->ilf_ino);
fs/xfs/xfs_ioctl.c
1001
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
1004
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_ioctl.c
1027
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl.c
1030
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl.c
1041
error = xfs_sync_sb_buf(mp, true);
fs/xfs/xfs_ioctl.c
1047
mutex_lock(&mp->m_growlock);
fs/xfs/xfs_ioctl.c
1048
error = xfs_update_secondary_sbs(mp);
fs/xfs/xfs_ioctl.c
1049
mutex_unlock(&mp->m_growlock);
fs/xfs/xfs_ioctl.c
1051
invalidate_bdev(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_ioctl.c
1052
if (xfs_has_rtsb(mp) && mp->m_rtdev_targp)
fs/xfs/xfs_ioctl.c
1053
invalidate_bdev(mp->m_rtdev_targp->bt_bdev);
fs/xfs/xfs_ioctl.c
1112
struct xfs_mount *mp = XFS_I(file_inode(filp))->i_mount;
fs/xfs/xfs_ioctl.c
1120
if (xfs_is_readonly(mp))
fs/xfs/xfs_ioctl.c
1129
error = xfs_reserve_blocks(mp, XC_FREE_BLOCKS, fsop.resblks);
fs/xfs/xfs_ioctl.c
1135
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl.c
1136
fsop.resblks = mp->m_free[XC_FREE_BLOCKS].res_total;
fs/xfs/xfs_ioctl.c
1137
fsop.resblks_avail = mp->m_free[XC_FREE_BLOCKS].res_avail;
fs/xfs/xfs_ioctl.c
1138
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl.c
1147
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
1151
.allocino = percpu_counter_read_positive(&mp->m_icount),
fs/xfs/xfs_ioctl.c
1152
.freeino = percpu_counter_read_positive(&mp->m_ifree),
fs/xfs/xfs_ioctl.c
1153
.freedata = xfs_estimate_freecounter(mp, XC_FREE_BLOCKS) -
fs/xfs/xfs_ioctl.c
1154
xfs_freecounter_unavailable(mp, XC_FREE_BLOCKS),
fs/xfs/xfs_ioctl.c
1155
.freertx = xfs_estimate_freecounter(mp, XC_FREE_RTEXTENTS),
fs/xfs/xfs_ioctl.c
1186
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
1194
return xfs_ioc_trim(mp, arg);
fs/xfs/xfs_ioctl.c
1196
return xfs_ioc_getlabel(mp, arg);
fs/xfs/xfs_ioctl.c
1198
return xfs_ioc_setlabel(filp, mp, arg);
fs/xfs/xfs_ioctl.c
1203
xfs_warn_once(mp,
fs/xfs/xfs_ioctl.c
1236
return xfs_ioc_inumbers(mp, cmd, arg);
fs/xfs/xfs_ioctl.c
1239
return xfs_ioc_fsgeometry(mp, arg, 3);
fs/xfs/xfs_ioctl.c
1241
return xfs_ioc_fsgeometry(mp, arg, 4);
fs/xfs/xfs_ioctl.c
1243
return xfs_ioc_fsgeometry(mp, arg, 5);
fs/xfs/xfs_ioctl.c
1246
return xfs_ioc_ag_geometry(mp, arg);
fs/xfs/xfs_ioctl.c
1248
return xfs_ioc_rtgroup_geometry(mp, arg);
fs/xfs/xfs_ioctl.c
1316
return xfs_ioctl_fs_counts(mp, arg);
fs/xfs/xfs_ioctl.c
1331
error = xfs_growfs_data(mp, &in);
fs/xfs/xfs_ioctl.c
1345
error = xfs_growfs_log(mp, &in);
fs/xfs/xfs_ioctl.c
1359
error = xfs_growfs_rt(mp, &in);
fs/xfs/xfs_ioctl.c
1373
return xfs_fs_goingdown(mp, in);
fs/xfs/xfs_ioctl.c
1385
return xfs_errortag_add(mp, in.errtag);
fs/xfs/xfs_ioctl.c
1392
return xfs_errortag_clearall(mp);
fs/xfs/xfs_ioctl.c
1401
if (xfs_is_readonly(mp))
fs/xfs/xfs_ioctl.c
1411
trace_xfs_ioc_free_eofblocks(mp, &icw, _RET_IP_);
fs/xfs/xfs_ioctl.c
1413
guard(super_write)(mp->m_super);
fs/xfs/xfs_ioctl.c
1414
return xfs_blockgc_free_space(mp, &icw);
fs/xfs/xfs_ioctl.c
175
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
202
breq->startino = mp->m_sb.sb_rootino;
fs/xfs/xfs_ioctl.c
216
if (hdr->agno >= mp->m_sb.sb_agcount)
fs/xfs/xfs_ioctl.c
220
breq->startino = XFS_AGINO_TO_INO(mp, hdr->agno, 0);
fs/xfs/xfs_ioctl.c
221
else if (XFS_INO_TO_AGNO(mp, breq->startino) < hdr->agno)
fs/xfs/xfs_ioctl.c
227
if (XFS_INO_TO_AGNO(mp, breq->startino) > hdr->agno)
fs/xfs/xfs_ioctl.c
233
if (XFS_INO_TO_AGNO(mp, breq->startino) >= mp->m_sb.sb_agcount)
fs/xfs/xfs_ioctl.c
266
struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount;
fs/xfs/xfs_ioctl.c
269
.mp = mp,
fs/xfs/xfs_ioctl.c
277
if (xfs_is_shutdown(mp))
fs/xfs/xfs_ioctl.c
283
error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->bulkstat);
fs/xfs/xfs_ioctl.c
314
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
320
.mp = mp,
fs/xfs/xfs_ioctl.c
327
if (xfs_is_shutdown(mp))
fs/xfs/xfs_ioctl.c
336
error = xfs_bulk_ireq_setup(mp, &hdr, &breq, arg->inumbers);
fs/xfs/xfs_ioctl.c
356
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
363
xfs_fs_geometry(mp, &fsgeo, struct_version);
fs/xfs/xfs_ioctl.c
370
xfs_fsop_geom_health(mp, &fsgeo);
fs/xfs/xfs_ioctl.c
381
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
395
pag = xfs_perag_get(mp, ageo.ag_number);
fs/xfs/xfs_ioctl.c
411
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
424
if (!xfs_has_rtgroups(mp))
fs/xfs/xfs_ioctl.c
427
rtg = xfs_rtgroup_get(mp, rgeo.rg_number);
fs/xfs/xfs_ioctl.c
451
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
457
fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
fs/xfs/xfs_ioctl.c
465
xfs_extlen_to_rtxmod(mp, ip->i_extsize) > 0) {
fs/xfs/xfs_ioctl.c
470
fa->fsx_extsize = XFS_FSB_TO_B(mp, ip->i_extsize);
fs/xfs/xfs_ioctl.c
481
ip->i_cowextsize % mp->m_sb.sb_rextsize > 0) {
fs/xfs/xfs_ioctl.c
485
fa->fsx_cowextsize = XFS_FSB_TO_B(mp, ip->i_cowextsize);
fs/xfs/xfs_ioctl.c
530
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
545
(mp->m_ddev_targp->bt_daxdev == NULL ||
fs/xfs/xfs_ioctl.c
546
(mp->m_rtdev_targp &&
fs/xfs/xfs_ioctl.c
547
mp->m_rtdev_targp->bt_daxdev == NULL)))
fs/xfs/xfs_ioctl.c
553
if (mp->m_sb.sb_rblocks == 0 || mp->m_sb.sb_rextsize == 0 ||
fs/xfs/xfs_ioctl.c
554
xfs_extlen_to_rtxmod(mp, ip->i_extsize))
fs/xfs/xfs_ioctl.c
560
if (i_flags2 && !xfs_has_v3inodes(mp))
fs/xfs/xfs_ioctl.c
577
XFS_STATS_INC(mp, xs_ig_attrchg);
fs/xfs/xfs_ioctl.c
58
xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat);
fs/xfs/xfs_ioctl.c
586
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
592
if (xfs_has_dax_always(mp) || xfs_has_dax_never(mp))
fs/xfs/xfs_ioctl.c
613
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
617
if (xfs_is_readonly(mp))
fs/xfs/xfs_ioctl.c
620
if (xfs_is_shutdown(mp))
fs/xfs/xfs_ioctl.c
628
if (xfs_has_wsync(mp))
fs/xfs/xfs_ioctl.c
646
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
654
XFS_FSB_TO_B(mp, ip->i_extsize) != fa->fsx_extsize)
fs/xfs/xfs_ioctl.c
657
if (fa->fsx_extsize & mp->m_blockmask)
fs/xfs/xfs_ioctl.c
672
rtextsize_bytes = XFS_FSB_TO_B(mp, mp->m_sb.sb_rextsize);
fs/xfs/xfs_ioctl.c
678
XFS_B_TO_FSB(mp, fa->fsx_extsize),
fs/xfs/xfs_ioctl.c
688
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
696
if (fa->fsx_cowextsize & mp->m_blockmask)
fs/xfs/xfs_ioctl.c
703
XFS_B_TO_FSB(mp, fa->fsx_cowextsize),
fs/xfs/xfs_ioctl.c
730
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_ioctl.c
757
if (fa->fsx_valid && XFS_IS_QUOTA_ON(mp)) {
fs/xfs/xfs_ioctl.c
801
if (XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_ioctl.c
814
ip->i_extsize = XFS_B_TO_FSB(mp, fa->fsx_extsize);
fs/xfs/xfs_ioctl.c
818
if (xfs_has_v3inodes(mp)) {
fs/xfs/xfs_ioctl.c
820
ip->i_cowextsize = XFS_B_TO_FSB(mp, fa->fsx_cowextsize);
fs/xfs/xfs_ioctl.c
83
struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount;
fs/xfs/xfs_ioctl.c
86
.mp = mp,
fs/xfs/xfs_ioctl.c
979
struct xfs_mount *mp,
fs/xfs/xfs_ioctl.c
982
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_ioctl.c
989
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl.c
99
if (xfs_is_shutdown(mp))
fs/xfs/xfs_ioctl.c
991
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_ioctl32.c
167
xfs_bulkstat_to_bstat(breq->mp, &bs1, bstat);
fs/xfs/xfs_ioctl32.c
203
struct xfs_mount *mp = XFS_I(file_inode(file))->i_mount;
fs/xfs/xfs_ioctl32.c
207
.mp = mp,
fs/xfs/xfs_ioctl32.c
244
if (xfs_is_shutdown(mp))
fs/xfs/xfs_ioctl32.c
35
struct xfs_mount *mp,
fs/xfs/xfs_ioctl32.c
40
xfs_fs_geometry(mp, &fsgeo, 3);
fs/xfs/xfs_iomap.c
1001
end_fsb = xfs_iomap_end_fsb(mp, offset, length);
fs/xfs/xfs_iomap.c
1019
length = XFS_FSB_TO_B(mp, cmap.br_startoff + cmap.br_blockcount);
fs/xfs/xfs_iomap.c
108
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1127
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1128
const xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
1129
const xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + length);
fs/xfs/xfs_iomap.c
1144
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
1147
if (!xfs_can_sw_atomic_write(mp)) {
fs/xfs/xfs_iomap.c
1148
ASSERT(xfs_can_sw_atomic_write(mp));
fs/xfs/xfs_iomap.c
1188
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_iomap.c
1191
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, resaligned);
fs/xfs/xfs_iomap.c
1195
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_iomap.c
140
if (XFS_IS_REALTIME_INODE(ip) && xfs_has_rtgroups(mp) &&
fs/xfs/xfs_iomap.c
141
xfs_rtbno_is_group_start(mp, imap->br_startblock))
fs/xfs/xfs_iomap.c
144
iomap->offset = XFS_FSB_TO_B(mp, imap->br_startoff);
fs/xfs/xfs_iomap.c
145
iomap->length = XFS_FSB_TO_B(mp, imap->br_blockcount);
fs/xfs/xfs_iomap.c
1467
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1505
error = xfs_bmap_extsize_align(mp, got, &prev, extsz, 0, eof,
fs/xfs/xfs_iomap.c
1529
error = xfs_dec_frextents(mp, xfs_blen_to_rtbxlen(mp, alen));
fs/xfs/xfs_iomap.c
1536
error = xfs_dec_fdblocks(mp, fdblocks, false);
fs/xfs/xfs_iomap.c
1564
xfs_add_frextents(mp, xfs_blen_to_rtbxlen(mp, alen));
fs/xfs/xfs_iomap.c
1566
if (XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_iomap.c
1595
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1596
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
1597
xfs_fileoff_t end_fsb = xfs_iomap_end_fsb(mp, offset, count);
fs/xfs/xfs_iomap.c
1610
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
1621
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(&ip->i_df)) ||
fs/xfs/xfs_iomap.c
1622
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/xfs_iomap.c
1628
XFS_STATS_INC(mp, xs_blk_mapw);
fs/xfs/xfs_iomap.c
1644
if (!IS_ALIGNED(offset, mp->m_sb.sb_blocksize) ||
fs/xfs/xfs_iomap.c
1645
!IS_ALIGNED(offset + count, mp->m_sb.sb_blocksize) ||
fs/xfs/xfs_iomap.c
1691
XFS_B_TO_FSB(mp, 1024 * PAGE_SIZE));
fs/xfs/xfs_iomap.c
1713
xfs_warn_ratelimited(mp,
fs/xfs/xfs_iomap.c
1728
error = xfs_dec_fdblocks(mp, indlen, false);
fs/xfs/xfs_iomap.c
1742
trace_xfs_iomap_alloc(ip, offset, XFS_FSB_TO_B(mp, count_fsb),
fs/xfs/xfs_iomap.c
1764
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
1765
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
1766
xfs_fileoff_t end_fsb = xfs_iomap_end_fsb(mp, offset, count);
fs/xfs/xfs_iomap.c
1777
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
1797
if (XFS_IS_CORRUPT(mp, !xfs_ifork_has_extents(&ip->i_df)) ||
fs/xfs/xfs_iomap.c
1798
XFS_TEST_ERROR(mp, XFS_ERRTAG_BMAPIFORMAT)) {
fs/xfs/xfs_iomap.c
1804
XFS_STATS_INC(mp, xs_blk_mapw);
fs/xfs/xfs_iomap.c
1833
xfs_fileoff_t eof_fsb = XFS_B_TO_FSB(mp, XFS_ISIZE(ip));
fs/xfs/xfs_iomap.c
1856
min(count, XFS_FSB_TO_B(mp, imap.br_blockcount));
fs/xfs/xfs_iomap.c
1860
XFS_B_TO_FSB(mp, foffset));
fs/xfs/xfs_iomap.c
188
struct xfs_mount *mp,
fs/xfs/xfs_iomap.c
192
ASSERT(offset <= mp->m_super->s_maxbytes);
fs/xfs/xfs_iomap.c
193
return min(XFS_B_TO_FSB(mp, offset + count),
fs/xfs/xfs_iomap.c
1931
end_fsb = xfs_iomap_end_fsb(mp, offset, count);
fs/xfs/xfs_iomap.c
194
XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes));
fs/xfs/xfs_iomap.c
1942
if (xfs_has_allocsize(mp))
fs/xfs/xfs_iomap.c
1943
prealloc_blocks = mp->m_allocsize_blocks;
fs/xfs/xfs_iomap.c
1955
end_offset = XFS_ALLOC_ALIGN(mp, offset + count - 1);
fs/xfs/xfs_iomap.c
1956
p_end_fsb = XFS_B_TO_FSBT(mp, end_offset) +
fs/xfs/xfs_iomap.c
1964
XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes));
fs/xfs/xfs_iomap.c
201
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2105
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2107
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
2108
xfs_fileoff_t end_fsb = xfs_iomap_end_fsb(mp, offset, length);
fs/xfs/xfs_iomap.c
2116
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
213
if (mp->m_swidth && xfs_has_swalloc(mp))
fs/xfs/xfs_iomap.c
214
align = mp->m_swidth;
fs/xfs/xfs_iomap.c
215
else if (mp->m_dalign)
fs/xfs/xfs_iomap.c
2150
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2151
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
2152
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + length);
fs/xfs/xfs_iomap.c
216
align = mp->m_dalign;
fs/xfs/xfs_iomap.c
2160
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
2179
data_fsb = xfs_iomap_end_fsb(mp, offset, length);
fs/xfs/xfs_iomap.c
218
if (align && XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, align))
fs/xfs/xfs_iomap.c
2238
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
2239
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
2240
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + length);
fs/xfs/xfs_iomap.c
2246
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
273
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
288
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_iomap.c
291
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, resaligned);
fs/xfs/xfs_iomap.c
317
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, 0) << 1;
fs/xfs/xfs_iomap.c
321
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_iomap.c
36
#define XFS_ALLOC_ALIGN(mp, off) \
fs/xfs/xfs_iomap.c
37
(((off) >> mp->m_allocsize_log) << mp->m_allocsize_log)
fs/xfs/xfs_iomap.c
447
struct xfs_mount *mp,
fs/xfs/xfs_iomap.c
454
freesp = xfs_estimate_freecounter(mp, idx);
fs/xfs/xfs_iomap.c
485
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
487
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
501
if (XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, mp->m_allocsize_blocks))
fs/xfs/xfs_iomap.c
508
if (XFS_ISIZE(ip) < XFS_FSB_TO_B(mp, mp->m_dalign) ||
fs/xfs/xfs_iomap.c
511
return mp->m_allocsize_blocks;
fs/xfs/xfs_iomap.c
538
alloc_blocks = XFS_B_TO_FSB(mp, offset);
fs/xfs/xfs_iomap.c
552
freesp = xfs_rtbxlen_to_blen(mp,
fs/xfs/xfs_iomap.c
553
xfs_iomap_freesp(mp, XC_FREE_RTEXTENTS,
fs/xfs/xfs_iomap.c
554
mp->m_low_rtexts, &shift));
fs/xfs/xfs_iomap.c
556
freesp = xfs_iomap_freesp(mp, XC_FREE_BLOCKS, mp->m_low_space,
fs/xfs/xfs_iomap.c
602
if (alloc_blocks < mp->m_allocsize_blocks)
fs/xfs/xfs_iomap.c
603
alloc_blocks = mp->m_allocsize_blocks;
fs/xfs/xfs_iomap.c
605
mp->m_allocsize_blocks);
fs/xfs/xfs_iomap.c
616
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
630
offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
631
count_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + count);
fs/xfs/xfs_iomap.c
644
resblks = XFS_DIOSTRAT_SPACE_RES(mp, 0) << 1;
fs/xfs/xfs_iomap.c
661
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, resblks,
fs/xfs/xfs_iomap.c
686
i_size = XFS_FSB_TO_B(mp, offset_fsb + count_fsb);
fs/xfs/xfs_iomap.c
819
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
820
xfs_fsize_t len = XFS_FSB_TO_B(mp, end_fsb - offset_fsb);
fs/xfs/xfs_iomap.c
854
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iomap.c
856
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_iomap.c
857
xfs_fileoff_t end_fsb = xfs_iomap_end_fsb(mp, offset, length);
fs/xfs/xfs_iomap.c
868
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iomap.c
933
length = XFS_FSB_TO_B(mp, end_fsb) - offset;
fs/xfs/xfs_iomap.c
977
((offset | length) & mp->m_blockmask))
fs/xfs/xfs_iops.c
1000
xfs_zoned_space_unreserve(mp, &ac);
fs/xfs/xfs_iops.c
1051
resblks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_iops.c
1053
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, resblks,
fs/xfs/xfs_iops.c
1116
XFS_STATS_INC(mp, xs_ig_attrchg);
fs/xfs/xfs_iops.c
1118
if (xfs_has_wsync(mp))
fs/xfs/xfs_iops.c
1191
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1216
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_fsyncts, 0, 0, 0, &tp);
fs/xfs/xfs_iops.c
1235
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1238
if (xfs_trans_alloc(mp, &M_RES(mp)->tr_fsyncts, 0, 0, 0, &tp))
fs/xfs/xfs_iops.c
1366
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
1373
if (mp->m_sb.sb_blocksize != PAGE_SIZE)
fs/xfs/xfs_iops.c
542
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
549
return XFS_FSB_TO_B(mp, xfs_get_extsz_hint(ip) ? : 1);
fs/xfs/xfs_iops.c
564
if (xfs_has_large_iosize(mp)) {
fs/xfs/xfs_iops.c
565
if (mp->m_swidth)
fs/xfs/xfs_iops.c
566
return XFS_FSB_TO_B(mp, mp->m_swidth);
fs/xfs/xfs_iops.c
567
if (xfs_has_allocsize(mp))
fs/xfs/xfs_iops.c
568
return 1U << mp->m_allocsize_log;
fs/xfs/xfs_iops.c
571
return max_t(uint32_t, PAGE_SIZE, mp->m_sb.sb_blocksize);
fs/xfs/xfs_iops.c
606
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
619
return mp->m_sb.sb_blocksize;
fs/xfs/xfs_iops.c
628
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
637
return mp->m_sb.sb_blocksize;
fs/xfs/xfs_iops.c
647
return XFS_FSB_TO_B(mp, mp->m_groups[XG_TYPE_RTG].awu_max);
fs/xfs/xfs_iops.c
648
return XFS_FSB_TO_B(mp, mp->m_groups[XG_TYPE_AG].awu_max);
fs/xfs/xfs_iops.c
691
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
697
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iops.c
711
stat->blocks = XFS_FSB_TO_BB(mp, ip->i_nblocks + ip->i_delayed_blks);
fs/xfs/xfs_iops.c
713
if (xfs_has_v3inodes(mp)) {
fs/xfs/xfs_iops.c
762
struct xfs_mount *mp = XFS_I(d_inode(dentry))->i_mount;
fs/xfs/xfs_iops.c
764
if (xfs_is_readonly(mp))
fs/xfs/xfs_iops.c
767
if (xfs_is_shutdown(mp))
fs/xfs/xfs_iops.c
786
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_iops.c
806
if (XFS_IS_QUOTA_ON(mp) && (mask & (ATTR_UID|ATTR_GID))) {
fs/xfs/xfs_iops.c
809
if ((mask & ATTR_UID) && XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_iops.c
816
if ((mask & ATTR_GID) && XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_iops.c
849
if (XFS_IS_UQUOTA_ON(mp) &&
fs/xfs/xfs_iops.c
854
if (XFS_IS_GQUOTA_ON(mp) &&
fs/xfs/xfs_iops.c
856
ASSERT(xfs_has_pquotino(mp) || !XFS_IS_PQUOTA_ON(mp));
fs/xfs/xfs_iops.c
864
XFS_STATS_INC(mp, xs_ig_attrchg);
fs/xfs/xfs_iops.c
866
if (xfs_has_wsync(mp))
fs/xfs/xfs_iops.c
915
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_iops.c
972
error = xfs_zoned_space_reserve(mp, 1,
fs/xfs/xfs_itable.c
126
if (IS_PRIVATE(inode) || xfs_is_sb_inum(mp, ino)) {
fs/xfs/xfs_itable.c
166
if (xfs_has_v3inodes(mp)) {
fs/xfs/xfs_itable.c
181
buf->bs_blksize = mp->m_sb.sb_blocksize;
fs/xfs/xfs_itable.c
187
buf->bs_blksize = mp->m_sb.sb_blocksize;
fs/xfs/xfs_itable.c
226
xfs_warn_ratelimited(breq->mp,
fs/xfs/xfs_itable.c
242
tp = xfs_trans_alloc_empty(breq->mp);
fs/xfs/xfs_itable.c
243
error = xfs_bulkstat_one_int(breq->mp, breq->idmap, tp,
fs/xfs/xfs_itable.c
260
struct xfs_mount *mp,
fs/xfs/xfs_itable.c
268
error = xfs_bulkstat_one_int(mp, bc->breq->idmap, tp, ino, data);
fs/xfs/xfs_itable.c
289
struct xfs_mount *mp,
fs/xfs/xfs_itable.c
292
xfs_agnumber_t agno = XFS_INO_TO_AGNO(mp, startino);
fs/xfs/xfs_itable.c
293
xfs_agino_t agino = XFS_INO_TO_AGINO(mp, startino);
fs/xfs/xfs_itable.c
295
return agno >= mp->m_sb.sb_agcount ||
fs/xfs/xfs_itable.c
296
startino != XFS_AGINO_TO_INO(mp, agno, agino);
fs/xfs/xfs_itable.c
313
xfs_warn_ratelimited(breq->mp,
fs/xfs/xfs_itable.c
317
if (xfs_bulkstat_already_done(breq->mp, breq->startino))
fs/xfs/xfs_itable.c
329
tp = xfs_trans_alloc_empty(breq->mp);
fs/xfs/xfs_itable.c
330
error = xfs_iwalk(breq->mp, tp, breq->startino, breq->iwalk_flags,
fs/xfs/xfs_itable.c
351
struct xfs_mount *mp,
fs/xfs/xfs_itable.c
373
bs1->bs_extsize = XFS_FSB_TO_B(mp, bstat->bs_extsize_blks);
fs/xfs/xfs_itable.c
381
bs1->bs_cowextsize = XFS_FSB_TO_B(mp, bstat->bs_cowextsize_blks);
fs/xfs/xfs_itable.c
409
struct xfs_mount *mp,
fs/xfs/xfs_itable.c
416
.xi_startino = XFS_AGINO_TO_INO(mp, agno, irec->ir_startino),
fs/xfs/xfs_itable.c
428
ic->breq->startino = XFS_AGINO_TO_INO(mp, agno, irec->ir_startino) +
fs/xfs/xfs_itable.c
448
if (xfs_bulkstat_already_done(breq->mp, breq->startino))
fs/xfs/xfs_itable.c
455
tp = xfs_trans_alloc_empty(breq->mp);
fs/xfs/xfs_itable.c
456
error = xfs_inobt_walk(breq->mp, tp, breq->startino, breq->iwalk_flags,
fs/xfs/xfs_itable.c
65
struct xfs_mount *mp,
fs/xfs/xfs_itable.c
71
struct user_namespace *sb_userns = mp->m_super->s_user_ns;
fs/xfs/xfs_itable.c
80
error = xfs_iget(mp, tp, ino,
fs/xfs/xfs_itable.c
93
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_itable.h
10
struct xfs_mount *mp;
fs/xfs/xfs_itable.h
59
void xfs_bulkstat_to_bstat(struct xfs_mount *mp, struct xfs_bstat *bs1,
fs/xfs/xfs_iunlink_item.c
146
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_iunlink_item.c
164
xfs_log_item_init(mp, &iup->item, XFS_LI_IUNLINK,
fs/xfs/xfs_iwalk.c
102
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/xfs_iwalk.c
107
agbno = XFS_AGINO_TO_AGBNO(mp, irec->ir_startino);
fs/xfs/xfs_iwalk.c
115
xfs_buf_readahead(mp->m_ddev_targp,
fs/xfs/xfs_iwalk.c
117
igeo->blocks_per_cluster * mp->m_bsize,
fs/xfs/xfs_iwalk.c
176
struct xfs_mount *mp = iwag->mp;
fs/xfs/xfs_iwalk.c
191
error = iwag->inobt_walk_fn(mp, tp, pag_agno(pag), irec,
fs/xfs/xfs_iwalk.c
209
error = iwag->iwalk_fn(mp, tp,
fs/xfs/xfs_iwalk.c
262
struct xfs_mount *mp = iwag->mp;
fs/xfs/xfs_iwalk.c
302
if (XFS_IS_CORRUPT(mp, *has_more != 1)) {
fs/xfs/xfs_iwalk.c
352
struct xfs_mount *mp = iwag->mp;
fs/xfs/xfs_iwalk.c
356
next_agino = XFS_INO_TO_AGINO(mp, iwag->lastino) + 1;
fs/xfs/xfs_iwalk.c
381
iwag->tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_iwalk.c
396
struct xfs_mount *mp = iwag->mp;
fs/xfs/xfs_iwalk.c
405
ASSERT(pag_agno(pag) == XFS_INO_TO_AGNO(mp, iwag->startino));
fs/xfs/xfs_iwalk.c
406
agino = XFS_INO_TO_AGINO(mp, iwag->startino);
fs/xfs/xfs_iwalk.c
426
XFS_IS_CORRUPT(mp, iwag->lastino >= rec_fsino)) {
fs/xfs/xfs_iwalk.c
446
xfs_iwalk_ichunk_ra(mp, pag, irec);
fs/xfs/xfs_iwalk.c
539
struct xfs_mount *mp = iwag->mp;
fs/xfs/xfs_iwalk.c
54
struct xfs_mount *mp;
fs/xfs/xfs_iwalk.c
543
start_agno = XFS_INO_TO_AGNO(iwag->mp, iwag->startino);
fs/xfs/xfs_iwalk.c
544
ASSERT(start_agno < iwag->mp->m_sb.sb_agcount);
fs/xfs/xfs_iwalk.c
551
while ((iwag->pag = xfs_perag_next_from(mp, iwag->pag, start_agno))) {
fs/xfs/xfs_iwalk.c
558
XFS_AGINO_TO_INO(mp, pag_agno(iwag->pag) + 1, 0);
fs/xfs/xfs_iwalk.c
573
struct xfs_mount *mp,
fs/xfs/xfs_iwalk.c
582
.mp = mp,
fs/xfs/xfs_iwalk.c
600
struct xfs_mount *mp,
fs/xfs/xfs_iwalk.c
617
iwag->tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_iwalk.c
636
struct xfs_mount *mp,
fs/xfs/xfs_iwalk.c
644
xfs_agnumber_t start_agno = XFS_INO_TO_AGNO(mp, startino);
fs/xfs/xfs_iwalk.c
649
ASSERT(start_agno < mp->m_sb.sb_agcount);
fs/xfs/xfs_iwalk.c
652
error = xfs_pwork_init(mp, &pctl, xfs_iwalk_ag_work, "xfs_iwalk");
fs/xfs/xfs_iwalk.c
656
while ((pag = xfs_perag_next_from(mp, pag, start_agno))) {
fs/xfs/xfs_iwalk.c
664
iwag->mp = mp;
fs/xfs/xfs_iwalk.c
677
startino = XFS_AGINO_TO_INO(mp, pag_agno(pag) + 1, 0);
fs/xfs/xfs_iwalk.c
734
struct xfs_mount *mp,
fs/xfs/xfs_iwalk.c
743
.mp = mp,
fs/xfs/xfs_iwalk.c
98
struct xfs_mount *mp,
fs/xfs/xfs_iwalk.h
18
typedef int (*xfs_iwalk_fn)(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/xfs_iwalk.h
21
int xfs_iwalk(struct xfs_mount *mp, struct xfs_trans *tp, xfs_ino_t startino,
fs/xfs/xfs_iwalk.h
24
int xfs_iwalk_threaded(struct xfs_mount *mp, xfs_ino_t startino,
fs/xfs/xfs_iwalk.h
34
typedef int (*xfs_inobt_walk_fn)(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/xfs_iwalk.h
39
int xfs_inobt_walk(struct xfs_mount *mp, struct xfs_trans *tp,
fs/xfs/xfs_log.c
1002
struct xfs_mount *mp)
fs/xfs/xfs_log.c
1004
xfs_log_clean(mp);
fs/xfs/xfs_log.c
1012
xlog_wait_iclog_completion(mp->m_log);
fs/xfs/xfs_log.c
1014
xfs_buftarg_drain(mp->m_ddev_targp);
fs/xfs/xfs_log.c
1016
xfs_trans_ail_destroy(mp);
fs/xfs/xfs_log.c
1018
xfs_sysfs_del(&mp->m_log->l_kobj);
fs/xfs/xfs_log.c
1020
xlog_dealloc_log(mp->m_log);
fs/xfs/xfs_log.c
1025
struct xfs_mount *mp,
fs/xfs/xfs_log.c
1030
item->li_log = mp->m_log;
fs/xfs/xfs_log.c
1031
item->li_ailp = mp->m_ail;
fs/xfs/xfs_log.c
1047
struct xfs_mount *mp)
fs/xfs/xfs_log.c
1049
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
1093
struct xfs_mount *mp)
fs/xfs/xfs_log.c
1095
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
1136
struct xfs_mount *mp)
fs/xfs/xfs_log.c
1141
if (!xlog_is_shutdown(mp->m_log)) {
fs/xfs/xfs_log.c
1142
ASSERT(xlog_cil_empty(mp->m_log));
fs/xfs/xfs_log.c
1143
ASSERT(xlog_iclogs_empty(mp->m_log));
fs/xfs/xfs_log.c
1144
ASSERT(!xfs_ail_min_lsn(mp->m_log->l_ailp));
fs/xfs/xfs_log.c
1147
if (!xfs_log_writable(mp))
fs/xfs/xfs_log.c
1161
need_covered = xfs_log_need_covered(mp);
fs/xfs/xfs_log.c
1162
if (!need_covered && !xfs_has_lazysbcount(mp))
fs/xfs/xfs_log.c
1175
error = xfs_sync_sb(mp, true);
fs/xfs/xfs_log.c
1178
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/xfs_log.c
1179
} while (xfs_log_need_covered(mp));
fs/xfs/xfs_log.c
1230
struct xfs_mount *mp,
fs/xfs/xfs_log.c
1233
if (mp->m_logbufs <= 0)
fs/xfs/xfs_log.c
1234
mp->m_logbufs = XLOG_MAX_ICLOGS;
fs/xfs/xfs_log.c
1235
if (mp->m_logbsize <= 0)
fs/xfs/xfs_log.c
1236
mp->m_logbsize = XLOG_BIG_RECORD_BSIZE;
fs/xfs/xfs_log.c
1238
log->l_iclog_bufs = mp->m_logbufs;
fs/xfs/xfs_log.c
1239
log->l_iclog_size = mp->m_logbsize;
fs/xfs/xfs_log.c
1247
DIV_ROUND_UP(mp->m_logbsize, XLOG_HEADER_CYCLE_SIZE) - 1);
fs/xfs/xfs_log.c
1252
struct xfs_mount *mp)
fs/xfs/xfs_log.c
1254
queue_delayed_work(mp->m_sync_workqueue, &mp->m_log->l_work,
fs/xfs/xfs_log.c
1268
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_log.c
1270
if (!xfs_sb_has_incompat_log_feature(&mp->m_sb,
fs/xfs/xfs_log.c
1277
xfs_clear_incompat_log_features(mp);
fs/xfs/xfs_log.c
1291
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_log.c
1294
if (xfs_fs_writable(mp, SB_FREEZE_WRITE) && xfs_log_need_covered(mp)) {
fs/xfs/xfs_log.c
1307
xfs_sync_sb(mp, true);
fs/xfs/xfs_log.c
1309
xfs_log_force(mp, 0);
fs/xfs/xfs_log.c
1312
xfs_ail_push_all(mp->m_ail);
fs/xfs/xfs_log.c
1315
xfs_log_work_queue(mp);
fs/xfs/xfs_log.c
1325
struct xfs_mount *mp,
fs/xfs/xfs_log.c
1339
xfs_warn(mp, "Log allocation failed: No memory!");
fs/xfs/xfs_log.c
1343
log->l_mp = mp;
fs/xfs/xfs_log.c
1358
if (xfs_has_logv2(mp) && mp->m_sb.sb_logsunit > 1)
fs/xfs/xfs_log.c
1359
log->l_iclog_roundoff = mp->m_sb.sb_logsunit;
fs/xfs/xfs_log.c
1360
else if (mp->m_sb.sb_logsectsize > 0)
fs/xfs/xfs_log.c
1361
log->l_iclog_roundoff = mp->m_sb.sb_logsectsize;
fs/xfs/xfs_log.c
1369
if (xfs_has_sector(mp)) {
fs/xfs/xfs_log.c
1370
log2_size = mp->m_sb.sb_logsectlog;
fs/xfs/xfs_log.c
1372
xfs_warn(mp, "Log sector size too small (0x%x < 0x%x)",
fs/xfs/xfs_log.c
1378
if (log2_size > mp->m_sectbb_log) {
fs/xfs/xfs_log.c
1379
xfs_warn(mp, "Log sector size too large (0x%x > 0x%x)",
fs/xfs/xfs_log.c
1380
log2_size, mp->m_sectbb_log);
fs/xfs/xfs_log.c
1386
!xfs_has_logv2(mp)) {
fs/xfs/xfs_log.c
1387
xfs_warn(mp,
fs/xfs/xfs_log.c
1395
xlog_get_iclog_buffer_size(mp, log);
fs/xfs/xfs_log.c
1425
memcpy(&iclog->ic_header->h_fs_uuid, &mp->m_sb.sb_uuid,
fs/xfs/xfs_log.c
1447
0, mp->m_super->s_id);
fs/xfs/xfs_log.c
1825
struct xfs_mount *mp,
fs/xfs/xfs_log.c
1828
xfs_warn(mp, "ticket reservation summary:");
fs/xfs/xfs_log.c
1829
xfs_warn(mp, " unit res = %d bytes", ticket->t_unit_res);
fs/xfs/xfs_log.c
1830
xfs_warn(mp, " current res = %d bytes", ticket->t_curr_res);
fs/xfs/xfs_log.c
1831
xfs_warn(mp, " original count = %d", ticket->t_ocnt);
fs/xfs/xfs_log.c
1832
xfs_warn(mp, " remaining count = %d", ticket->t_cnt);
fs/xfs/xfs_log.c
1842
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_log.c
1846
xfs_warn(mp, "transaction summary:");
fs/xfs/xfs_log.c
1847
xfs_warn(mp, " log res = %d", tp->t_log_res);
fs/xfs/xfs_log.c
1848
xfs_warn(mp, " log count = %d", tp->t_log_count);
fs/xfs/xfs_log.c
1849
xfs_warn(mp, " flags = 0x%x", tp->t_flags);
fs/xfs/xfs_log.c
1851
xlog_print_tic_res(mp, tp->t_ticket);
fs/xfs/xfs_log.c
1859
xfs_warn(mp, "log item: ");
fs/xfs/xfs_log.c
1860
xfs_warn(mp, " type = 0x%x", lip->li_type);
fs/xfs/xfs_log.c
1861
xfs_warn(mp, " flags = 0x%lx", lip->li_flags);
fs/xfs/xfs_log.c
1864
xfs_warn(mp, " niovecs = %d", lv->lv_niovecs);
fs/xfs/xfs_log.c
1865
xfs_warn(mp, " alloc_size = %d", lv->lv_alloc_size);
fs/xfs/xfs_log.c
1866
xfs_warn(mp, " bytes = %d", lv->lv_bytes);
fs/xfs/xfs_log.c
1867
xfs_warn(mp, " buf used= %d", lv->lv_buf_used);
fs/xfs/xfs_log.c
1874
xfs_warn(mp, " iovec[%d]", i);
fs/xfs/xfs_log.c
1875
xfs_warn(mp, " type = 0x%x", vec->i_type);
fs/xfs/xfs_log.c
1876
xfs_warn(mp, " len = %d", vec->i_len);
fs/xfs/xfs_log.c
1877
xfs_warn(mp, " first %d bytes of iovec[%d]:", dumplen, i);
fs/xfs/xfs_log.c
2793
struct xfs_mount *mp,
fs/xfs/xfs_log.c
2796
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
2799
XFS_STATS_INC(mp, xs_log_force);
fs/xfs/xfs_log.c
2800
trace_xfs_log_force(mp, 0, _RET_IP_);
fs/xfs/xfs_log.c
288
struct xfs_mount *mp)
fs/xfs/xfs_log.c
296
if (xfs_has_norecovery(mp))
fs/xfs/xfs_log.c
2971
struct xfs_mount *mp,
fs/xfs/xfs_log.c
2976
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
298
if (xfs_readonly_buftarg(mp->m_ddev_targp))
fs/xfs/xfs_log.c
2981
XFS_STATS_INC(mp, xs_log_force);
fs/xfs/xfs_log.c
2982
trace_xfs_log_force(mp, seq, _RET_IP_);
fs/xfs/xfs_log.c
2990
XFS_STATS_INC(mp, xs_log_force_sleep);
fs/xfs/xfs_log.c
300
if (xfs_readonly_buftarg(mp->m_log->l_targ))
fs/xfs/xfs_log.c
302
if (xlog_is_shutdown(mp->m_log))
fs/xfs/xfs_log.c
3112
struct xfs_mount *mp,
fs/xfs/xfs_log.c
3115
return xlog_calc_unit_res(mp->m_log, unit_bytes, NULL);
fs/xfs/xfs_log.c
312
struct xfs_mount *mp,
fs/xfs/xfs_log.c
315
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
322
XFS_STATS_INC(mp, xs_try_logspace);
fs/xfs/xfs_log.c
3442
struct xfs_mount *mp,
fs/xfs/xfs_log.c
3445
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
3453
if (xfs_has_norecovery(mp))
fs/xfs/xfs_log.c
3463
valid = xlog_valid_lsn(mp->m_log, lsn);
fs/xfs/xfs_log.c
3468
xfs_warn(mp,
fs/xfs/xfs_log.c
367
struct xfs_mount *mp,
fs/xfs/xfs_log.c
373
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
381
XFS_STATS_INC(mp, xs_try_logspace);
fs/xfs/xfs_log.c
39
struct xfs_mount *mp,
fs/xfs/xfs_log.c
535
xfs_mount_t *mp,
fs/xfs/xfs_log.c
544
if (!xfs_has_norecovery(mp)) {
fs/xfs/xfs_log.c
545
xfs_notice(mp, "Mounting V%d Filesystem %pU",
fs/xfs/xfs_log.c
546
XFS_SB_VERSION_NUM(&mp->m_sb),
fs/xfs/xfs_log.c
547
&mp->m_sb.sb_uuid);
fs/xfs/xfs_log.c
549
xfs_notice(mp,
fs/xfs/xfs_log.c
551
XFS_SB_VERSION_NUM(&mp->m_sb),
fs/xfs/xfs_log.c
552
&mp->m_sb.sb_uuid);
fs/xfs/xfs_log.c
553
ASSERT(xfs_is_readonly(mp));
fs/xfs/xfs_log.c
556
log = xlog_alloc_log(mp, log_target, blk_offset, num_bblks);
fs/xfs/xfs_log.c
561
mp->m_log = log;
fs/xfs/xfs_log.c
581
min_logfsbs = xfs_log_calc_minimum_size(mp);
fs/xfs/xfs_log.c
582
if (mp->m_sb.sb_logblocks < min_logfsbs) {
fs/xfs/xfs_log.c
583
xfs_warn(mp,
fs/xfs/xfs_log.c
585
mp->m_sb.sb_logblocks, min_logfsbs);
fs/xfs/xfs_log.c
591
if (xfs_has_crc(mp)) {
fs/xfs/xfs_log.c
592
xfs_crit(mp, "AAIEEE! Log failed size checks. Abort!");
fs/xfs/xfs_log.c
597
xfs_crit(mp, "Log size out of supported range.");
fs/xfs/xfs_log.c
598
xfs_crit(mp,
fs/xfs/xfs_log.c
605
error = xfs_trans_ail_init(mp);
fs/xfs/xfs_log.c
607
xfs_warn(mp, "AIL initialisation failed: error %d", error);
fs/xfs/xfs_log.c
610
log->l_ailp = mp->m_ail;
fs/xfs/xfs_log.c
616
if (!xfs_has_norecovery(mp)) {
fs/xfs/xfs_log.c
619
xfs_warn(mp, "log mount/recovery failed: error %d",
fs/xfs/xfs_log.c
626
error = xfs_sysfs_init(&log->l_kobj, &xfs_log_ktype, &mp->m_kobj,
fs/xfs/xfs_log.c
644
xfs_trans_ail_destroy(mp);
fs/xfs/xfs_log.c
663
struct xfs_mount *mp)
fs/xfs/xfs_log.c
665
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
668
if (xfs_has_norecovery(mp)) {
fs/xfs/xfs_log.c
669
ASSERT(xfs_is_readonly(mp));
fs/xfs/xfs_log.c
693
mp->m_super->s_flags |= SB_ACTIVE;
fs/xfs/xfs_log.c
694
xfs_log_work_queue(mp);
fs/xfs/xfs_log.c
697
mp->m_super->s_flags &= ~SB_ACTIVE;
fs/xfs/xfs_log.c
698
evict_inodes(mp->m_super);
fs/xfs/xfs_log.c
711
xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_log.c
712
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/xfs_log.c
714
xfs_notice(mp, "Ending recovery (logdev: %s)",
fs/xfs/xfs_log.c
715
mp->m_logname ? mp->m_logname : "internal");
fs/xfs/xfs_log.c
717
xfs_info(mp, "Ending clean mount");
fs/xfs/xfs_log.c
719
xfs_buftarg_drain(mp->m_ddev_targp);
fs/xfs/xfs_log.c
735
struct xfs_mount *mp)
fs/xfs/xfs_log.c
737
xlog_recover_cancel(mp->m_log);
fs/xfs/xfs_log.c
738
xfs_log_unmount(mp);
fs/xfs/xfs_log.c
862
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_log.c
867
error = xfs_log_reserve(mp, 600, 1, &tic, 0);
fs/xfs/xfs_log.c
878
xfs_alert(mp, "%s: unmount record failed", __func__);
fs/xfs/xfs_log.c
912
struct xfs_mount *mp)
fs/xfs/xfs_log.c
914
struct xlog *log = mp->m_log;
fs/xfs/xfs_log.c
916
if (!xfs_log_writable(mp))
fs/xfs/xfs_log.c
919
xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_log.c
930
if (xfs_fs_has_sickness(mp, XFS_SICK_FS_COUNTERS) ||
fs/xfs/xfs_log.c
931
XFS_TEST_ERROR(mp, XFS_ERRTAG_FORCE_SUMMARY_RECALC)) {
fs/xfs/xfs_log.c
932
xfs_alert(mp, "%s: will fix summary counters at next mount",
fs/xfs/xfs_log.c
951
struct xfs_mount *mp)
fs/xfs/xfs_log.c
958
if (xfs_clear_incompat_log_features(mp)) {
fs/xfs/xfs_log.c
961
error = xfs_sync_sb(mp, false);
fs/xfs/xfs_log.c
963
xfs_warn(mp,
fs/xfs/xfs_log.c
967
cancel_delayed_work_sync(&mp->m_log->l_work);
fs/xfs/xfs_log.c
968
xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_log.c
977
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/xfs_log.c
978
xfs_buftarg_wait(mp->m_ddev_targp);
fs/xfs/xfs_log.c
979
xfs_buf_lock(mp->m_sb_bp);
fs/xfs/xfs_log.c
980
xfs_buf_unlock(mp->m_sb_bp);
fs/xfs/xfs_log.c
982
return xfs_log_cover(mp);
fs/xfs/xfs_log.c
987
struct xfs_mount *mp)
fs/xfs/xfs_log.c
989
xfs_log_quiesce(mp);
fs/xfs/xfs_log.c
990
xfs_log_unmount_write(mp);
fs/xfs/xfs_log.h
116
int xfs_log_force(struct xfs_mount *mp, uint flags);
fs/xfs/xfs_log.h
117
int xfs_log_force_seq(struct xfs_mount *mp, xfs_csn_t seq, uint flags,
fs/xfs/xfs_log.h
119
int xfs_log_mount(struct xfs_mount *mp,
fs/xfs/xfs_log.h
123
int xfs_log_mount_finish(struct xfs_mount *mp);
fs/xfs/xfs_log.h
125
xfs_lsn_t xlog_assign_tail_lsn(struct xfs_mount *mp);
fs/xfs/xfs_log.h
126
xfs_lsn_t xlog_assign_tail_lsn_locked(struct xfs_mount *mp);
fs/xfs/xfs_log.h
127
void xfs_log_space_wake(struct xfs_mount *mp);
fs/xfs/xfs_log.h
128
int xfs_log_reserve(struct xfs_mount *mp, int length, int count,
fs/xfs/xfs_log.h
130
int xfs_log_regrant(struct xfs_mount *mp, struct xlog_ticket *tic);
fs/xfs/xfs_log.h
131
void xfs_log_unmount(struct xfs_mount *mp);
fs/xfs/xfs_log.h
132
bool xfs_log_writable(struct xfs_mount *mp);
fs/xfs/xfs_log.h
140
void xfs_log_work_queue(struct xfs_mount *mp);
fs/xfs/xfs_log.h
141
int xfs_log_quiesce(struct xfs_mount *mp);
fs/xfs/xfs_log.h
142
void xfs_log_clean(struct xfs_mount *mp);
fs/xfs/xfs_log_cil.c
1009
xfs_discard_extents(mp, &ctx->busy_extents);
fs/xfs/xfs_log_cil.c
978
struct xfs_mount *mp = ctx->cil->xc_log->l_mp;
fs/xfs/xfs_log_cil.c
999
xfs_has_discard(mp) && !abort);
fs/xfs/xfs_log_priv.h
523
void xlog_print_tic_res(struct xfs_mount *mp, struct xlog_ticket *ticket);
fs/xfs/xfs_log_recover.c
1744
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
1750
error = xfs_iget(mp, NULL, ino, 0, 0, ipp);
fs/xfs/xfs_log_recover.c
1776
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
1784
error = xlog_recover_iget(mp, ino, &ip);
fs/xfs/xfs_log_recover.c
193
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
196
xfs_debug(mp, "%s: SB : uuid = %pU, fmt = %d",
fs/xfs/xfs_log_recover.c
197
__func__, &mp->m_sb.sb_uuid, XLOG_FMT);
fs/xfs/xfs_log_recover.c
198
xfs_debug(mp, " log : uuid = %pU, fmt = %d",
fs/xfs/xfs_log_recover.c
202
#define xlog_header_check_dump(mp, head)
fs/xfs/xfs_log_recover.c
210
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
220
if (XFS_IS_CORRUPT(mp, head->h_fmt != cpu_to_be32(XLOG_FMT))) {
fs/xfs/xfs_log_recover.c
221
xfs_warn(mp,
fs/xfs/xfs_log_recover.c
223
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
226
if (XFS_IS_CORRUPT(mp, !uuid_equal(&mp->m_sb.sb_uuid,
fs/xfs/xfs_log_recover.c
228
xfs_warn(mp,
fs/xfs/xfs_log_recover.c
230
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
241
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
2513
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
252
xfs_warn(mp, "null uuid in log - IRIX style log");
fs/xfs/xfs_log_recover.c
253
} else if (XFS_IS_CORRUPT(mp, !uuid_equal(&mp->m_sb.sb_uuid,
fs/xfs/xfs_log_recover.c
2534
error = xfs_trans_alloc(mp, &resv, dfc->dfc_blkres,
fs/xfs/xfs_log_recover.c
2537
xlog_force_shutdown(mp->m_log, SHUTDOWN_LOG_IO_ERROR);
fs/xfs/xfs_log_recover.c
255
xfs_warn(mp, "log has mismatched uuid - can't recover");
fs/xfs/xfs_log_recover.c
256
xlog_header_check_dump(mp, head);
fs/xfs/xfs_log_recover.c
2560
struct xfs_mount *mp,
fs/xfs/xfs_log_recover.c
2568
xfs_defer_ops_capture_abort(mp, dfc);
fs/xfs/xfs_log_recover.c
2685
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_log_recover.c
2692
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_clearagi, 0, 0, 0, &tp);
fs/xfs/xfs_log_recover.c
2715
xfs_warn(mp, "%s: failed to clear agi %d. Continuing.", __func__,
fs/xfs/xfs_log_recover.c
2726
struct xfs_mount *mp = pag_mount(pag);
fs/xfs/xfs_log_recover.c
2734
error = xfs_iget(mp, NULL, xfs_agino_to_ino(pag, agino), 0, 0,
fs/xfs/xfs_log_recover.c
2757
error = xfs_inodegc_flush(mp);
fs/xfs/xfs_log_recover.c
2772
error2 = xfs_inodegc_flush(mp);
fs/xfs/xfs_log_recover.c
2955
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_log_recover.c
2959
if (XFS_IS_CORRUPT(mp,
fs/xfs/xfs_log_recover.c
2966
if (xfs_has_logv2(mp)) {
fs/xfs/xfs_log_recover.c
2967
if (XFS_IS_CORRUPT(mp, h_version != XLOG_VERSION_2))
fs/xfs/xfs_log_recover.c
2970
if (XFS_IS_CORRUPT(mp, h_version != XLOG_VERSION_1))
fs/xfs/xfs_log_recover.c
2979
if (XFS_IS_CORRUPT(mp, hlen <= 0 || hlen > bufsize))
fs/xfs/xfs_log_recover.c
2982
if (XFS_IS_CORRUPT(mp, blkno > log->l_logBBsize || blkno > INT_MAX))
fs/xfs/xfs_log_recover.c
3357
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_log_recover.c
3358
struct xfs_buf *bp = mp->m_sb_bp;
fs/xfs/xfs_log_recover.c
3359
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_log_recover.c
3405
mp->m_features |= xfs_sb_version_to_features(sbp);
fs/xfs/xfs_log_recover.c
3406
xfs_reinit_percpu_counters(mp);
fs/xfs/xfs_message.c
120
struct xfs_mount *mp = bp->b_mount;
fs/xfs/xfs_message.c
131
__xfs_printk(KERN_ALERT, mp, &vaf);
fs/xfs/xfs_message.c
137
struct xfs_mount *mp,
fs/xfs/xfs_message.c
160
if (xfs_should_warn(mp, features[feat].opstate))
fs/xfs/xfs_message.c
161
xfs_warn(mp,
fs/xfs/xfs_message.c
20
const struct xfs_mount *mp,
fs/xfs/xfs_message.c
23
if (mp && mp->m_super) {
fs/xfs/xfs_message.c
24
printk("%sXFS (%s): %pV\n", level, mp->m_super->s_id, vaf);
fs/xfs/xfs_message.c
33
const struct xfs_mount *mp,
fs/xfs/xfs_message.c
44
__xfs_printk(kern_level, mp, &vaf);
fs/xfs/xfs_message.c
56
const struct xfs_mount *mp,
fs/xfs/xfs_message.c
65
xfs_alert(mp, "Transforming an alert into a BUG.");
fs/xfs/xfs_message.c
74
__xfs_printk(KERN_ALERT, mp, &vaf);
fs/xfs/xfs_message.c
82
struct xfs_mount *mp,
fs/xfs/xfs_message.c
87
xfs_warn(mp, "Assertion failed: %s, file: %s, line: %d",
fs/xfs/xfs_message.c
94
struct xfs_mount *mp,
fs/xfs/xfs_message.c
99
xfs_emerg(mp, "Assertion failed: %s, file: %s, line: %d",
fs/xfs/xfs_message.h
10
void xfs_printk_level(const char *kern_level, const struct xfs_mount *mp,
fs/xfs/xfs_message.h
100
void xfs_warn_experimental(struct xfs_mount *mp, enum xfs_experimental_feat f);
fs/xfs/xfs_message.h
13
#define xfs_printk_index_wrap(kern_level, mp, fmt, ...) \
fs/xfs/xfs_message.h
16
xfs_printk_level(kern_level, mp, fmt, ##__VA_ARGS__); \
fs/xfs/xfs_message.h
18
#define xfs_emerg(mp, fmt, ...) \
fs/xfs/xfs_message.h
19
xfs_printk_index_wrap(KERN_EMERG, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
20
#define xfs_alert(mp, fmt, ...) \
fs/xfs/xfs_message.h
21
xfs_printk_index_wrap(KERN_ALERT, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
22
#define xfs_crit(mp, fmt, ...) \
fs/xfs/xfs_message.h
23
xfs_printk_index_wrap(KERN_CRIT, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
24
#define xfs_err(mp, fmt, ...) \
fs/xfs/xfs_message.h
25
xfs_printk_index_wrap(KERN_ERR, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
26
#define xfs_warn(mp, fmt, ...) \
fs/xfs/xfs_message.h
27
xfs_printk_index_wrap(KERN_WARNING, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
28
#define xfs_notice(mp, fmt, ...) \
fs/xfs/xfs_message.h
29
xfs_printk_index_wrap(KERN_NOTICE, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
30
#define xfs_info(mp, fmt, ...) \
fs/xfs/xfs_message.h
31
xfs_printk_index_wrap(KERN_INFO, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
33
#define xfs_debug(mp, fmt, ...) \
fs/xfs/xfs_message.h
34
xfs_printk_index_wrap(KERN_DEBUG, mp, fmt, ##__VA_ARGS__)
fs/xfs/xfs_message.h
36
#define xfs_debug(mp, fmt, ...) do {} while (0)
fs/xfs/xfs_message.h
39
#define xfs_alert_tag(mp, tag, fmt, ...) \
fs/xfs/xfs_message.h
42
_xfs_alert_tag(mp, tag, fmt, ##__VA_ARGS__); \
fs/xfs/xfs_message.h
46
void _xfs_alert_tag(const struct xfs_mount *mp, uint32_t tag,
fs/xfs/xfs_message.h
85
void assfail(struct xfs_mount *mp, char *expr, char *f, int l);
fs/xfs/xfs_message.h
86
void asswarn(struct xfs_mount *mp, char *expr, char *f, int l);
fs/xfs/xfs_mount.c
1001
xfs_warn(mp, "Failed per-ag init: %d", error);
fs/xfs/xfs_mount.c
1005
error = xfs_initialize_rtgroups(mp, 0, sbp->sb_rgcount,
fs/xfs/xfs_mount.c
1006
mp->m_sb.sb_rextents);
fs/xfs/xfs_mount.c
1008
xfs_warn(mp, "Failed rtgroup init: %d", error);
fs/xfs/xfs_mount.c
1012
if (XFS_IS_CORRUPT(mp, !sbp->sb_logblocks)) {
fs/xfs/xfs_mount.c
1013
xfs_warn(mp, "no log defined");
fs/xfs/xfs_mount.c
1018
error = xfs_inodegc_register_shrinker(mp);
fs/xfs/xfs_mount.c
1026
if (xfs_is_resuming_quotaon(mp))
fs/xfs/xfs_mount.c
1027
xfs_qm_resume_quotaon(mp);
fs/xfs/xfs_mount.c
1034
error = xfs_log_mount(mp, mp->m_logdev_targp,
fs/xfs/xfs_mount.c
1035
XFS_FSB_TO_DADDR(mp, sbp->sb_logstart),
fs/xfs/xfs_mount.c
1036
XFS_FSB_TO_BB(mp, sbp->sb_logblocks));
fs/xfs/xfs_mount.c
1038
xfs_warn(mp, "log mount failed");
fs/xfs/xfs_mount.c
1047
if (xfs_clear_resuming_quotaon(mp) && xlog_recovery_needed(mp->m_log))
fs/xfs/xfs_mount.c
1048
xfs_qm_resume_quotaon(mp);
fs/xfs/xfs_mount.c
105
xfs_warn(mp, "Filesystem has duplicate UUID %pU - can't mount", uuid);
fs/xfs/xfs_mount.c
1054
if (xfs_sb_version_haslogxattrs(&mp->m_sb))
fs/xfs/xfs_mount.c
1055
xfs_set_using_logged_xattrs(mp);
fs/xfs/xfs_mount.c
1057
xfs_clear_using_logged_xattrs(mp);
fs/xfs/xfs_mount.c
1060
xfs_inodegc_start(mp);
fs/xfs/xfs_mount.c
1061
xfs_blockgc_start(mp);
fs/xfs/xfs_mount.c
1063
if (xfs_has_metadir(mp)) {
fs/xfs/xfs_mount.c
1064
error = xfs_mount_setup_metadir(mp);
fs/xfs/xfs_mount.c
1073
error = xfs_iget(mp, NULL, sbp->sb_rootino, XFS_IGET_UNTRUSTED,
fs/xfs/xfs_mount.c
1076
xfs_warn(mp,
fs/xfs/xfs_mount.c
1084
if (XFS_IS_CORRUPT(mp, !S_ISDIR(VFS_I(rip)->i_mode))) {
fs/xfs/xfs_mount.c
1085
xfs_warn(mp, "corrupted root inode %llu: not a directory",
fs/xfs/xfs_mount.c
1091
mp->m_rootip = rip; /* save it */
fs/xfs/xfs_mount.c
1098
error = xfs_rtmount_inodes(mp);
fs/xfs/xfs_mount.c
1103
xfs_warn(mp, "failed to read RT inodes");
fs/xfs/xfs_mount.c
1108
error = xfs_check_summary_counts(mp);
fs/xfs/xfs_mount.c
111
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
1117
if (mp->m_update_sb && !xfs_is_readonly(mp)) {
fs/xfs/xfs_mount.c
1118
error = xfs_sync_sb(mp, false);
fs/xfs/xfs_mount.c
1120
xfs_warn(mp, "failed to write sb changes");
fs/xfs/xfs_mount.c
1128
if (XFS_IS_QUOTA_ON(mp)) {
fs/xfs/xfs_mount.c
1129
error = xfs_qm_newmount(mp, "amount, "aflags);
fs/xfs/xfs_mount.c
113
uuid_t *uuid = &mp->m_sb.sb_uuid;
fs/xfs/xfs_mount.c
1138
if (mp->m_sb.sb_qflags & XFS_ALL_QUOTA_ACCT) {
fs/xfs/xfs_mount.c
1139
xfs_notice(mp, "resetting quota flags");
fs/xfs/xfs_mount.c
1140
error = xfs_mount_reset_sbqflags(mp);
fs/xfs/xfs_mount.c
1154
error = xfs_fs_reserve_ag_blocks(mp);
fs/xfs/xfs_mount.c
1156
xfs_warn(mp,
fs/xfs/xfs_mount.c
1158
error = xfs_log_mount_finish(mp);
fs/xfs/xfs_mount.c
1159
xfs_fs_unreserve_ag_blocks(mp);
fs/xfs/xfs_mount.c
116
if (xfs_has_nouuid(mp))
fs/xfs/xfs_mount.c
1161
xfs_warn(mp, "log mount finish failed");
fs/xfs/xfs_mount.c
1174
if (xfs_is_readonly(mp) && !xfs_has_norecovery(mp))
fs/xfs/xfs_mount.c
1175
xfs_log_clean(mp);
fs/xfs/xfs_mount.c
1177
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_mount.c
1178
error = xfs_mount_zones(mp);
fs/xfs/xfs_mount.c
1187
ASSERT(mp->m_qflags == 0);
fs/xfs/xfs_mount.c
1188
mp->m_qflags = quotaflags;
fs/xfs/xfs_mount.c
1190
xfs_qm_mount_quotas(mp);
fs/xfs/xfs_mount.c
1204
if (!xfs_is_readonly(mp)) {
fs/xfs/xfs_mount.c
1206
error = xfs_reserve_blocks(mp, i,
fs/xfs/xfs_mount.c
1207
xfs_default_resblks(mp, i));
fs/xfs/xfs_mount.c
1209
xfs_warn(mp,
fs/xfs/xfs_mount.c
1215
error = xfs_fs_reserve_ag_blocks(mp);
fs/xfs/xfs_mount.c
1219
xfs_zone_gc_start(mp);
fs/xfs/xfs_mount.c
1227
error = xfs_set_max_atomic_write_opt(mp, mp->m_awu_max_bytes);
fs/xfs/xfs_mount.c
1234
xfs_fs_unreserve_ag_blocks(mp);
fs/xfs/xfs_mount.c
1235
xfs_qm_unmount_quotas(mp);
fs/xfs/xfs_mount.c
1236
if (xfs_has_zoned(mp))
fs/xfs/xfs_mount.c
1237
xfs_unmount_zones(mp);
fs/xfs/xfs_mount.c
1239
xfs_rtunmount_inodes(mp);
fs/xfs/xfs_mount.c
1243
xfs_qm_unmount(mp);
fs/xfs/xfs_mount.c
1245
if (mp->m_metadirip)
fs/xfs/xfs_mount.c
1246
xfs_irele(mp->m_metadirip);
fs/xfs/xfs_mount.c
1254
xfs_inodegc_flush(mp);
fs/xfs/xfs_mount.c
1267
xfs_unmount_flush_inodes(mp);
fs/xfs/xfs_mount.c
1268
xfs_log_mount_cancel(mp);
fs/xfs/xfs_mount.c
1270
shrinker_free(mp->m_inodegc_shrinker);
fs/xfs/xfs_mount.c
1272
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_mount.c
1273
xfs_buftarg_drain(mp->m_logdev_targp);
fs/xfs/xfs_mount.c
1274
xfs_buftarg_drain(mp->m_ddev_targp);
fs/xfs/xfs_mount.c
1276
xfs_free_rtgroups(mp, 0, mp->m_sb.sb_rgcount);
fs/xfs/xfs_mount.c
1278
xfs_free_perag_range(mp, 0, mp->m_sb.sb_agcount);
fs/xfs/xfs_mount.c
1280
xfs_da_unmount(mp);
fs/xfs/xfs_mount.c
1282
xfs_uuid_unmount(mp);
fs/xfs/xfs_mount.c
1284
xfs_errortag_del(mp);
fs/xfs/xfs_mount.c
1286
xfs_mount_sysfs_del(mp);
fs/xfs/xfs_mount.c
1288
xchk_stats_unregister(mp->m_scrub_stats);
fs/xfs/xfs_mount.c
1299
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
1311
xfs_inodegc_flush(mp);
fs/xfs/xfs_mount.c
1313
xfs_blockgc_stop(mp);
fs/xfs/xfs_mount.c
1314
if (!test_bit(XFS_OPSTATE_READONLY, &mp->m_opstate))
fs/xfs/xfs_mount.c
1315
xfs_zone_gc_stop(mp);
fs/xfs/xfs_mount.c
1316
xfs_fs_unreserve_ag_blocks(mp);
fs/xfs/xfs_mount.c
1317
xfs_qm_unmount_quotas(mp);
fs/xfs/xfs_mount.c
1318
if (xfs_has_zoned(mp))
fs/xfs/xfs_mount.c
1319
xfs_unmount_zones(mp);
fs/xfs/xfs_mount.c
1320
xfs_rtunmount_inodes(mp);
fs/xfs/xfs_mount.c
1321
xfs_irele(mp->m_rootip);
fs/xfs/xfs_mount.c
1322
if (mp->m_metadirip)
fs/xfs/xfs_mount.c
1323
xfs_irele(mp->m_metadirip);
fs/xfs/xfs_mount.c
1325
xfs_unmount_flush_inodes(mp);
fs/xfs/xfs_mount.c
1327
xfs_qm_unmount(mp);
fs/xfs/xfs_mount.c
1343
error = xfs_reserve_blocks(mp, XC_FREE_BLOCKS, 0);
fs/xfs/xfs_mount.c
1345
xfs_warn(mp, "Unable to free reserved block pool. "
fs/xfs/xfs_mount.c
1347
xfs_unmount_check(mp);
fs/xfs/xfs_mount.c
1353
xfs_set_done_with_log_incompat(mp);
fs/xfs/xfs_mount.c
1354
xfs_log_unmount(mp);
fs/xfs/xfs_mount.c
1355
xfs_da_unmount(mp);
fs/xfs/xfs_mount.c
1356
xfs_uuid_unmount(mp);
fs/xfs/xfs_mount.c
1359
xfs_errortag_clearall(mp);
fs/xfs/xfs_mount.c
1361
shrinker_free(mp->m_inodegc_shrinker);
fs/xfs/xfs_mount.c
1362
xfs_free_rtgroups(mp, 0, mp->m_sb.sb_rgcount);
fs/xfs/xfs_mount.c
1363
xfs_free_perag_range(mp, 0, mp->m_sb.sb_agcount);
fs/xfs/xfs_mount.c
1364
xfs_errortag_del(mp);
fs/xfs/xfs_mount.c
1365
xchk_stats_unregister(mp->m_scrub_stats);
fs/xfs/xfs_mount.c
1366
xfs_mount_sysfs_del(mp);
fs/xfs/xfs_mount.c
1377
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1381
if ((mp->m_super->s_writers.frozen >= level) ||
fs/xfs/xfs_mount.c
1382
xfs_is_shutdown(mp) || xfs_is_readonly(mp))
fs/xfs/xfs_mount.c
1398
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1403
return mp->m_alloc_set_aside + atomic64_read(&mp->m_allocbt_blks);
fs/xfs/xfs_mount.c
1408
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1412
struct xfs_freecounter *counter = &mp->m_free[ctr];
fs/xfs/xfs_mount.c
1424
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
1433
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
1440
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1445
struct xfs_freecounter *counter = &mp->m_free[ctr];
fs/xfs/xfs_mount.c
1477
xfs_freecounter_unavailable(mp, ctr),
fs/xfs/xfs_mount.c
1483
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
1489
xfs_warn_once(mp,
fs/xfs/xfs_mount.c
1494
trace_xfs_freecounter_reserved(mp, ctr, delta, _RET_IP_);
fs/xfs/xfs_mount.c
1495
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
1502
trace_xfs_freecounter_enospc(mp, ctr, delta, _RET_IP_);
fs/xfs/xfs_mount.c
1503
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
1512
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
1514
struct xfs_buf *bp = mp->m_sb_bp;
fs/xfs/xfs_mount.c
1517
mp->m_sb_bp = NULL;
fs/xfs/xfs_mount.c
1527
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1530
if (xfs_readonly_buftarg(mp->m_ddev_targp) ||
fs/xfs/xfs_mount.c
1531
xfs_readonly_buftarg(mp->m_logdev_targp) ||
fs/xfs/xfs_mount.c
1532
(mp->m_rtdev_targp && xfs_readonly_buftarg(mp->m_rtdev_targp))) {
fs/xfs/xfs_mount.c
1533
xfs_notice(mp, "%s required on read-only device.", message);
fs/xfs/xfs_mount.c
1534
xfs_notice(mp, "write access unavailable, cannot proceed.");
fs/xfs/xfs_mount.c
1543
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
1545
if (!xfs_has_lazysbcount(mp))
fs/xfs/xfs_mount.c
1548
xfs_fs_mark_sick(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/xfs_mount.c
1557
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1572
error = xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_mount.c
1575
xfs_ail_push_all(mp->m_ail);
fs/xfs/xfs_mount.c
1581
xfs_buf_lock(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1582
xfs_buf_hold(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1584
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_mount.c
1589
if (xfs_sb_has_incompat_log_feature(&mp->m_sb, feature))
fs/xfs/xfs_mount.c
1597
dsb = mp->m_sb_bp->b_addr;
fs/xfs/xfs_mount.c
1598
xfs_sb_to_disk(dsb, &mp->m_sb);
fs/xfs/xfs_mount.c
1600
error = xfs_bwrite(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1608
xfs_sb_add_incompat_log_features(&mp->m_sb, feature);
fs/xfs/xfs_mount.c
1609
xfs_buf_relse(mp->m_sb_bp);
fs/xfs/xfs_mount.c
161
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
1612
return xfs_sync_sb(mp, false);
fs/xfs/xfs_mount.c
1614
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_mount.c
1616
xfs_buf_relse(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1632
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
1636
if (!xfs_has_crc(mp) ||
fs/xfs/xfs_mount.c
1637
!xfs_sb_has_incompat_log_feature(&mp->m_sb,
fs/xfs/xfs_mount.c
1639
xfs_is_shutdown(mp) ||
fs/xfs/xfs_mount.c
1640
!xfs_is_done_with_log_incompat(mp))
fs/xfs/xfs_mount.c
1648
xfs_buf_lock(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1649
xfs_buf_hold(mp->m_sb_bp);
fs/xfs/xfs_mount.c
1651
if (xfs_sb_has_incompat_log_feature(&mp->m_sb,
fs/xfs/xfs_mount.c
1653
xfs_sb_remove_incompat_log_features(&mp->m_sb);
fs/xfs/xfs_mount.c
1657
xfs_buf_relse(mp->m_sb_bp);
fs/xfs/xfs_mount.c
166
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_mount.c
1679
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_mount.c
1682
percpu_counter_add_batch(&mp->m_delalloc_rtextents,
fs/xfs/xfs_mount.c
1683
xfs_blen_to_rtbxlen(mp, data_delta),
fs/xfs/xfs_mount.c
1689
percpu_counter_add_batch(&mp->m_delalloc_blks, data_delta + ind_delta,
fs/xfs/xfs_mount.c
171
ASSERT(mp->m_sb_bp == NULL);
fs/xfs/xfs_mount.c
172
ASSERT(mp->m_ddev_targp != NULL);
fs/xfs/xfs_mount.c
182
sector_size = mp->m_ddev_targp->bt_logical_sectorsize;
fs/xfs/xfs_mount.c
186
error = xfs_buf_read_uncached(mp->m_ddev_targp, XFS_SB_DADDR,
fs/xfs/xfs_mount.c
190
xfs_warn(mp, "SB validate failed with error %d.", error);
fs/xfs/xfs_mount.c
208
xfs_warn(mp, "Invalid superblock magic number");
fs/xfs/xfs_mount.c
218
xfs_warn(mp, "device supports %u byte sectors (not %u)",
fs/xfs/xfs_mount.c
235
mp->m_features |= xfs_sb_version_to_features(sbp);
fs/xfs/xfs_mount.c
236
xfs_reinit_percpu_counters(mp);
fs/xfs/xfs_mount.c
242
if (xfs_sb_version_haslogxattrs(&mp->m_sb))
fs/xfs/xfs_mount.c
243
xfs_set_using_logged_xattrs(mp);
fs/xfs/xfs_mount.c
252
mp->m_sb_bp = bp;
fs/xfs/xfs_mount.c
269
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
273
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_mount.c
276
calc_ino = xfs_ialloc_calc_rootino(mp, new_dalign);
fs/xfs/xfs_mount.c
277
trace_xfs_check_new_dalign(mp, new_dalign, calc_ino);
fs/xfs/xfs_mount.c
284
xfs_warn(mp,
fs/xfs/xfs_mount.c
292
xfs_warn(mp, "Skipping superblock stripe alignment update.");
fs/xfs/xfs_mount.c
305
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
307
if (mp->m_dalign == 0)
fs/xfs/xfs_mount.c
314
if ((BBTOB(mp->m_dalign) & mp->m_blockmask) ||
fs/xfs/xfs_mount.c
315
(BBTOB(mp->m_swidth) & mp->m_blockmask)) {
fs/xfs/xfs_mount.c
316
xfs_warn(mp,
fs/xfs/xfs_mount.c
318
mp->m_sb.sb_blocksize);
fs/xfs/xfs_mount.c
325
mp->m_dalign = XFS_BB_TO_FSBT(mp, mp->m_dalign);
fs/xfs/xfs_mount.c
326
if (mp->m_dalign && (mp->m_sb.sb_agblocks % mp->m_dalign)) {
fs/xfs/xfs_mount.c
327
xfs_warn(mp,
fs/xfs/xfs_mount.c
329
mp->m_sb.sb_agblocks);
fs/xfs/xfs_mount.c
333
if (!mp->m_dalign) {
fs/xfs/xfs_mount.c
334
xfs_warn(mp,
fs/xfs/xfs_mount.c
336
mp->m_dalign, mp->m_sb.sb_blocksize);
fs/xfs/xfs_mount.c
340
mp->m_swidth = XFS_BB_TO_FSBT(mp, mp->m_swidth);
fs/xfs/xfs_mount.c
342
if (!xfs_has_dalign(mp)) {
fs/xfs/xfs_mount.c
343
xfs_warn(mp,
fs/xfs/xfs_mount.c
354
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
356
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_mount.c
358
if (mp->m_dalign) {
fs/xfs/xfs_mount.c
362
if (sbp->sb_unit == mp->m_dalign &&
fs/xfs/xfs_mount.c
363
sbp->sb_width == mp->m_swidth)
fs/xfs/xfs_mount.c
366
error = xfs_check_new_dalign(mp, mp->m_dalign, &update_sb);
fs/xfs/xfs_mount.c
370
sbp->sb_unit = mp->m_dalign;
fs/xfs/xfs_mount.c
371
sbp->sb_width = mp->m_swidth;
fs/xfs/xfs_mount.c
372
mp->m_update_sb = true;
fs/xfs/xfs_mount.c
373
} else if (!xfs_has_noalign(mp) && xfs_has_dalign(mp)) {
fs/xfs/xfs_mount.c
374
mp->m_dalign = sbp->sb_unit;
fs/xfs/xfs_mount.c
375
mp->m_swidth = sbp->sb_width;
fs/xfs/xfs_mount.c
386
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
388
uint64_t dblocks = mp->m_sb.sb_dblocks;
fs/xfs/xfs_mount.c
389
uint64_t rtexts = mp->m_sb.sb_rextents;
fs/xfs/xfs_mount.c
396
mp->m_low_space[i] = dblocks * (i + 1);
fs/xfs/xfs_mount.c
397
mp->m_low_rtexts[i] = rtexts * (i + 1);
fs/xfs/xfs_mount.c
406
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
412
d = (xfs_daddr_t)XFS_FSB_TO_BB(mp, mp->m_sb.sb_dblocks);
fs/xfs/xfs_mount.c
413
if (XFS_BB_TO_FSB(mp, d) != mp->m_sb.sb_dblocks) {
fs/xfs/xfs_mount.c
414
xfs_warn(mp, "filesystem size mismatch detected");
fs/xfs/xfs_mount.c
417
error = xfs_buf_read_uncached(mp->m_ddev_targp,
fs/xfs/xfs_mount.c
418
d - XFS_FSS_TO_BB(mp, 1),
fs/xfs/xfs_mount.c
419
XFS_FSS_TO_BB(mp, 1), &bp, NULL);
fs/xfs/xfs_mount.c
421
xfs_warn(mp, "last sector read failed");
fs/xfs/xfs_mount.c
426
if (mp->m_logdev_targp == mp->m_ddev_targp)
fs/xfs/xfs_mount.c
429
d = (xfs_daddr_t)XFS_FSB_TO_BB(mp, mp->m_sb.sb_logblocks);
fs/xfs/xfs_mount.c
430
if (XFS_BB_TO_FSB(mp, d) != mp->m_sb.sb_logblocks) {
fs/xfs/xfs_mount.c
431
xfs_warn(mp, "log size mismatch detected");
fs/xfs/xfs_mount.c
434
error = xfs_buf_read_uncached(mp->m_logdev_targp,
fs/xfs/xfs_mount.c
435
d - XFS_FSB_TO_BB(mp, 1),
fs/xfs/xfs_mount.c
436
XFS_FSB_TO_BB(mp, 1), &bp, NULL);
fs/xfs/xfs_mount.c
438
xfs_warn(mp, "log device read failed");
fs/xfs/xfs_mount.c
450
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
452
mp->m_qflags = 0;
fs/xfs/xfs_mount.c
455
if (mp->m_sb.sb_qflags == 0)
fs/xfs/xfs_mount.c
457
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
458
mp->m_sb.sb_qflags = 0;
fs/xfs/xfs_mount.c
459
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_mount.c
461
if (!xfs_fs_writable(mp, SB_FREEZE_WRITE))
fs/xfs/xfs_mount.c
464
return xfs_sync_sb(mp, false);
fs/xfs/xfs_mount.c
475
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
489
return min(div_u64(mp->m_sb.sb_dblocks, 20), 8192ULL);
fs/xfs/xfs_mount.c
492
if (IS_ENABLED(CONFIG_XFS_RT) && xfs_has_zoned(mp))
fs/xfs/xfs_mount.c
493
return xfs_zoned_default_resblks(mp, ctr);
fs/xfs/xfs_mount.c
504
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
512
if (mp->m_sb.sb_inprogress) {
fs/xfs/xfs_mount.c
513
xfs_err(mp, "sb_inprogress set after log recovery??");
fs/xfs/xfs_mount.c
529
if (xfs_is_clean(mp) &&
fs/xfs/xfs_mount.c
530
(mp->m_sb.sb_fdblocks > mp->m_sb.sb_dblocks ||
fs/xfs/xfs_mount.c
531
!xfs_verify_icount(mp, mp->m_sb.sb_icount) ||
fs/xfs/xfs_mount.c
532
mp->m_sb.sb_ifree > mp->m_sb.sb_icount))
fs/xfs/xfs_mount.c
533
xfs_fs_mark_sick(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/xfs_mount.c
546
if ((xfs_has_lazysbcount(mp) && !xfs_is_clean(mp)) ||
fs/xfs/xfs_mount.c
547
xfs_fs_has_sickness(mp, XFS_SICK_FS_COUNTERS)) {
fs/xfs/xfs_mount.c
548
error = xfs_initialize_perag_data(mp, mp->m_sb.sb_agcount);
fs/xfs/xfs_mount.c
565
if (xfs_has_realtime(mp) && !xfs_has_zoned(mp) && !xfs_is_clean(mp)) {
fs/xfs/xfs_mount.c
566
error = xfs_rtalloc_reinit_frextents(mp);
fs/xfs/xfs_mount.c
576
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
578
if (xfs_is_shutdown(mp))
fs/xfs/xfs_mount.c
581
if (percpu_counter_sum(&mp->m_ifree) >
fs/xfs/xfs_mount.c
582
percpu_counter_sum(&mp->m_icount)) {
fs/xfs/xfs_mount.c
583
xfs_alert(mp, "ifree/icount mismatch at unmount");
fs/xfs/xfs_mount.c
584
xfs_fs_mark_sick(mp, XFS_SICK_FS_COUNTERS);
fs/xfs/xfs_mount.c
617
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
619
xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_mount.c
620
xfs_extent_busy_wait_all(mp);
fs/xfs/xfs_mount.c
623
xfs_set_unmounting(mp);
fs/xfs/xfs_mount.c
625
xfs_inodegc_stop(mp);
fs/xfs/xfs_mount.c
626
cancel_delayed_work_sync(&mp->m_reclaim_work);
fs/xfs/xfs_mount.c
627
xfs_ail_push_all_sync(mp->m_ail);
fs/xfs/xfs_mount.c
628
xfs_reclaim_inodes(mp);
fs/xfs/xfs_mount.c
629
xfs_health_unmount(mp);
fs/xfs/xfs_mount.c
630
xfs_healthmon_unmount(mp);
fs/xfs/xfs_mount.c
635
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
637
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/xfs_mount.c
639
igeo->attr_fork_offset = xfs_bmap_compute_attr_offset(mp);
fs/xfs/xfs_mount.c
640
ASSERT(igeo->attr_fork_offset < XFS_LITINO(mp));
fs/xfs/xfs_mount.c
642
xfs_ialloc_setup_geometry(mp);
fs/xfs/xfs_mount.c
648
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
653
error = xfs_metafile_iget(mp, mp->m_sb.sb_metadirino, XFS_METAFILE_DIR,
fs/xfs/xfs_mount.c
654
&mp->m_metadirip);
fs/xfs/xfs_mount.c
656
xfs_warn(mp, "Failed to load metadir root directory, error %d",
fs/xfs/xfs_mount.c
66
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
664
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
668
levels = max(mp->m_alloc_maxlevels, M_IGEO(mp)->inobt_maxlevels);
fs/xfs/xfs_mount.c
669
levels = max(levels, mp->m_rmap_maxlevels);
fs/xfs/xfs_mount.c
670
mp->m_agbtree_maxlevels = max(levels, mp->m_refc_maxlevels);
fs/xfs/xfs_mount.c
674
static inline xfs_extlen_t xfs_calc_atomic_write_max(struct xfs_mount *mp)
fs/xfs/xfs_mount.c
676
return rounddown_pow_of_two(XFS_B_TO_FSB(mp, MAX_RW_COUNT));
fs/xfs/xfs_mount.c
68
uuid_t *uuid = &mp->m_sb.sb_uuid;
fs/xfs/xfs_mount.c
692
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
695
struct xfs_groups *g = &mp->m_groups[type];
fs/xfs/xfs_mount.c
696
struct xfs_buftarg *btp = xfs_group_type_buftarg(mp, type);
fs/xfs/xfs_mount.c
708
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
711
struct xfs_groups *g = &mp->m_groups[type];
fs/xfs/xfs_mount.c
713
const xfs_extlen_t max_write = xfs_calc_atomic_write_max(mp);
fs/xfs/xfs_mount.c
714
const xfs_extlen_t max_ioend = xfs_reflink_max_atomic_cow(mp);
fs/xfs/xfs_mount.c
715
const xfs_extlen_t max_gsize = xfs_calc_group_awu_max(mp, type);
fs/xfs/xfs_mount.c
718
trace_xfs_calc_atomic_write_unit_max(mp, type, max_write, max_ioend,
fs/xfs/xfs_mount.c
72
super_set_uuid(mp->m_super, uuid->b, sizeof(*uuid));
fs/xfs/xfs_mount.c
728
struct xfs_mount *mp,
fs/xfs/xfs_mount.c
731
const xfs_filblks_t new_max_fsbs = XFS_B_TO_FSBT(mp, new_max_bytes);
fs/xfs/xfs_mount.c
732
const xfs_extlen_t max_write = xfs_calc_atomic_write_max(mp);
fs/xfs/xfs_mount.c
734
max(mp->m_groups[XG_TYPE_AG].blocks,
fs/xfs/xfs_mount.c
735
mp->m_groups[XG_TYPE_RTG].blocks);
fs/xfs/xfs_mount.c
737
max(xfs_calc_group_awu_max(mp, XG_TYPE_AG),
fs/xfs/xfs_mount.c
738
xfs_calc_group_awu_max(mp, XG_TYPE_RTG));
fs/xfs/xfs_mount.c
74
if (xfs_has_nouuid(mp))
fs/xfs/xfs_mount.c
748
xfs_warn(mp,
fs/xfs/xfs_mount.c
754
if (new_max_bytes & mp->m_blockmask) {
fs/xfs/xfs_mount.c
755
xfs_warn(mp,
fs/xfs/xfs_mount.c
762
xfs_warn(mp,
fs/xfs/xfs_mount.c
765
XFS_FSB_TO_B(mp, max_write) >> 10);
fs/xfs/xfs_mount.c
770
xfs_warn(mp,
fs/xfs/xfs_mount.c
773
XFS_FSB_TO_B(mp, max_group) >> 10);
fs/xfs/xfs_mount.c
778
xfs_warn(mp,
fs/xfs/xfs_mount.c
78
xfs_warn(mp, "Filesystem has null UUID - can't mount");
fs/xfs/xfs_mount.c
781
XFS_FSB_TO_B(mp, max_group_write) >> 10);
fs/xfs/xfs_mount.c
785
if (xfs_has_reflink(mp))
fs/xfs/xfs_mount.c
789
if (mp->m_ddev_targp->bt_awu_max ||
fs/xfs/xfs_mount.c
790
(mp->m_rtdev_targp && mp->m_rtdev_targp->bt_awu_max)) {
fs/xfs/xfs_mount.c
792
xfs_warn(mp,
fs/xfs/xfs_mount.c
798
xfs_warn(mp,
fs/xfs/xfs_mount.c
805
error = xfs_calc_atomic_write_reservation(mp, new_max_fsbs);
fs/xfs/xfs_mount.c
807
xfs_warn(mp,
fs/xfs/xfs_mount.c
813
xfs_calc_atomic_write_unit_max(mp, XG_TYPE_AG);
fs/xfs/xfs_mount.c
814
xfs_calc_atomic_write_unit_max(mp, XG_TYPE_RTG);
fs/xfs/xfs_mount.c
815
mp->m_awu_max_bytes = new_max_bytes;
fs/xfs/xfs_mount.c
822
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
824
mp->m_rtbtree_maxlevels = max(mp->m_rtrmap_maxlevels,
fs/xfs/xfs_mount.c
825
mp->m_rtrefc_maxlevels);
fs/xfs/xfs_mount.c
840
struct xfs_mount *mp)
fs/xfs/xfs_mount.c
842
struct xfs_sb *sbp = &(mp->m_sb);
fs/xfs/xfs_mount.c
844
struct xfs_ino_geometry *igeo = M_IGEO(mp);
fs/xfs/xfs_mount.c
850
xfs_sb_mount_common(mp, sbp);
fs/xfs/xfs_mount.c
869
xfs_warn(mp, "correcting sb_features alignment problem");
fs/xfs/xfs_mount.c
871
mp->m_update_sb = true;
fs/xfs/xfs_mount.c
876
if (!(mp->m_sb.sb_versionnum & XFS_SB_VERSION_NLINKBIT)) {
fs/xfs/xfs_mount.c
877
mp->m_sb.sb_versionnum |= XFS_SB_VERSION_NLINKBIT;
fs/xfs/xfs_mount.c
878
mp->m_features |= XFS_FEAT_NLINK;
fs/xfs/xfs_mount.c
879
mp->m_update_sb = true;
fs/xfs/xfs_mount.c
888
error = xfs_validate_new_dalign(mp);
fs/xfs/xfs_mount.c
892
xfs_alloc_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
893
xfs_bmap_compute_maxlevels(mp, XFS_DATA_FORK);
fs/xfs/xfs_mount.c
894
xfs_bmap_compute_maxlevels(mp, XFS_ATTR_FORK);
fs/xfs/xfs_mount.c
895
xfs_mount_setup_inode_geom(mp);
fs/xfs/xfs_mount.c
896
xfs_rmapbt_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
897
xfs_rtrmapbt_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
898
xfs_refcountbt_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
899
xfs_rtrefcountbt_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
901
xfs_agbtree_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
902
xfs_rtbtree_compute_maxlevels(mp);
fs/xfs/xfs_mount.c
911
error = xfs_update_alignment(mp);
fs/xfs/xfs_mount.c
916
mp->m_fail_unmount = true;
fs/xfs/xfs_mount.c
918
error = xfs_mount_sysfs_init(mp);
fs/xfs/xfs_mount.c
922
xchk_stats_register(mp->m_scrub_stats, mp->m_debugfs);
fs/xfs/xfs_mount.c
924
error = xfs_errortag_init(mp);
fs/xfs/xfs_mount.c
928
error = xfs_uuid_mount(mp);
fs/xfs/xfs_mount.c
936
mp->m_allocsize_log =
fs/xfs/xfs_mount.c
937
max_t(uint32_t, sbp->sb_blocklog, mp->m_allocsize_log);
fs/xfs/xfs_mount.c
938
mp->m_allocsize_blocks = 1U << (mp->m_allocsize_log - sbp->sb_blocklog);
fs/xfs/xfs_mount.c
941
xfs_set_low_space_thresholds(mp);
fs/xfs/xfs_mount.c
948
if (xfs_has_sparseinodes(mp) &&
fs/xfs/xfs_mount.c
949
mp->m_sb.sb_spino_align !=
fs/xfs/xfs_mount.c
950
XFS_B_TO_FSBT(mp, igeo->inode_cluster_size_raw)) {
fs/xfs/xfs_mount.c
951
xfs_warn(mp,
fs/xfs/xfs_mount.c
953
mp->m_sb.sb_spino_align,
fs/xfs/xfs_mount.c
954
XFS_B_TO_FSBT(mp, igeo->inode_cluster_size_raw));
fs/xfs/xfs_mount.c
962
error = xfs_check_sizes(mp);
fs/xfs/xfs_mount.c
969
error = xfs_rtmount_init(mp);
fs/xfs/xfs_mount.c
971
xfs_warn(mp, "RT mount failed");
fs/xfs/xfs_mount.c
979
mp->m_fixedfsid[0] =
fs/xfs/xfs_mount.c
982
mp->m_fixedfsid[1] = get_unaligned_be32(&sbp->sb_uuid.b[0]);
fs/xfs/xfs_mount.c
984
error = xfs_da_mount(mp);
fs/xfs/xfs_mount.c
986
xfs_warn(mp, "Failed dir/attr init: %d", error);
fs/xfs/xfs_mount.c
993
xfs_trans_init(mp);
fs/xfs/xfs_mount.c
998
error = xfs_initialize_perag(mp, 0, sbp->sb_agcount,
fs/xfs/xfs_mount.c
999
mp->m_sb.sb_dblocks, &mp->m_maxagi);
fs/xfs/xfs_mount.h
351
#define M_IGEO(mp) (&(mp)->m_ino_geo)
fs/xfs/xfs_mount.h
408
static inline bool xfs_has_ ## name (const struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
410
return mp->m_features & XFS_FEAT_ ## NAME; \
fs/xfs/xfs_mount.h
416
static inline void xfs_add_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
418
mp->m_features |= XFS_FEAT_ ## NAME; \
fs/xfs/xfs_mount.h
419
xfs_sb_version_add ## name(&mp->m_sb); \
fs/xfs/xfs_mount.h
446
static inline bool xfs_has_rtgroups(const struct xfs_mount *mp)
fs/xfs/xfs_mount.h
449
return xfs_has_metadir(mp);
fs/xfs/xfs_mount.h
452
static inline bool xfs_has_rtsb(const struct xfs_mount *mp)
fs/xfs/xfs_mount.h
455
return xfs_has_rtgroups(mp) &&
fs/xfs/xfs_mount.h
456
xfs_has_realtime(mp) &&
fs/xfs/xfs_mount.h
457
!xfs_has_zoned(mp);
fs/xfs/xfs_mount.h
460
static inline bool xfs_has_rtrmapbt(const struct xfs_mount *mp)
fs/xfs/xfs_mount.h
462
return xfs_has_rtgroups(mp) && xfs_has_realtime(mp) &&
fs/xfs/xfs_mount.h
463
xfs_has_rmapbt(mp);
fs/xfs/xfs_mount.h
466
static inline bool xfs_has_rtreflink(const struct xfs_mount *mp)
fs/xfs/xfs_mount.h
468
return xfs_has_metadir(mp) && xfs_has_realtime(mp) &&
fs/xfs/xfs_mount.h
469
xfs_has_reflink(mp);
fs/xfs/xfs_mount.h
472
static inline bool xfs_has_nonzoned(const struct xfs_mount *mp)
fs/xfs/xfs_mount.h
474
return !xfs_has_zoned(mp);
fs/xfs/xfs_mount.h
477
static inline bool xfs_can_sw_atomic_write(struct xfs_mount *mp)
fs/xfs/xfs_mount.h
479
return xfs_has_reflink(mp);
fs/xfs/xfs_mount.h
487
static inline bool xfs_has_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
490
(mp->m_features & XFS_FEAT_ ## NAME); \
fs/xfs/xfs_mount.h
495
static inline void xfs_add_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
498
mp->m_features |= XFS_FEAT_ ## NAME; \
fs/xfs/xfs_mount.h
499
xfs_sb_version_add ## name(&mp->m_sb); \
fs/xfs/xfs_mount.h
512
static inline void xfs_add_attr2(struct xfs_mount *mp)
fs/xfs/xfs_mount.h
515
xfs_sb_version_addattr2(&mp->m_sb);
fs/xfs/xfs_mount.h
586
static inline bool xfs_is_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
588
return test_bit(XFS_OPSTATE_ ## NAME, &mp->m_opstate); \
fs/xfs/xfs_mount.h
590
static inline bool xfs_clear_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
592
return test_and_clear_bit(XFS_OPSTATE_ ## NAME, &mp->m_opstate); \
fs/xfs/xfs_mount.h
594
static inline bool xfs_set_ ## name (struct xfs_mount *mp) \
fs/xfs/xfs_mount.h
596
return test_and_set_bit(XFS_OPSTATE_ ## NAME, &mp->m_opstate); \
fs/xfs/xfs_mount.h
610
static inline bool xfs_is_quotacheck_running(struct xfs_mount *mp)
fs/xfs/xfs_mount.h
614
static inline bool xfs_is_resuming_quotaon(struct xfs_mount *mp)
fs/xfs/xfs_mount.h
621
static inline bool xfs_clear_resuming_quotaon(struct xfs_mount *mp)
fs/xfs/xfs_mount.h
631
xfs_should_warn(struct xfs_mount *mp, long nr)
fs/xfs/xfs_mount.h
633
return !test_and_set_bit(nr, &mp->m_opstate);
fs/xfs/xfs_mount.h
64
struct xfs_mount *mp;
fs/xfs/xfs_mount.h
657
void xfs_do_force_shutdown(struct xfs_mount *mp, uint32_t flags, char *fname,
fs/xfs/xfs_mount.h
682
xfs_daddr_to_agno(struct xfs_mount *mp, xfs_daddr_t d)
fs/xfs/xfs_mount.h
684
xfs_rfsblock_t ld = XFS_BB_TO_FSBT(mp, d);
fs/xfs/xfs_mount.h
685
do_div(ld, mp->m_sb.sb_agblocks);
fs/xfs/xfs_mount.h
690
xfs_daddr_to_agbno(struct xfs_mount *mp, xfs_daddr_t d)
fs/xfs/xfs_mount.h
692
xfs_rfsblock_t ld = XFS_BB_TO_FSBT(mp, d);
fs/xfs/xfs_mount.h
693
return (xfs_agblock_t) do_div(ld, mp->m_sb.sb_agblocks);
fs/xfs/xfs_mount.h
697
uint64_t xfs_default_resblks(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
699
extern int xfs_mountfs(xfs_mount_t *mp);
fs/xfs/xfs_mount.h
711
uint64_t xfs_freecounter_unavailable(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
717
static inline s64 xfs_sum_freecounter(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
720
return percpu_counter_sum_positive(&mp->m_free[ctr].count);
fs/xfs/xfs_mount.h
727
static inline s64 xfs_sum_freecounter_raw(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
730
return percpu_counter_sum(&mp->m_free[ctr].count);
fs/xfs/xfs_mount.h
737
static inline s64 xfs_estimate_freecounter(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
740
return percpu_counter_read_positive(&mp->m_free[ctr].count);
fs/xfs/xfs_mount.h
743
static inline int xfs_compare_freecounter(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
746
return __percpu_counter_compare(&mp->m_free[ctr].count, rhs, batch);
fs/xfs/xfs_mount.h
749
static inline void xfs_set_freecounter(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
752
percpu_counter_set(&mp->m_free[ctr].count, val);
fs/xfs/xfs_mount.h
755
int xfs_dec_freecounter(struct xfs_mount *mp, enum xfs_free_counter ctr,
fs/xfs/xfs_mount.h
757
void xfs_add_freecounter(struct xfs_mount *mp, enum xfs_free_counter ctr,
fs/xfs/xfs_mount.h
760
static inline int xfs_dec_fdblocks(struct xfs_mount *mp, uint64_t delta,
fs/xfs/xfs_mount.h
763
return xfs_dec_freecounter(mp, XC_FREE_BLOCKS, delta, reserved);
fs/xfs/xfs_mount.h
766
static inline void xfs_add_fdblocks(struct xfs_mount *mp, uint64_t delta)
fs/xfs/xfs_mount.h
768
xfs_add_freecounter(mp, XC_FREE_BLOCKS, delta);
fs/xfs/xfs_mount.h
771
static inline int xfs_dec_frextents(struct xfs_mount *mp, uint64_t delta)
fs/xfs/xfs_mount.h
773
return xfs_dec_freecounter(mp, XC_FREE_RTEXTENTS, delta, false);
fs/xfs/xfs_mount.h
776
static inline void xfs_add_frextents(struct xfs_mount *mp, uint64_t delta)
fs/xfs/xfs_mount.h
778
xfs_add_freecounter(mp, XC_FREE_RTEXTENTS, delta);
fs/xfs/xfs_mount.h
783
extern bool xfs_fs_writable(struct xfs_mount *mp, int level);
fs/xfs/xfs_mount.h
793
struct xfs_error_cfg * xfs_error_get_cfg(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
795
void xfs_force_summary_recalc(struct xfs_mount *mp);
fs/xfs/xfs_mount.h
796
int xfs_add_incompat_log_feature(struct xfs_mount *mp, uint32_t feature);
fs/xfs/xfs_mount.h
797
bool xfs_clear_incompat_log_features(struct xfs_mount *mp);
fs/xfs/xfs_mount.h
800
static inline void xfs_mod_sb_delalloc(struct xfs_mount *mp, int64_t delta)
fs/xfs/xfs_mount.h
802
percpu_counter_add(&mp->m_delalloc_blks, delta);
fs/xfs/xfs_mount.h
805
int xfs_set_max_atomic_write_opt(struct xfs_mount *mp,
fs/xfs/xfs_mount.h
810
struct xfs_mount *mp,
fs/xfs/xfs_mount.h
815
return mp->m_ddev_targp;
fs/xfs/xfs_mount.h
817
return mp->m_rtdev_targp;
fs/xfs/xfs_notify_failure.c
108
pgoff = xfs_failure_pgoff(mp, rec, notify);
fs/xfs/xfs_notify_failure.c
109
pgcnt = xfs_failure_pgcnt(mp, rec, notify);
fs/xfs/xfs_notify_failure.c
130
struct xfs_mount *mp)
fs/xfs/xfs_notify_failure.c
132
struct super_block *sb = mp->m_super;
fs/xfs/xfs_notify_failure.c
137
xfs_emerg(mp, "already frozen by kernel, err=%d", error);
fs/xfs/xfs_notify_failure.c
144
struct xfs_mount *mp,
fs/xfs/xfs_notify_failure.c
147
struct super_block *sb = mp->m_super;
fs/xfs/xfs_notify_failure.c
153
xfs_emerg(mp, "still frozen after notify failure, err=%d",
fs/xfs/xfs_notify_failure.c
205
struct xfs_mount *mp,
fs/xfs/xfs_notify_failure.c
218
error = xfs_dax_translate_range(mp->m_logdev_targp,
fs/xfs/xfs_notify_failure.c
223
xfs_healthmon_report_media(mp, XFS_DEV_LOG, daddr, bblen);
fs/xfs/xfs_notify_failure.c
234
xfs_err(mp, "ondisk log corrupt, shutting down fs!");
fs/xfs/xfs_notify_failure.c
235
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_ONDISK);
fs/xfs/xfs_notify_failure.c
241
struct xfs_mount *mp,
fs/xfs/xfs_notify_failure.c
258
error = xfs_dax_translate_range(xfs_group_type_buftarg(mp, type),
fs/xfs/xfs_notify_failure.c
263
xfs_healthmon_report_media(mp,
fs/xfs/xfs_notify_failure.c
267
if (!xfs_has_rmapbt(mp)) {
fs/xfs/xfs_notify_failure.c
268
xfs_debug(mp, "notify_failure() needs rmapbt enabled!");
fs/xfs/xfs_notify_failure.c
273
start_bno = xfs_daddr_to_rtb(mp, daddr);
fs/xfs/xfs_notify_failure.c
274
end_bno = xfs_daddr_to_rtb(mp, daddr + bblen - 1);
fs/xfs/xfs_notify_failure.c
276
start_bno = XFS_DADDR_TO_FSB(mp, daddr);
fs/xfs/xfs_notify_failure.c
277
end_bno = XFS_DADDR_TO_FSB(mp, daddr + bblen - 1);
fs/xfs/xfs_notify_failure.c
281
xfs_info(mp, "Device is about to be removed!");
fs/xfs/xfs_notify_failure.c
290
kernel_frozen = xfs_dax_notify_failure_freeze(mp) == 0;
fs/xfs/xfs_notify_failure.c
293
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_notify_failure.c
294
start_gno = xfs_fsb_to_gno(mp, start_bno, type);
fs/xfs/xfs_notify_failure.c
295
end_gno = xfs_fsb_to_gno(mp, end_bno, type);
fs/xfs/xfs_notify_failure.c
296
while ((xg = xfs_group_next_range(mp, xg, start_gno, end_gno, type))) {
fs/xfs/xfs_notify_failure.c
311
cur = xfs_rmapbt_init_cursor(mp, tp, agf_bp, pag);
fs/xfs/xfs_notify_failure.c
325
xfs_fsb_to_gbno(mp, start_bno, type);
fs/xfs/xfs_notify_failure.c
328
xfs_fsb_to_gbno(mp, end_bno, type);
fs/xfs/xfs_notify_failure.c
357
xfs_force_shutdown(mp, SHUTDOWN_FORCE_UMOUNT);
fs/xfs/xfs_notify_failure.c
359
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_ONDISK);
fs/xfs/xfs_notify_failure.c
366
xfs_dax_notify_failure_thaw(mp, kernel_frozen);
fs/xfs/xfs_notify_failure.c
378
struct xfs_mount *mp = dax_holder(dax_dev);
fs/xfs/xfs_notify_failure.c
380
if (!(mp->m_super->s_flags & SB_BORN)) {
fs/xfs/xfs_notify_failure.c
381
xfs_warn(mp, "filesystem is not ready for notify_failure()!");
fs/xfs/xfs_notify_failure.c
385
if (mp->m_logdev_targp != mp->m_ddev_targp &&
fs/xfs/xfs_notify_failure.c
386
mp->m_logdev_targp->bt_daxdev == dax_dev) {
fs/xfs/xfs_notify_failure.c
387
return xfs_dax_notify_logdev_failure(mp, offset, len, mf_flags);
fs/xfs/xfs_notify_failure.c
390
return xfs_dax_notify_dev_failure(mp, offset, len, mf_flags,
fs/xfs/xfs_notify_failure.c
391
(mp->m_rtdev_targp && mp->m_rtdev_targp->bt_daxdev == dax_dev) ?
fs/xfs/xfs_notify_failure.c
41
struct xfs_mount *mp,
fs/xfs/xfs_notify_failure.c
45
loff_t pos = XFS_FSB_TO_B(mp, rec->rm_offset);
fs/xfs/xfs_notify_failure.c
48
pos += XFS_FSB_TO_B(mp,
fs/xfs/xfs_notify_failure.c
55
struct xfs_mount *mp,
fs/xfs/xfs_notify_failure.c
70
return XFS_FSB_TO_B(mp, end_cross - start_cross) >> PAGE_SHIFT;
fs/xfs/xfs_notify_failure.c
79
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/xfs_notify_failure.c
97
error = xfs_iget(mp, cur->bc_tp, rec->rm_owner, XFS_IGET_INCORE,
fs/xfs/xfs_platform.h
257
#define XFS_IS_CORRUPT(mp, expr) \
fs/xfs/xfs_platform.h
258
(unlikely(expr) ? xfs_corruption_error(#expr, XFS_ERRLEVEL_LOW, (mp), \
fs/xfs/xfs_platform.h
274
#define XFS_IS_REALTIME_MOUNT(mp) ((mp)->m_rtdev_targp ? 1 : 0)
fs/xfs/xfs_platform.h
277
#define XFS_IS_REALTIME_MOUNT(mp) (0)
fs/xfs/xfs_pnfs.c
117
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_pnfs.c
127
if (xfs_is_shutdown(mp))
fs/xfs/xfs_pnfs.c
155
limit = mp->m_super->s_maxbytes;
fs/xfs/xfs_pnfs.c
171
end_fsb = XFS_B_TO_FSB(mp, (xfs_ufsize_t)offset + length);
fs/xfs/xfs_pnfs.c
172
offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_pnfs.c
212
*device_generation = mp->m_generation;
fs/xfs/xfs_pnfs.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_pnfs.c
313
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
fs/xfs/xfs_pnfs.c
59
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_pnfs.c
64
memcpy(buf, &mp->m_sb.sb_uuid, sizeof(uuid_t));
fs/xfs/xfs_pwork.c
47
error = pctl->work_fn(pctl->mp, pwork);
fs/xfs/xfs_pwork.c
61
struct xfs_mount *mp,
fs/xfs/xfs_pwork.c
72
trace_xfs_pwork_init(mp, nr_threads, current->pid);
fs/xfs/xfs_pwork.c
81
pctl->mp = mp;
fs/xfs/xfs_pwork.h
12
typedef int (*xfs_pwork_work_fn)(struct xfs_mount *mp, struct xfs_pwork *pwork);
fs/xfs/xfs_pwork.h
19
struct xfs_mount *mp;
fs/xfs/xfs_pwork.h
53
int xfs_pwork_init(struct xfs_mount *mp, struct xfs_pwork_ctl *pctl,
fs/xfs/xfs_qm.c
1000
xfs_alert(mp, "%s failed (error %d)!", __func__, error);
fs/xfs/xfs_qm.c
1012
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
1027
j = (int)XFS_FSB_TO_B(mp, XFS_DQUOT_CLUSTER_SIZE_FSB) /
fs/xfs/xfs_qm.c
1029
ASSERT(mp->m_quotainfo->qi_dqperchunk == j);
fs/xfs/xfs_qm.c
1032
for (j = 0; j < mp->m_quotainfo->qi_dqperchunk; j++) {
fs/xfs/xfs_qm.c
1043
if (xfs_dqblk_verify(mp, &dqb[j], id + j) ||
fs/xfs/xfs_qm.c
1045
xfs_dqblk_repair(mp, &dqb[j], id + j, type);
fs/xfs/xfs_qm.c
1068
if (xfs_has_bigtime(mp))
fs/xfs/xfs_qm.c
1072
if (xfs_has_crc(mp)) {
fs/xfs/xfs_qm.c
1082
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
1104
error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp,
fs/xfs/xfs_qm.c
1105
XFS_FSB_TO_DADDR(mp, bno),
fs/xfs/xfs_qm.c
1106
mp->m_quotainfo->qi_dqchunklen, 0, &bp,
fs/xfs/xfs_qm.c
1117
error = xfs_trans_read_buf(mp, NULL, mp->m_ddev_targp,
fs/xfs/xfs_qm.c
1118
XFS_FSB_TO_DADDR(mp, bno),
fs/xfs/xfs_qm.c
1119
mp->m_quotainfo->qi_dqchunklen, 0, &bp,
fs/xfs/xfs_qm.c
1132
xfs_qm_reset_dqcounts(mp, bp, firstid, type);
fs/xfs/xfs_qm.c
1138
firstid += mp->m_quotainfo->qi_dqperchunk;
fs/xfs/xfs_qm.c
1150
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
1177
maxlblkcnt = XFS_B_TO_FSB(mp, mp->m_super->s_maxbytes);
fs/xfs/xfs_qm.c
1206
mp->m_quotainfo->qi_dqperchunk;
fs/xfs/xfs_qm.c
1215
xfs_buf_readahead(mp->m_ddev_targp,
fs/xfs/xfs_qm.c
1216
XFS_FSB_TO_DADDR(mp, rablkno),
fs/xfs/xfs_qm.c
1217
mp->m_quotainfo->qi_dqchunklen,
fs/xfs/xfs_qm.c
1226
error = xfs_qm_reset_dqcounts_all(mp, firstid,
fs/xfs/xfs_qm.c
1255
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
1261
error = xfs_qm_dqget(mp, id, type, true, &dqp);
fs/xfs/xfs_qm.c
1317
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
1327
ASSERT(XFS_IS_QUOTA_ON(mp));
fs/xfs/xfs_qm.c
1333
if (xfs_is_quota_inode(&mp->m_sb, ino))
fs/xfs/xfs_qm.c
1340
error = xfs_iget(mp, tp, ino, XFS_IGET_DONTCACHE, 0, &ip);
fs/xfs/xfs_qm.c
1354
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_qm.c
1389
if (XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1396
if (XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1403
if (XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1458
xfs_mount_t *mp)
fs/xfs/xfs_qm.c
1463
struct xfs_inode *uip = mp->m_quotainfo->qi_uquotaip;
fs/xfs/xfs_qm.c
1464
struct xfs_inode *gip = mp->m_quotainfo->qi_gquotaip;
fs/xfs/xfs_qm.c
1465
struct xfs_inode *pip = mp->m_quotainfo->qi_pquotaip;
fs/xfs/xfs_qm.c
1470
ASSERT(XFS_IS_QUOTA_ON(mp));
fs/xfs/xfs_qm.c
1472
xfs_notice(mp, "Quotacheck needed: Please wait.");
fs/xfs/xfs_qm.c
1480
error = xfs_qm_reset_dqcounts_buf(mp, uip, XFS_DQTYPE_USER,
fs/xfs/xfs_qm.c
1488
error = xfs_qm_reset_dqcounts_buf(mp, gip, XFS_DQTYPE_GROUP,
fs/xfs/xfs_qm.c
1496
error = xfs_qm_reset_dqcounts_buf(mp, pip, XFS_DQTYPE_PROJ,
fs/xfs/xfs_qm.c
1503
xfs_set_quotacheck_running(mp);
fs/xfs/xfs_qm.c
1504
error = xfs_iwalk_threaded(mp, 0, 0, xfs_qm_dqusage_adjust, 0, true,
fs/xfs/xfs_qm.c
1506
xfs_clear_quotacheck_running(mp);
fs/xfs/xfs_qm.c
1520
if (XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1521
error = xfs_qm_dquot_walk(mp, XFS_DQTYPE_USER, xfs_qm_flush_one,
fs/xfs/xfs_qm.c
1524
if (XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1525
error2 = xfs_qm_dquot_walk(mp, XFS_DQTYPE_GROUP, xfs_qm_flush_one,
fs/xfs/xfs_qm.c
1530
if (XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1531
error2 = xfs_qm_dquot_walk(mp, XFS_DQTYPE_PROJ, xfs_qm_flush_one,
fs/xfs/xfs_qm.c
1556
mp->m_qflags &= ~XFS_ALL_QUOTA_CHKD;
fs/xfs/xfs_qm.c
1557
mp->m_qflags |= flags;
fs/xfs/xfs_qm.c
1563
xfs_warn(mp,
fs/xfs/xfs_qm.c
1569
ASSERT(mp->m_quotainfo != NULL);
fs/xfs/xfs_qm.c
1570
xfs_qm_destroy_quotainfo(mp);
fs/xfs/xfs_qm.c
1571
if (xfs_mount_reset_sbqflags(mp)) {
fs/xfs/xfs_qm.c
1572
xfs_warn(mp,
fs/xfs/xfs_qm.c
1575
xfs_fs_mark_sick(mp, XFS_SICK_FS_QUOTACHECK);
fs/xfs/xfs_qm.c
1577
xfs_notice(mp, "Quotacheck: Done.");
fs/xfs/xfs_qm.c
1578
xfs_fs_mark_healthy(mp, XFS_SICK_FS_QUOTACHECK);
fs/xfs/xfs_qm.c
1593
xfs_inodegc_flush(mp);
fs/xfs/xfs_qm.c
1594
xfs_qm_dqpurge_all(mp);
fs/xfs/xfs_qm.c
1610
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
1620
if (mp->m_sb.sb_rextents &&
fs/xfs/xfs_qm.c
1621
(!xfs_has_rtgroups(mp) || xfs_has_zoned(mp))) {
fs/xfs/xfs_qm.c
1622
xfs_notice(mp, "Cannot turn on quotas for realtime filesystem");
fs/xfs/xfs_qm.c
1623
mp->m_qflags = 0;
fs/xfs/xfs_qm.c
1627
ASSERT(XFS_IS_QUOTA_ON(mp));
fs/xfs/xfs_qm.c
1633
error = xfs_qm_init_quotainfo(mp);
fs/xfs/xfs_qm.c
1638
ASSERT(mp->m_quotainfo == NULL);
fs/xfs/xfs_qm.c
1639
mp->m_qflags = 0;
fs/xfs/xfs_qm.c
1645
if (XFS_QM_NEED_QUOTACHECK(mp)) {
fs/xfs/xfs_qm.c
1646
error = xfs_qm_quotacheck(mp);
fs/xfs/xfs_qm.c
1657
if (!XFS_IS_UQUOTA_ON(mp))
fs/xfs/xfs_qm.c
1658
mp->m_qflags &= ~XFS_UQUOTA_CHKD;
fs/xfs/xfs_qm.c
1659
if (!XFS_IS_GQUOTA_ON(mp))
fs/xfs/xfs_qm.c
1660
mp->m_qflags &= ~XFS_GQUOTA_CHKD;
fs/xfs/xfs_qm.c
1661
if (!XFS_IS_PQUOTA_ON(mp))
fs/xfs/xfs_qm.c
1662
mp->m_qflags &= ~XFS_PQUOTA_CHKD;
fs/xfs/xfs_qm.c
1669
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
1670
sbf = mp->m_sb.sb_qflags;
fs/xfs/xfs_qm.c
1671
mp->m_sb.sb_qflags = mp->m_qflags & XFS_MOUNT_QUOTA_ALL;
fs/xfs/xfs_qm.c
1672
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
1674
if (sbf != (mp->m_qflags & XFS_MOUNT_QUOTA_ALL)) {
fs/xfs/xfs_qm.c
1675
if (xfs_sync_sb(mp, false)) {
fs/xfs/xfs_qm.c
1682
ASSERT(!(XFS_IS_QUOTA_ON(mp)));
fs/xfs/xfs_qm.c
1683
xfs_alert(mp, "%s: Superblock update failed!",
fs/xfs/xfs_qm.c
1689
xfs_warn(mp, "Failed to initialize disk quotas, err %d.", error);
fs/xfs/xfs_qm.c
1704
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
1712
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_qm.c
1713
if (xfs_has_metadir(mp)) {
fs/xfs/xfs_qm.c
1733
xfs_mount_t *mp)
fs/xfs/xfs_qm.c
1741
ASSERT(mp->m_quotainfo);
fs/xfs/xfs_qm.c
1746
if (xfs_has_quota(mp)) {
fs/xfs/xfs_qm.c
1747
if (XFS_IS_UQUOTA_ON(mp) &&
fs/xfs/xfs_qm.c
1748
mp->m_sb.sb_uquotino != NULLFSINO) {
fs/xfs/xfs_qm.c
1749
ASSERT(mp->m_sb.sb_uquotino > 0);
fs/xfs/xfs_qm.c
1750
error = xfs_qm_qino_load(mp, XFS_DQTYPE_USER, &uip);
fs/xfs/xfs_qm.c
1754
if (XFS_IS_GQUOTA_ON(mp) &&
fs/xfs/xfs_qm.c
1755
mp->m_sb.sb_gquotino != NULLFSINO) {
fs/xfs/xfs_qm.c
1756
ASSERT(mp->m_sb.sb_gquotino > 0);
fs/xfs/xfs_qm.c
1757
error = xfs_qm_qino_load(mp, XFS_DQTYPE_GROUP, &gip);
fs/xfs/xfs_qm.c
1761
if (XFS_IS_PQUOTA_ON(mp) &&
fs/xfs/xfs_qm.c
1762
mp->m_sb.sb_pquotino != NULLFSINO) {
fs/xfs/xfs_qm.c
1763
ASSERT(mp->m_sb.sb_pquotino > 0);
fs/xfs/xfs_qm.c
1764
error = xfs_qm_qino_load(mp, XFS_DQTYPE_PROJ, &pip);
fs/xfs/xfs_qm.c
1778
if (XFS_IS_UQUOTA_ON(mp) && uip == NULL) {
fs/xfs/xfs_qm.c
1779
error = xfs_qm_qino_alloc(mp, &uip,
fs/xfs/xfs_qm.c
1786
if (XFS_IS_GQUOTA_ON(mp) && gip == NULL) {
fs/xfs/xfs_qm.c
1787
error = xfs_qm_qino_alloc(mp, &gip,
fs/xfs/xfs_qm.c
1794
if (XFS_IS_PQUOTA_ON(mp) && pip == NULL) {
fs/xfs/xfs_qm.c
1795
error = xfs_qm_qino_alloc(mp, &pip,
fs/xfs/xfs_qm.c
1801
mp->m_quotainfo->qi_uquotaip = uip;
fs/xfs/xfs_qm.c
1802
mp->m_quotainfo->qi_gquotaip = gip;
fs/xfs/xfs_qm.c
1803
mp->m_quotainfo->qi_pquotaip = pip;
fs/xfs/xfs_qm.c
1821
struct xfs_mount *mp = dqp->q_mount;
fs/xfs/xfs_qm.c
1822
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_qm.c
1857
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
1865
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_qm.c
1877
if (XFS_NOT_DQATTACHED(mp, ip)) {
fs/xfs/xfs_qm.c
1885
if ((flags & XFS_QMOPT_UQUOTA) && XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1888
error = xfs_qm_dqget(mp, from_kuid(user_ns, uid),
fs/xfs/xfs_qm.c
1903
if ((flags & XFS_QMOPT_GQUOTA) && XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1906
error = xfs_qm_dqget(mp, from_kgid(user_ns, gid),
fs/xfs/xfs_qm.c
1917
if ((flags & XFS_QMOPT_PQUOTA) && XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
1920
error = xfs_qm_dqget(mp, prid,
fs/xfs/xfs_qm.c
2032
struct xfs_mount *mp = i_tab[0]->i_mount;
fs/xfs/xfs_qm.c
2035
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_qm.c
2046
if (XFS_NOT_DQATTACHED(mp, ip)) {
fs/xfs/xfs_qm.c
205
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
2064
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_qm.c
2066
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_qm.c
207
xfs_qm_dquot_walk(mp, XFS_DQTYPE_USER, xfs_qm_dqpurge, NULL);
fs/xfs/xfs_qm.c
2072
if (udqp && XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
2078
if (gdqp && XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
208
xfs_qm_dquot_walk(mp, XFS_DQTYPE_GROUP, xfs_qm_dqpurge, NULL);
fs/xfs/xfs_qm.c
2084
if (pdqp && XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
209
xfs_qm_dquot_walk(mp, XFS_DQTYPE_PROJ, xfs_qm_dqpurge, NULL);
fs/xfs/xfs_qm.c
217
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
219
if (mp->m_quotainfo) {
fs/xfs/xfs_qm.c
220
xfs_qm_dqpurge_all(mp);
fs/xfs/xfs_qm.c
221
xfs_qm_destroy_quotainfo(mp);
fs/xfs/xfs_qm.c
227
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
229
struct xfs_rtgroup *rtg = xfs_rtgroup_grab(mp, 0);
fs/xfs/xfs_qm.c
267
xfs_mount_t *mp)
fs/xfs/xfs_qm.c
273
ASSERT(mp->m_rootip);
fs/xfs/xfs_qm.c
274
xfs_qm_dqdetach(mp->m_rootip);
fs/xfs/xfs_qm.c
280
if (!xfs_has_rtgroups(mp))
fs/xfs/xfs_qm.c
281
xfs_qm_unmount_rt(mp);
fs/xfs/xfs_qm.c
286
if (mp->m_quotainfo)
fs/xfs/xfs_qm.c
287
xfs_qm_destroy_quotainos(mp->m_quotainfo);
fs/xfs/xfs_qm.c
294
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm.c
296
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_qm.c
298
if (!XFS_NOT_DQATTACHED(mp, ip))
fs/xfs/xfs_qm.c
300
if (xfs_is_quota_inode(&mp->m_sb, ip->i_ino))
fs/xfs/xfs_qm.c
319
xfs_mount_t *mp = ip->i_mount;
fs/xfs/xfs_qm.c
328
if (XFS_IS_UQUOTA_ON(mp) && !ip->i_udquot) {
fs/xfs/xfs_qm.c
336
if (XFS_IS_GQUOTA_ON(mp) && !ip->i_gdquot) {
fs/xfs/xfs_qm.c
344
if (XFS_IS_PQUOTA_ON(mp) && !ip->i_pdquot) {
fs/xfs/xfs_qm.c
40
STATIC int xfs_qm_init_quotainos(struct xfs_mount *mp);
fs/xfs/xfs_qm.c
41
STATIC int xfs_qm_init_quotainfo(struct xfs_mount *mp);
fs/xfs/xfs_qm.c
537
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
545
error = xfs_qm_dqget_uncached(mp, 0, type, &dqp);
fs/xfs/xfs_qm.c
55
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
567
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
570
struct xfs_quotainfo *qinf = mp->m_quotainfo;
fs/xfs/xfs_qm.c
588
error = xfs_qm_dqget_uncached(mp, 0, type, &dqp);
fs/xfs/xfs_qm.c
60
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_qm.c
609
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
615
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_qm.c
625
if (XFS_IS_UQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
632
if (XFS_IS_GQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
639
if (XFS_IS_PQUOTA_ON(mp)) {
fs/xfs/xfs_qm.c
655
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
661
error = xfs_dqinode_mkdir_parent(mp, &qi->qi_dirip);
fs/xfs/xfs_qm.c
669
if (XFS_IS_CORRUPT(mp, qi->qi_dirip == NULL))
fs/xfs/xfs_qm.c
673
if (XFS_IS_UQUOTA_ON(mp) && !qi->qi_uquotaip) {
fs/xfs/xfs_qm.c
680
if (XFS_IS_GQUOTA_ON(mp) && !qi->qi_gquotaip) {
fs/xfs/xfs_qm.c
687
if (XFS_IS_PQUOTA_ON(mp) && !qi->qi_pquotaip) {
fs/xfs/xfs_qm.c
703
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
708
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_sb, 0, 0, 0, &tp);
fs/xfs/xfs_qm.c
712
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
714
xfs_add_quota(mp);
fs/xfs/xfs_qm.c
717
mp->m_sb.sb_qflags = mp->m_qflags & XFS_ALL_QUOTA_ACCT;
fs/xfs/xfs_qm.c
719
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
731
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
733
struct xfs_quotainfo *qi = mp->m_quotainfo;
fs/xfs/xfs_qm.c
736
if (!xfs_has_quota(mp)) {
fs/xfs/xfs_qm.c
737
error = xfs_qm_prep_metadir_sb(mp);
fs/xfs/xfs_qm.c
742
error = xfs_qm_load_metadir_qinos(mp, qi);
fs/xfs/xfs_qm.c
746
error = xfs_qm_create_metadir_qinos(mp, qi);
fs/xfs/xfs_qm.c
757
xfs_qm_destroy_quotainos(mp->m_quotainfo);
fs/xfs/xfs_qm.c
767
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
772
ASSERT(XFS_IS_QUOTA_ON(mp));
fs/xfs/xfs_qm.c
774
qinf = mp->m_quotainfo = kzalloc_obj(struct xfs_quotainfo,
fs/xfs/xfs_qm.c
785
if (xfs_has_metadir(mp))
fs/xfs/xfs_qm.c
786
error = xfs_qm_init_metadir_qinos(mp);
fs/xfs/xfs_qm.c
788
error = xfs_qm_init_quotainos(mp);
fs/xfs/xfs_qm.c
801
qinf->qi_dqchunklen = XFS_FSB_TO_BB(mp, XFS_DQUOT_CLUSTER_SIZE_FSB);
fs/xfs/xfs_qm.c
803
if (xfs_has_bigtime(mp)) {
fs/xfs/xfs_qm.c
812
trace_xfs_quota_expiry_range(mp, qinf->qi_expiry_min,
fs/xfs/xfs_qm.c
815
mp->m_qflags |= (mp->m_sb.sb_qflags & XFS_ALL_QUOTA_CHKD);
fs/xfs/xfs_qm.c
817
xfs_qm_init_timelimits(mp, XFS_DQTYPE_USER);
fs/xfs/xfs_qm.c
818
xfs_qm_init_timelimits(mp, XFS_DQTYPE_GROUP);
fs/xfs/xfs_qm.c
819
xfs_qm_init_timelimits(mp, XFS_DQTYPE_PROJ);
fs/xfs/xfs_qm.c
821
if (XFS_IS_UQUOTA_ON(mp))
fs/xfs/xfs_qm.c
822
xfs_qm_set_defquota(mp, XFS_DQTYPE_USER, qinf);
fs/xfs/xfs_qm.c
823
if (XFS_IS_GQUOTA_ON(mp))
fs/xfs/xfs_qm.c
824
xfs_qm_set_defquota(mp, XFS_DQTYPE_GROUP, qinf);
fs/xfs/xfs_qm.c
825
if (XFS_IS_PQUOTA_ON(mp))
fs/xfs/xfs_qm.c
826
xfs_qm_set_defquota(mp, XFS_DQTYPE_PROJ, qinf);
fs/xfs/xfs_qm.c
829
mp->m_super->s_id);
fs/xfs/xfs_qm.c
854
mp->m_quotainfo = NULL;
fs/xfs/xfs_qm.c
865
struct xfs_mount *mp)
fs/xfs/xfs_qm.c
869
qi = mp->m_quotainfo;
fs/xfs/xfs_qm.c
878
mp->m_quotainfo = NULL;
fs/xfs/xfs_qm.c
898
struct xfs_mount *mp,
fs/xfs/xfs_qm.c
915
if (!xfs_has_pquotino(mp) &&
fs/xfs/xfs_qm.c
920
(mp->m_sb.sb_gquotino != NULLFSINO)) {
fs/xfs/xfs_qm.c
921
ino = mp->m_sb.sb_gquotino;
fs/xfs/xfs_qm.c
922
if (XFS_IS_CORRUPT(mp,
fs/xfs/xfs_qm.c
923
mp->m_sb.sb_pquotino != NULLFSINO)) {
fs/xfs/xfs_qm.c
924
xfs_fs_mark_sick(mp, XFS_SICK_FS_PQUOTA);
fs/xfs/xfs_qm.c
928
(mp->m_sb.sb_pquotino != NULLFSINO)) {
fs/xfs/xfs_qm.c
929
ino = mp->m_sb.sb_pquotino;
fs/xfs/xfs_qm.c
930
if (XFS_IS_CORRUPT(mp,
fs/xfs/xfs_qm.c
931
mp->m_sb.sb_gquotino != NULLFSINO)) {
fs/xfs/xfs_qm.c
932
xfs_fs_mark_sick(mp, XFS_SICK_FS_GQUOTA);
fs/xfs/xfs_qm.c
937
error = xfs_metafile_iget(mp, ino, metafile_type, ipp);
fs/xfs/xfs_qm.c
941
mp->m_sb.sb_gquotino = NULLFSINO;
fs/xfs/xfs_qm.c
942
mp->m_sb.sb_pquotino = NULLFSINO;
fs/xfs/xfs_qm.c
947
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_create,
fs/xfs/xfs_qm.c
948
need_alloc ? XFS_QM_QINOCREATE_SPACE_RES(mp) : 0,
fs/xfs/xfs_qm.c
967
if (xfs_has_metadir(mp))
fs/xfs/xfs_qm.c
976
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
978
ASSERT(!xfs_has_quota(mp));
fs/xfs/xfs_qm.c
980
xfs_add_quota(mp);
fs/xfs/xfs_qm.c
981
mp->m_sb.sb_uquotino = NULLFSINO;
fs/xfs/xfs_qm.c
982
mp->m_sb.sb_gquotino = NULLFSINO;
fs/xfs/xfs_qm.c
983
mp->m_sb.sb_pquotino = NULLFSINO;
fs/xfs/xfs_qm.c
986
mp->m_sb.sb_qflags = mp->m_qflags & XFS_ALL_QUOTA_ACCT;
fs/xfs/xfs_qm.c
989
mp->m_sb.sb_uquotino = (*ipp)->i_ino;
fs/xfs/xfs_qm.c
991
mp->m_sb.sb_gquotino = (*ipp)->i_ino;
fs/xfs/xfs_qm.c
993
mp->m_sb.sb_pquotino = (*ipp)->i_ino;
fs/xfs/xfs_qm.c
994
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_qm.c
999
ASSERT(xfs_is_shutdown(mp));
fs/xfs/xfs_qm.h
101
return mp->m_quotainfo->qi_uquotaip;
fs/xfs/xfs_qm.h
103
return mp->m_quotainfo->qi_gquotaip;
fs/xfs/xfs_qm.h
105
return mp->m_quotainfo->qi_pquotaip;
fs/xfs/xfs_qm.h
157
extern int xfs_qm_scall_getquota(struct xfs_mount *mp,
fs/xfs/xfs_qm.h
161
extern int xfs_qm_scall_getquota_next(struct xfs_mount *mp,
fs/xfs/xfs_qm.h
165
extern int xfs_qm_scall_setqlim(struct xfs_mount *mp,
fs/xfs/xfs_qm.h
188
int xfs_qm_qino_load(struct xfs_mount *mp, xfs_dqtype_t type,
fs/xfs/xfs_qm.h
97
xfs_quota_inode(struct xfs_mount *mp, xfs_dqtype_t type)
fs/xfs/xfs_qm_bhv.c
101
(xfs_dev_is_read_only(mp, "changing quota state") ||
fs/xfs/xfs_qm_bhv.c
102
xfs_has_norecovery(mp));
fs/xfs/xfs_qm_bhv.c
107
xfs_mount_t *mp,
fs/xfs/xfs_qm_bhv.c
114
quotaondisk = xfs_has_quota(mp) &&
fs/xfs/xfs_qm_bhv.c
115
(mp->m_sb.sb_qflags & XFS_ALL_QUOTA_ACCT);
fs/xfs/xfs_qm_bhv.c
118
uquotaondisk = mp->m_sb.sb_qflags & XFS_UQUOTA_ACCT;
fs/xfs/xfs_qm_bhv.c
119
pquotaondisk = mp->m_sb.sb_qflags & XFS_PQUOTA_ACCT;
fs/xfs/xfs_qm_bhv.c
120
gquotaondisk = mp->m_sb.sb_qflags & XFS_GQUOTA_ACCT;
fs/xfs/xfs_qm_bhv.c
131
if (xfs_qm_validate_state_change(mp, uquotaondisk,
fs/xfs/xfs_qm_bhv.c
134
if (xfs_has_metadir(mp))
fs/xfs/xfs_qm_bhv.c
135
xfs_warn(mp,
fs/xfs/xfs_qm_bhv.c
138
xfs_warn(mp, "please mount with%s%s%s%s.",
fs/xfs/xfs_qm_bhv.c
146
if (XFS_IS_QUOTA_ON(mp) || quotaondisk) {
fs/xfs/xfs_qm_bhv.c
151
if (quotaondisk && !XFS_QM_NEED_QUOTACHECK(mp)) {
fs/xfs/xfs_qm_bhv.c
158
xfs_qm_mount_quotas(mp);
fs/xfs/xfs_qm_bhv.c
168
*quotaflags = mp->m_qflags;
fs/xfs/xfs_qm_bhv.c
169
mp->m_qflags = 0;
fs/xfs/xfs_qm_bhv.c
183
struct xfs_mount *mp)
fs/xfs/xfs_qm_bhv.c
185
if (!xfs_has_metadir(mp))
fs/xfs/xfs_qm_bhv.c
187
if (xfs_has_norecovery(mp))
fs/xfs/xfs_qm_bhv.c
190
mp->m_qflags = mp->m_sb.sb_qflags & (XFS_ALL_QUOTA_ACCT |
fs/xfs/xfs_qm_bhv.c
72
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_qm_bhv.c
75
if (!xfs_qm_dqget(mp, ip->i_projid, XFS_DQTYPE_PROJ, false, &dqp)) {
fs/xfs/xfs_qm_bhv.c
85
struct xfs_mount *mp,
fs/xfs/xfs_qm_bhv.c
93
state = ((uqd && !XFS_IS_UQUOTA_ON(mp)) ||
fs/xfs/xfs_qm_bhv.c
94
(!uqd && XFS_IS_UQUOTA_ON(mp)) ||
fs/xfs/xfs_qm_bhv.c
95
(gqd && !XFS_IS_GQUOTA_ON(mp)) ||
fs/xfs/xfs_qm_bhv.c
96
(!gqd && XFS_IS_GQUOTA_ON(mp)) ||
fs/xfs/xfs_qm_bhv.c
97
(pqd && !XFS_IS_PQUOTA_ON(mp)) ||
fs/xfs/xfs_qm_bhv.c
98
(!pqd && XFS_IS_PQUOTA_ON(mp)));
fs/xfs/xfs_qm_syscalls.c
102
xfs_mount_t *mp,
fs/xfs/xfs_qm_syscalls.c
107
if (!xfs_has_quota(mp) || flags == 0 ||
fs/xfs/xfs_qm_syscalls.c
109
xfs_debug(mp, "%s: flags=%x m_qflags=%x",
fs/xfs/xfs_qm_syscalls.c
110
__func__, flags, mp->m_qflags);
fs/xfs/xfs_qm_syscalls.c
115
error = xfs_qm_scall_trunc_qfile(mp, XFS_DQTYPE_USER);
fs/xfs/xfs_qm_syscalls.c
120
error = xfs_qm_scall_trunc_qfile(mp, XFS_DQTYPE_GROUP);
fs/xfs/xfs_qm_syscalls.c
125
error = xfs_qm_scall_trunc_qfile(mp, XFS_DQTYPE_PROJ);
fs/xfs/xfs_qm_syscalls.c
137
xfs_mount_t *mp,
fs/xfs/xfs_qm_syscalls.c
150
xfs_debug(mp, "%s: zero flags, m_qflags=%x",
fs/xfs/xfs_qm_syscalls.c
151
__func__, mp->m_qflags);
fs/xfs/xfs_qm_syscalls.c
160
if (((mp->m_sb.sb_qflags & XFS_UQUOTA_ACCT) == 0 &&
fs/xfs/xfs_qm_syscalls.c
162
((mp->m_sb.sb_qflags & XFS_GQUOTA_ACCT) == 0 &&
fs/xfs/xfs_qm_syscalls.c
164
((mp->m_sb.sb_qflags & XFS_PQUOTA_ACCT) == 0 &&
fs/xfs/xfs_qm_syscalls.c
166
xfs_debug(mp,
fs/xfs/xfs_qm_syscalls.c
168
__func__, flags, mp->m_sb.sb_qflags);
fs/xfs/xfs_qm_syscalls.c
174
if ((mp->m_qflags & flags) == flags)
fs/xfs/xfs_qm_syscalls.c
181
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_qm_syscalls.c
182
qf = mp->m_sb.sb_qflags;
fs/xfs/xfs_qm_syscalls.c
183
mp->m_sb.sb_qflags = qf | flags;
fs/xfs/xfs_qm_syscalls.c
184
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_qm_syscalls.c
192
error = xfs_sync_sb(mp, false);
fs/xfs/xfs_qm_syscalls.c
198
if (((mp->m_sb.sb_qflags & XFS_UQUOTA_ACCT) !=
fs/xfs/xfs_qm_syscalls.c
199
(mp->m_qflags & XFS_UQUOTA_ACCT)) ||
fs/xfs/xfs_qm_syscalls.c
200
((mp->m_sb.sb_qflags & XFS_PQUOTA_ACCT) !=
fs/xfs/xfs_qm_syscalls.c
201
(mp->m_qflags & XFS_PQUOTA_ACCT)) ||
fs/xfs/xfs_qm_syscalls.c
202
((mp->m_sb.sb_qflags & XFS_GQUOTA_ACCT) !=
fs/xfs/xfs_qm_syscalls.c
203
(mp->m_qflags & XFS_GQUOTA_ACCT)))
fs/xfs/xfs_qm_syscalls.c
206
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_qm_syscalls.c
212
mutex_lock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/xfs_qm_syscalls.c
213
mp->m_qflags |= (flags & XFS_ALL_QUOTA_ENFD);
fs/xfs/xfs_qm_syscalls.c
214
mutex_unlock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/xfs_qm_syscalls.c
227
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
236
xfs_debug(mp, "%shard %lld < %ssoft %lld", tag, hard, tag,
fs/xfs/xfs_qm_syscalls.c
24
xfs_mount_t *mp,
fs/xfs/xfs_qm_syscalls.c
253
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
264
res->timer = xfs_dquot_set_timeout(mp, timer);
fs/xfs/xfs_qm_syscalls.c
273
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
278
struct xfs_quotainfo *q = mp->m_quotainfo;
fs/xfs/xfs_qm_syscalls.c
299
error = xfs_qm_dqget(mp, id, type, true, &dqp);
fs/xfs/xfs_qm_syscalls.c
307
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_qm_setqlim, 0, 0, 0, &tp);
fs/xfs/xfs_qm_syscalls.c
32
if ((mp->m_qflags & flags) == 0)
fs/xfs/xfs_qm_syscalls.c
333
(xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_spc_hardlimit) :
fs/xfs/xfs_qm_syscalls.c
336
(xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_spc_softlimit) :
fs/xfs/xfs_qm_syscalls.c
341
if (xfs_setqlim_limits(mp, res, qlim, hard, soft, "blk"))
fs/xfs/xfs_qm_syscalls.c
344
xfs_setqlim_timer(mp, res, qlim, newlim->d_spc_timer);
fs/xfs/xfs_qm_syscalls.c
348
(xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_rt_spc_hardlimit) :
fs/xfs/xfs_qm_syscalls.c
351
(xfs_qcnt_t) XFS_B_TO_FSB(mp, newlim->d_rt_spc_softlimit) :
fs/xfs/xfs_qm_syscalls.c
356
xfs_setqlim_limits(mp, res, qlim, hard, soft, "rtb");
fs/xfs/xfs_qm_syscalls.c
358
xfs_setqlim_timer(mp, res, qlim, newlim->d_rt_spc_timer);
fs/xfs/xfs_qm_syscalls.c
370
xfs_setqlim_limits(mp, res, qlim, hard, soft, "ino");
fs/xfs/xfs_qm_syscalls.c
372
xfs_setqlim_timer(mp, res, qlim, newlim->d_ino_timer);
fs/xfs/xfs_qm_syscalls.c
397
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
40
xfs_info(mp, "disabling of quota accounting not supported.");
fs/xfs/xfs_qm_syscalls.c
403
dst->d_spc_hardlimit = XFS_FSB_TO_B(mp, dqp->q_blk.hardlimit);
fs/xfs/xfs_qm_syscalls.c
404
dst->d_spc_softlimit = XFS_FSB_TO_B(mp, dqp->q_blk.softlimit);
fs/xfs/xfs_qm_syscalls.c
407
dst->d_space = XFS_FSB_TO_B(mp, dqp->q_blk.reserved);
fs/xfs/xfs_qm_syscalls.c
413
dst->d_rt_spc_hardlimit = XFS_FSB_TO_B(mp, dqp->q_rtb.hardlimit);
fs/xfs/xfs_qm_syscalls.c
414
dst->d_rt_spc_softlimit = XFS_FSB_TO_B(mp, dqp->q_rtb.softlimit);
fs/xfs/xfs_qm_syscalls.c
415
dst->d_rt_space = XFS_FSB_TO_B(mp, dqp->q_rtb.reserved);
fs/xfs/xfs_qm_syscalls.c
42
mutex_lock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/xfs_qm_syscalls.c
43
mp->m_qflags &= ~(flags & XFS_ALL_QUOTA_ENFD);
fs/xfs/xfs_qm_syscalls.c
434
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
44
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_qm_syscalls.c
447
xfs_inodegc_push(mp);
fs/xfs/xfs_qm_syscalls.c
45
mp->m_sb.sb_qflags = mp->m_qflags;
fs/xfs/xfs_qm_syscalls.c
453
error = xfs_qm_dqget(mp, id, type, false, &dqp);
fs/xfs/xfs_qm_syscalls.c
46
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_qm_syscalls.c
467
xfs_qm_scall_getquota_fill_qc(mp, type, dqp, dst);
fs/xfs/xfs_qm_syscalls.c
47
mutex_unlock(&mp->m_quotainfo->qi_quotaofflock);
fs/xfs/xfs_qm_syscalls.c
481
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
491
xfs_inodegc_push(mp);
fs/xfs/xfs_qm_syscalls.c
493
error = xfs_qm_dqget_next(mp, *id, type, &dqp);
fs/xfs/xfs_qm_syscalls.c
50
return xfs_sync_sb(mp, false);
fs/xfs/xfs_qm_syscalls.c
500
xfs_qm_scall_getquota_fill_qc(mp, type, dqp, dst);
fs/xfs/xfs_qm_syscalls.c
55
struct xfs_mount *mp,
fs/xfs/xfs_qm_syscalls.c
62
error = xfs_qm_qino_load(mp, type, &ip);
fs/xfs/xfs_qm_syscalls.c
70
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, 0, 0, 0, &tp);
fs/xfs/xfs_quota.h
124
void xfs_qm_resume_quotaon(struct xfs_mount *mp);
fs/xfs/xfs_quota.h
180
struct xfs_mount *mp, struct xfs_dquot *udqp,
fs/xfs/xfs_quota.h
202
#define xfs_qm_newmount(mp, a, b) (0)
fs/xfs/xfs_quota.h
203
#define xfs_qm_resume_quotaon(mp) ((void)0)
fs/xfs/xfs_quota.h
204
#define xfs_qm_mount_quotas(mp)
fs/xfs/xfs_quota.h
205
#define xfs_qm_unmount(mp)
fs/xfs/xfs_quota.h
206
#define xfs_qm_unmount_quotas(mp)
fs/xfs/xfs_quota.h
27
#define XFS_NOT_DQATTACHED(mp, ip) \
fs/xfs/xfs_quota.h
28
((XFS_IS_UQUOTA_ON(mp) && (ip)->i_udquot == NULL) || \
fs/xfs/xfs_quota.h
29
(XFS_IS_GQUOTA_ON(mp) && (ip)->i_gdquot == NULL) || \
fs/xfs/xfs_quota.h
30
(XFS_IS_PQUOTA_ON(mp) && (ip)->i_pdquot == NULL))
fs/xfs/xfs_quota.h
32
#define XFS_QM_NEED_QUOTACHECK(mp) \
fs/xfs/xfs_quota.h
33
((XFS_IS_UQUOTA_ON(mp) && \
fs/xfs/xfs_quota.h
34
(mp->m_sb.sb_qflags & XFS_UQUOTA_CHKD) == 0) || \
fs/xfs/xfs_quota.h
35
(XFS_IS_GQUOTA_ON(mp) && \
fs/xfs/xfs_quota.h
36
(mp->m_sb.sb_qflags & XFS_GQUOTA_CHKD) == 0) || \
fs/xfs/xfs_quota.h
37
(XFS_IS_PQUOTA_ON(mp) && \
fs/xfs/xfs_quota.h
38
(mp->m_sb.sb_qflags & XFS_PQUOTA_CHKD) == 0))
fs/xfs/xfs_quotaops.c
121
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
126
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
141
return xfs_qm_scall_setqlim(mp, 0, xfs_quota_type(type), &newlim);
fs/xfs/xfs_quotaops.c
170
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
174
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
177
return xfs_qm_scall_quotaon(mp, xfs_quota_flags(uflags));
fs/xfs/xfs_quotaops.c
185
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
189
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
192
return xfs_qm_scall_quotaoff(mp, xfs_quota_flags(uflags));
fs/xfs/xfs_quotaops.c
200
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
206
if (XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
219
return xfs_qm_scall_trunc_qfiles(mp, flags);
fs/xfs/xfs_quotaops.c
22
struct xfs_mount *mp,
fs/xfs/xfs_quotaops.c
228
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
231
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
235
return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq);
fs/xfs/xfs_quotaops.c
246
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
249
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
253
ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type),
fs/xfs/xfs_quotaops.c
269
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
273
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
276
return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid),
fs/xfs/xfs_quotaops.c
29
error = xfs_qm_qino_load(mp, type, &ip);
fs/xfs/xfs_quotaops.c
35
defq = xfs_get_defquota(mp->m_quotainfo, type);
fs/xfs/xfs_quotaops.c
61
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
62
struct xfs_quotainfo *q = mp->m_quotainfo;
fs/xfs/xfs_quotaops.c
66
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
69
if (XFS_IS_UQUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
71
if (XFS_IS_UQUOTA_ENFORCED(mp))
fs/xfs/xfs_quotaops.c
73
if (XFS_IS_GQUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
75
if (XFS_IS_GQUOTA_ENFORCED(mp))
fs/xfs/xfs_quotaops.c
77
if (XFS_IS_PQUOTA_ON(mp))
fs/xfs/xfs_quotaops.c
79
if (XFS_IS_PQUOTA_ENFORCED(mp))
fs/xfs/xfs_quotaops.c
82
error = xfs_qm_fill_state(&state->s_state[USRQUOTA], mp,
fs/xfs/xfs_quotaops.c
86
error = xfs_qm_fill_state(&state->s_state[GRPQUOTA], mp,
fs/xfs/xfs_quotaops.c
90
error = xfs_qm_fill_state(&state->s_state[PRJQUOTA], mp,
fs/xfs/xfs_refcount_item.c
146
struct xfs_mount *mp,
fs/xfs/xfs_refcount_item.c
162
xfs_log_item_init(mp, &cuip->cui_item, item_type, &xfs_cui_item_ops);
fs/xfs/xfs_refcount_item.c
314
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_refcount_item.c
320
cuip = xfs_cui_init(mp, item_type, count);
fs/xfs/xfs_refcount_item.c
322
list_sort(mp, items, xfs_refcount_update_diff_items);
fs/xfs/xfs_refcount_item.c
370
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_refcount_item.c
379
ri->ri_group = xfs_group_intent_get(mp, ri->ri_startblock,
fs/xfs/xfs_refcount_item.c
382
trace_xfs_refcount_defer(mp, ri);
fs/xfs/xfs_refcount_item.c
450
struct xfs_mount *mp,
fs/xfs/xfs_refcount_item.c
454
if (!xfs_has_reflink(mp))
fs/xfs/xfs_refcount_item.c
471
return xfs_verify_rtbext(mp, pmap->pe_startblock, pmap->pe_len);
fs/xfs/xfs_refcount_item.c
473
return xfs_verify_fsbext(mp, pmap->pe_startblock, pmap->pe_len);
fs/xfs/xfs_refcount_item.c
478
struct xfs_mount *mp,
fs/xfs/xfs_refcount_item.c
490
ri->ri_group = xfs_group_intent_get(mp, pmap->pe_startblock,
fs/xfs/xfs_refcount_item.c
510
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_refcount_item.c
521
if (!xfs_cui_validate_phys(mp, isrt,
fs/xfs/xfs_refcount_item.c
523
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
529
xfs_cui_recover_work(mp, dfp, isrt,
fs/xfs/xfs_refcount_item.c
545
resv = xlog_recover_resv(&M_RES(mp)->tr_itruncate);
fs/xfs/xfs_refcount_item.c
546
error = xfs_trans_alloc(mp, &resv, mp->m_refc_maxlevels * 2, 0,
fs/xfs/xfs_refcount_item.c
553
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
713
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_refcount_item.c
721
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
728
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
733
cuip = xfs_cui_init(mp, ITEM_TYPE(item), cui_formatp->cui_nextents);
fs/xfs/xfs_refcount_item.c
755
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_refcount_item.c
763
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
770
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_refcount_item.c
775
cuip = xfs_cui_init(mp, ITEM_TYPE(item), cui_formatp->cui_nextents);
fs/xfs/xfs_reflink.c
1002
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1008
offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_reflink.c
1009
end_fsb = XFS_B_TO_FSB(mp, offset + count);
fs/xfs/xfs_reflink.c
1016
XFS_NEXTENTADD_SPACE_RES(mp, 1, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
1018
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_atomic_ioend, resblks, 0,
fs/xfs/xfs_reflink.c
1046
struct xfs_mount *mp)
fs/xfs/xfs_reflink.c
1049
if (!xfs_can_sw_atomic_write(mp))
fs/xfs/xfs_reflink.c
1056
return rounddown_pow_of_two(xfs_calc_max_atomic_write_fsblocks(mp));
fs/xfs/xfs_reflink.c
1067
struct xfs_mount *mp)
fs/xfs/xfs_reflink.c
1073
if (!xfs_has_reflink(mp))
fs/xfs/xfs_reflink.c
1076
while ((pag = xfs_perag_next(mp, pag))) {
fs/xfs/xfs_reflink.c
1084
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/xfs_reflink.c
1179
struct xfs_mount *mp = src->i_mount;
fs/xfs/xfs_reflink.c
1186
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
fs/xfs/xfs_reflink.c
1238
struct xfs_mount *mp = dest->i_mount;
fs/xfs/xfs_reflink.c
1245
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
fs/xfs/xfs_reflink.c
1283
struct xfs_mount *mp,
fs/xfs/xfs_reflink.c
1291
if (!xfs_has_rmapbt(mp))
fs/xfs/xfs_reflink.c
1294
if (xfs_metafile_resv_critical(mp))
fs/xfs/xfs_reflink.c
1299
agno = XFS_FSB_TO_AGNO(mp, fsb);
fs/xfs/xfs_reflink.c
1300
pag = xfs_perag_get(mp, agno);
fs/xfs/xfs_reflink.c
1319
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1353
resblks = XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
1361
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_reflink.c
1365
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write,
fs/xfs/xfs_reflink.c
137
struct xfs_mount *mp,
fs/xfs/xfs_reflink.c
1413
error = xfs_reflink_ag_has_free_space(mp, ip,
fs/xfs/xfs_reflink.c
150
pag = xfs_perag_get(mp, XFS_FSB_TO_AGNO(mp, irec->br_startblock));
fs/xfs/xfs_reflink.c
1502
newlen = XFS_FSB_TO_B(mp, dmap->br_startoff + dmap->br_blockcount);
fs/xfs/xfs_reflink.c
151
orig_bno = XFS_FSB_TO_AGBNO(mp, irec->br_startblock);
fs/xfs/xfs_reflink.c
1536
struct xfs_mount *mp = src->i_mount;
fs/xfs/xfs_reflink.c
1537
xfs_fileoff_t srcoff = XFS_B_TO_FSBT(mp, pos_in);
fs/xfs/xfs_reflink.c
1538
xfs_fileoff_t destoff = XFS_B_TO_FSBT(mp, pos_out);
fs/xfs/xfs_reflink.c
1545
len = min_t(xfs_filblks_t, XFS_B_TO_FSB(mp, remap_len),
fs/xfs/xfs_reflink.c
157
cur = xfs_refcountbt_init_cursor(mp, tp, agbp, pag);
fs/xfs/xfs_reflink.c
1740
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1761
error = xfs_reflink_find_rtshared(mp, tp, &got,
fs/xfs/xfs_reflink.c
1764
error = xfs_reflink_find_shared(mp, tp, &got,
fs/xfs/xfs_reflink.c
181
struct xfs_mount *mp,
fs/xfs/xfs_reflink.c
1830
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
1835
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, 0, 0, 0, &tp);
fs/xfs/xfs_reflink.c
1912
struct xfs_mount *mp,
fs/xfs/xfs_reflink.c
1916
if (!xfs_has_rtgroups(mp))
fs/xfs/xfs_reflink.c
201
orig_bno = xfs_rtb_to_rgbno(mp, irec->br_startblock);
fs/xfs/xfs_reflink.c
202
rtg = xfs_rtgroup_get(mp, xfs_rtb_to_rgno(mp, irec->br_startblock));
fs/xfs/xfs_reflink.c
233
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
246
error = xfs_reflink_find_rtshared(mp, NULL, irec,
fs/xfs/xfs_reflink.c
249
error = xfs_reflink_find_shared(mp, NULL, irec,
fs/xfs/xfs_reflink.c
341
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
342
xfs_fileoff_t offset_fsb = XFS_B_TO_FSBT(mp, offset);
fs/xfs/xfs_reflink.c
343
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + count);
fs/xfs/xfs_reflink.c
440
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
451
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, 0);
fs/xfs/xfs_reflink.c
454
dblocks = XFS_DIOSTRAT_SPACE_RES(mp, resaligned);
fs/xfs/xfs_reflink.c
461
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, dblocks,
fs/xfs/xfs_reflink.c
507
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
517
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_write, 0, 0,
fs/xfs/xfs_reflink.c
907
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_reflink.c
912
resblks = XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK);
fs/xfs/xfs_reflink.c
913
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, resblks, 0,
fs/xfs/xfs_reflink.h
50
extern int xfs_reflink_recover_cow(struct xfs_mount *mp);
fs/xfs/xfs_reflink.h
69
bool xfs_reflink_supports_rextsize(struct xfs_mount *mp, unsigned int rextsize);
fs/xfs/xfs_reflink.h
71
xfs_extlen_t xfs_reflink_max_atomic_cow(struct xfs_mount *mp);
fs/xfs/xfs_rmap_item.c
146
struct xfs_mount *mp,
fs/xfs/xfs_rmap_item.c
163
xfs_log_item_init(mp, &ruip->rui_item, item_type, &xfs_rui_item_ops);
fs/xfs/xfs_rmap_item.c
339
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_rmap_item.c
345
ruip = xfs_rui_init(mp, item_type, count);
fs/xfs/xfs_rmap_item.c
347
list_sort(mp, items, xfs_rmap_update_diff_items);
fs/xfs/xfs_rmap_item.c
395
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_rmap_item.c
404
ri->ri_group = xfs_group_intent_get(mp, ri->ri_bmap.br_startblock,
fs/xfs/xfs_rmap_item.c
407
trace_xfs_rmap_defer(mp, ri);
fs/xfs/xfs_rmap_item.c
469
struct xfs_mount *mp,
fs/xfs/xfs_rmap_item.c
473
if (!xfs_has_rmapbt(mp))
fs/xfs/xfs_rmap_item.c
494
!xfs_verify_ino(mp, map->me_owner))
fs/xfs/xfs_rmap_item.c
497
if (!xfs_verify_fileext(mp, map->me_startoff, map->me_len))
fs/xfs/xfs_rmap_item.c
501
return xfs_verify_rtbext(mp, map->me_startblock, map->me_len);
fs/xfs/xfs_rmap_item.c
503
return xfs_verify_fsbext(mp, map->me_startblock, map->me_len);
fs/xfs/xfs_rmap_item.c
508
struct xfs_mount *mp,
fs/xfs/xfs_rmap_item.c
555
ri->ri_group = xfs_group_intent_get(mp, map->me_startblock,
fs/xfs/xfs_rmap_item.c
575
struct xfs_mount *mp = lip->li_log->l_mp;
fs/xfs/xfs_rmap_item.c
586
if (!xfs_rui_validate_map(mp, isrt,
fs/xfs/xfs_rmap_item.c
588
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
594
xfs_rui_recover_work(mp, dfp, isrt,
fs/xfs/xfs_rmap_item.c
598
resv = xlog_recover_resv(&M_RES(mp)->tr_itruncate);
fs/xfs/xfs_rmap_item.c
599
error = xfs_trans_alloc(mp, &resv, mp->m_rmap_maxlevels, 0,
fs/xfs/xfs_rmap_item.c
606
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
742
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_rmap_item.c
750
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
757
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
762
ruip = xfs_rui_init(mp, ITEM_TYPE(item), rui_formatp->rui_nextents);
fs/xfs/xfs_rmap_item.c
784
struct xfs_mount *mp = log->l_mp;
fs/xfs/xfs_rmap_item.c
792
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
799
XFS_CORRUPTION_ERROR(__func__, XFS_ERRLEVEL_LOW, mp,
fs/xfs/xfs_rmap_item.c
804
ruip = xfs_rui_init(mp, ITEM_TYPE(item), rui_formatp->rui_nextents);
fs/xfs/xfs_rtalloc.c
1013
if (mp->m_sb.sb_rbmblocks != nmp->m_sb.sb_rbmblocks ||
fs/xfs/xfs_rtalloc.c
1014
mp->m_rsumlevels != nmp->m_rsumlevels) {
fs/xfs/xfs_rtalloc.c
1045
mp->m_rsumlevels = nmp->m_rsumlevels;
fs/xfs/xfs_rtalloc.c
1046
mp->m_rsumblocks = nmp->m_rsumblocks;
fs/xfs/xfs_rtalloc.c
1051
xfs_trans_resv_calc(mp, &mp->m_resv);
fs/xfs/xfs_rtalloc.c
1059
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_rtalloc.c
1070
mp->m_features |= XFS_FEAT_REALTIME;
fs/xfs/xfs_rtalloc.c
1071
xfs_rtrmapbt_compute_maxlevels(mp);
fs/xfs/xfs_rtalloc.c
1072
xfs_rtrefcountbt_compute_maxlevels(mp);
fs/xfs/xfs_rtalloc.c
108
for (log = oargs->mp->m_rsumlevels - 1; log >= 0; log--) {
fs/xfs/xfs_rtalloc.c
1086
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1088
return mp->m_sb.sb_rextents -
fs/xfs/xfs_rtalloc.c
1089
((xfs_rtxnum_t)(mp->m_sb.sb_rgcount - 1) *
fs/xfs/xfs_rtalloc.c
109
for (bbno = oargs->mp->m_sb.sb_rbmblocks - 1;
fs/xfs/xfs_rtalloc.c
1090
mp->m_sb.sb_rgextents);
fs/xfs/xfs_rtalloc.c
1111
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
1116
ASSERT(!mp->m_sb.sb_rgcount || rgno >= mp->m_sb.sb_rgcount - 1);
fs/xfs/xfs_rtalloc.c
1118
if (mp->m_sb.sb_rgcount && rgno == mp->m_sb.sb_rgcount - 1) {
fs/xfs/xfs_rtalloc.c
1119
xfs_rtxnum_t nrext = xfs_last_rtgroup_extents(mp);
fs/xfs/xfs_rtalloc.c
1123
bmbno = xfs_rtbitmap_blockcount_len(mp, nrext) - 1;
fs/xfs/xfs_rtalloc.c
1124
div_u64_rem(nrext, xfs_rtbitmap_rtx_per_rbmblock(mp), &mod);
fs/xfs/xfs_rtalloc.c
1147
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
115
if (XFS_IS_CORRUPT(oargs->mp, sum < 0)) {
fs/xfs/xfs_rtalloc.c
1155
nmp = xfs_growfs_rt_alloc_fake_mount(mp, nrblocks, rextsize);
fs/xfs/xfs_rtalloc.c
1160
if (xfs_has_rtgroups(mp)) {
fs/xfs/xfs_rtalloc.c
1173
if (mp->m_sb.sb_rgcount &&
fs/xfs/xfs_rtalloc.c
1174
rtg_rgno(rtg) == mp->m_sb.sb_rgcount - 1)
fs/xfs/xfs_rtalloc.c
1181
orbmblocks = XFS_B_TO_FSB(mp, rbmip->i_disk_size);
fs/xfs/xfs_rtalloc.c
1182
orsumblocks = XFS_B_TO_FSB(mp, rsumip->i_disk_size);
fs/xfs/xfs_rtalloc.c
1198
struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1210
rtg = xfs_rtgroup_grab(mp, rgno);
fs/xfs/xfs_rtalloc.c
1220
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_rtalloc.c
1250
xfs_rtgroup_calc_geometry(mp, rtg, rtg_rgno(rtg), mp->m_sb.sb_rgcount,
fs/xfs/xfs_rtalloc.c
1251
mp->m_sb.sb_rextents);
fs/xfs/xfs_rtalloc.c
1263
const struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1271
nmp = xfs_growfs_rt_alloc_fake_mount(mp, rblocks, rextsize);
fs/xfs/xfs_rtalloc.c
1288
trace_xfs_growfs_check_rtgeom(mp, min_logfsbs);
fs/xfs/xfs_rtalloc.c
1290
if (min_logfsbs > mp->m_sb.sb_logblocks)
fs/xfs/xfs_rtalloc.c
1293
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_rtalloc.c
1294
uint32_t gblocks = mp->m_groups[XG_TYPE_RTG].blocks;
fs/xfs/xfs_rtalloc.c
1301
xfs_warn(mp,
fs/xfs/xfs_rtalloc.c
1323
struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1330
*new_rgcount = howmany_64(rblocks, mp->m_sb.sb_rgextents * rextsize);
fs/xfs/xfs_rtalloc.c
1335
if (!mp->m_rtdirip) {
fs/xfs/xfs_rtalloc.c
1338
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_rtalloc.c
1343
error = xfs_rtginode_mkdir_parent(mp);
fs/xfs/xfs_rtalloc.c
1353
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1355
if (!xfs_has_rtgroups(mp))
fs/xfs/xfs_rtalloc.c
1357
if (mp->m_sb.sb_rgcount == 0)
fs/xfs/xfs_rtalloc.c
1359
return xfs_rtgroup_extents(mp, mp->m_sb.sb_rgcount - 1) <
fs/xfs/xfs_rtalloc.c
1360
mp->m_sb.sb_rgextents;
fs/xfs/xfs_rtalloc.c
1368
struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1371
xfs_daddr_t daddr = XFS_FSB_TO_BB(mp, last_block);
fs/xfs/xfs_rtalloc.c
1375
if (XFS_BB_TO_FSB(mp, daddr) != last_block) {
fs/xfs/xfs_rtalloc.c
1376
xfs_warn(mp, "RT device size overflow: %llu != %llu",
fs/xfs/xfs_rtalloc.c
1377
XFS_BB_TO_FSB(mp, daddr), last_block);
fs/xfs/xfs_rtalloc.c
1381
error = xfs_buf_read_uncached(mp->m_rtdev_targp,
fs/xfs/xfs_rtalloc.c
1382
XFS_FSB_TO_BB(mp, mp->m_sb.sb_rtstart) + daddr,
fs/xfs/xfs_rtalloc.c
1383
XFS_FSB_TO_BB(mp, 1), &bp, NULL);
fs/xfs/xfs_rtalloc.c
1385
xfs_warn(mp, "cannot read last RT device sector (%lld)",
fs/xfs/xfs_rtalloc.c
1397
struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1400
xfs_rgnumber_t old_rgcount = mp->m_sb.sb_rgcount;
fs/xfs/xfs_rtalloc.c
1403
xfs_agblock_t old_rextsize = mp->m_sb.sb_rextsize;
fs/xfs/xfs_rtalloc.c
1410
if (!XFS_IS_REALTIME_MOUNT(mp))
fs/xfs/xfs_rtalloc.c
1413
if (!mutex_trylock(&mp->m_growlock))
fs/xfs/xfs_rtalloc.c
1418
if (in->newblocks <= mp->m_sb.sb_rblocks)
fs/xfs/xfs_rtalloc.c
1421
if (mp->m_sb.sb_rblocks > 0 && in->extsize != mp->m_sb.sb_rextsize)
fs/xfs/xfs_rtalloc.c
1425
if (XFS_FSB_TO_B(mp, in->extsize) > XFS_MAX_RTEXTSIZE ||
fs/xfs/xfs_rtalloc.c
1426
XFS_FSB_TO_B(mp, in->extsize) < XFS_MIN_RTEXTSIZE)
fs/xfs/xfs_rtalloc.c
1431
if (!xfs_has_rtgroups(mp)) {
fs/xfs/xfs_rtalloc.c
1432
if (xfs_has_rmapbt(mp))
fs/xfs/xfs_rtalloc.c
1434
if (xfs_has_quota(mp))
fs/xfs/xfs_rtalloc.c
1436
if (xfs_has_reflink(mp))
fs/xfs/xfs_rtalloc.c
1438
} else if (xfs_has_reflink(mp) &&
fs/xfs/xfs_rtalloc.c
1439
!xfs_reflink_supports_rextsize(mp, in->extsize))
fs/xfs/xfs_rtalloc.c
144
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_rtalloc.c
1442
error = xfs_sb_validate_fsb_count(&mp->m_sb, in->newblocks);
fs/xfs/xfs_rtalloc.c
1446
error = xfs_rt_check_size(mp, in->newblocks - 1);
fs/xfs/xfs_rtalloc.c
1458
error = xfs_growfs_check_rtgeom(mp, mp->m_sb.sb_dblocks, in->newblocks,
fs/xfs/xfs_rtalloc.c
1463
if (xfs_has_rtgroups(mp)) {
fs/xfs/xfs_rtalloc.c
1464
error = xfs_growfs_rt_prep_groups(mp, in->newblocks,
fs/xfs/xfs_rtalloc.c
1470
if (xfs_grow_last_rtg(mp)) {
fs/xfs/xfs_rtalloc.c
1471
error = xfs_growfs_rtg(mp, old_rgcount - 1, in->newblocks,
fs/xfs/xfs_rtalloc.c
1480
error = xfs_rtgroup_alloc(mp, rgno, new_rgcount, rextents);
fs/xfs/xfs_rtalloc.c
1484
error = xfs_growfs_rtg(mp, rgno, in->newblocks, in->extsize);
fs/xfs/xfs_rtalloc.c
1488
rtg = xfs_rtgroup_grab(mp, rgno);
fs/xfs/xfs_rtalloc.c
1492
xfs_rtgroup_free(mp, rgno);
fs/xfs/xfs_rtalloc.c
1499
error = xfs_growfs_rt_fixup_extsize(mp);
fs/xfs/xfs_rtalloc.c
1507
if (!xfs_is_shutdown(mp)) {
fs/xfs/xfs_rtalloc.c
1508
int error2 = xfs_update_secondary_sbs(mp);
fs/xfs/xfs_rtalloc.c
1514
error2 = xfs_metafile_resv_init(mp);
fs/xfs/xfs_rtalloc.c
1520
mutex_unlock(&mp->m_growlock);
fs/xfs/xfs_rtalloc.c
1527
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1532
if (!xfs_has_rtsb(mp))
fs/xfs/xfs_rtalloc.c
1534
if (mp->m_sb.sb_rblocks == 0)
fs/xfs/xfs_rtalloc.c
1536
if (mp->m_rtdev_targp == NULL) {
fs/xfs/xfs_rtalloc.c
1537
xfs_warn(mp,
fs/xfs/xfs_rtalloc.c
1543
error = xfs_buf_read_uncached(mp->m_rtdev_targp, XFS_RTSB_DADDR,
fs/xfs/xfs_rtalloc.c
1544
mp->m_sb.sb_blocksize >> BBSHIFT, &bp,
fs/xfs/xfs_rtalloc.c
1547
xfs_warn(mp, "rt sb validate failed with error %d.", error);
fs/xfs/xfs_rtalloc.c
1554
mp->m_rtsb_bp = bp;
fs/xfs/xfs_rtalloc.c
1562
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1564
struct xfs_buf *bp = mp->m_rtsb_bp;
fs/xfs/xfs_rtalloc.c
1570
mp->m_rtsb_bp = NULL;
fs/xfs/xfs_rtalloc.c
1579
struct xfs_mount *mp) /* file system mount structure */
fs/xfs/xfs_rtalloc.c
1581
if (mp->m_sb.sb_rblocks == 0)
fs/xfs/xfs_rtalloc.c
1583
if (mp->m_rtdev_targp == NULL) {
fs/xfs/xfs_rtalloc.c
1584
xfs_warn(mp,
fs/xfs/xfs_rtalloc.c
1589
mp->m_rsumblocks = xfs_rtsummary_blockcount(mp, &mp->m_rsumlevels);
fs/xfs/xfs_rtalloc.c
1591
return xfs_rt_check_size(mp, mp->m_sb.sb_rblocks - 1);
fs/xfs/xfs_rtalloc.c
1613
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1620
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/xfs_rtalloc.c
1631
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_rtalloc.c
1632
mp->m_sb.sb_frextents = val;
fs/xfs/xfs_rtalloc.c
1633
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_rtalloc.c
1634
xfs_set_freecounter(mp, XC_FREE_RTEXTENTS, mp->m_sb.sb_frextents);
fs/xfs/xfs_rtalloc.c
1669
struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
1688
if (xfs_has_zoned(mp))
fs/xfs/xfs_rtalloc.c
1690
return xfs_alloc_rsum_cache(rtg, mp->m_sb.sb_rbmblocks);
fs/xfs/xfs_rtalloc.c
1699
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1705
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_rtalloc.c
1706
if (xfs_has_rtgroups(mp) && mp->m_sb.sb_rgcount > 0) {
fs/xfs/xfs_rtalloc.c
1712
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/xfs_rtalloc.c
1713
error = xfs_rtmount_rtg(mp, tp, rtg);
fs/xfs/xfs_rtalloc.c
1716
xfs_rtunmount_inodes(mp);
fs/xfs/xfs_rtalloc.c
1728
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
1732
while ((rtg = xfs_rtgroup_next(mp, rtg)))
fs/xfs/xfs_rtalloc.c
1734
xfs_rtginode_irele(&mp->m_rtdirip);
fs/xfs/xfs_rtalloc.c
174
xfs_rtx_to_rbmblock(mp, preblock), -1);
fs/xfs/xfs_rtalloc.c
1750
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
1770
b = (mp->m_sb.sb_rextents * ((resid << 1) + 1ULL)) >>
fs/xfs/xfs_rtalloc.c
1772
if (b >= mp->m_sb.sb_rextents)
fs/xfs/xfs_rtalloc.c
1773
div64_u64_rem(b, mp->m_sb.sb_rextents, &b);
fs/xfs/xfs_rtalloc.c
1774
if (b + len > mp->m_sb.sb_rextents)
fs/xfs/xfs_rtalloc.c
1775
b = mp->m_sb.sb_rextents - len;
fs/xfs/xfs_rtalloc.c
1827
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
1830
xfs_extlen_t minlen = xfs_rtxlen_to_extlen(mp, minlen_rtx);
fs/xfs/xfs_rtalloc.c
1831
xfs_extlen_t len = xfs_rtxlen_to_extlen(mp, len_rtx);
fs/xfs/xfs_rtalloc.c
1836
xfs_rtxlen_to_extlen(mp, maxlen_rtx), &rgbno, &len,
fs/xfs/xfs_rtalloc.c
185
xfs_rtx_to_rbmblock(mp, preblock), 1);
fs/xfs/xfs_rtalloc.c
1856
*resrtx = xfs_rgbno_to_rtx(mp, aligned_rgbno);
fs/xfs/xfs_rtalloc.c
1857
*reslen = xfs_extlen_to_rtxlen(mp,
fs/xfs/xfs_rtalloc.c
1860
*resrtx = xfs_rgbno_to_rtx(mp, rgbno);
fs/xfs/xfs_rtalloc.c
1861
*reslen = xfs_extlen_to_rtxlen(mp, len);
fs/xfs/xfs_rtalloc.c
1935
.mp = tp->t_mountp,
fs/xfs/xfs_rtalloc.c
1943
args.rtg = xfs_rtgroup_grab(args.mp, rgno);
fs/xfs/xfs_rtalloc.c
1967
if (!xfs_has_rtgroups(args.mp))
fs/xfs/xfs_rtalloc.c
197
xfs_rtx_to_rbmblock(mp, end + 1), 1);
fs/xfs/xfs_rtalloc.c
1978
start = xfs_rtb_to_rtx(args.mp, bno_hint);
fs/xfs/xfs_rtalloc.c
1979
else if (!xfs_has_rtgroups(args.mp) && initial_user_data)
fs/xfs/xfs_rtalloc.c
2001
if (xfs_has_rtgroups(args.mp))
fs/xfs/xfs_rtalloc.c
2006
if (xfs_has_rtgroups(args.mp)) {
fs/xfs/xfs_rtalloc.c
2023
*blen = xfs_rtxlen_to_extlen(args.mp, len);
fs/xfs/xfs_rtalloc.c
2047
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_rtalloc.c
2058
start_rgno = xfs_rtb_to_rgno(mp, bno_hint);
fs/xfs/xfs_rtalloc.c
2060
start_rgno = (atomic_inc_return(&mp->m_rtgrotor) - 1) %
fs/xfs/xfs_rtalloc.c
2061
mp->m_sb.sb_rgcount;
fs/xfs/xfs_rtalloc.c
2074
if (++rgno == mp->m_sb.sb_rgcount)
fs/xfs/xfs_rtalloc.c
2090
struct xfs_mount *mp = ap->ip->i_mount;
fs/xfs/xfs_rtalloc.c
2092
xfs_extlen_t minlen = mp->m_sb.sb_rextsize;
fs/xfs/xfs_rtalloc.c
2098
align = mp->m_sb.sb_rextsize;
fs/xfs/xfs_rtalloc.c
2106
if (align == mp->m_sb.sb_rextsize)
fs/xfs/xfs_rtalloc.c
2110
error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev, align, 1,
fs/xfs/xfs_rtalloc.c
2115
ASSERT(xfs_extlen_to_rtxmod(mp, ap->length) == 0);
fs/xfs/xfs_rtalloc.c
2135
*ralen = xfs_extlen_to_rtxlen(mp, min(ap->length, XFS_MAX_BMBT_EXTLEN));
fs/xfs/xfs_rtalloc.c
2136
*raminlen = max_t(xfs_rtxlen_t, 1, xfs_extlen_to_rtxlen(mp, minlen));
fs/xfs/xfs_rtalloc.c
2148
*prod = xfs_extlen_to_rtxlen(mp, align);
fs/xfs/xfs_rtalloc.c
253
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_rtalloc.c
267
end = min(args->rtg->rtg_extents, xfs_rbmblock_to_rtx(mp, bbno + 1)) -
fs/xfs/xfs_rtalloc.c
269
for (i = xfs_rbmblock_to_rtx(mp, bbno); i <= end; i++) {
fs/xfs/xfs_rtalloc.c
414
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_rtalloc.c
440
bbno = xfs_rtx_to_rbmblock(mp, start);
fs/xfs/xfs_rtalloc.c
453
error = xfs_rtany_summary(args, log2len, mp->m_rsumlevels - 1,
fs/xfs/xfs_rtalloc.c
496
else if (maxlog < mp->m_blkbit_log)
fs/xfs/xfs_rtalloc.c
499
maxblocks = 2 << (maxlog - mp->m_blkbit_log);
fs/xfs/xfs_rtalloc.c
528
else if (i > 0 && (int)bbno + i < mp->m_sb.sb_rbmblocks - 1)
fs/xfs/xfs_rtalloc.c
534
else if (i <= 0 && (int)bbno - i < mp->m_sb.sb_rbmblocks - 1)
fs/xfs/xfs_rtalloc.c
564
for (i = 0; i < args->mp->m_sb.sb_rbmblocks; i++) {
fs/xfs/xfs_rtalloc.c
590
if (xfs_rtx_to_rbmblock(args->mp, n) > i + 1)
fs/xfs/xfs_rtalloc.c
591
i = xfs_rtx_to_rbmblock(args->mp, n) - 1;
fs/xfs/xfs_rtalloc.c
627
for (l = xfs_highbit32(maxlen); l < args->mp->m_rsumlevels; l++) {
fs/xfs/xfs_rtalloc.c
698
struct xfs_mount *mp)
fs/xfs/xfs_rtalloc.c
700
struct xfs_inode *ip = mp->m_rootip;
fs/xfs/xfs_rtalloc.c
709
error = xfs_trans_alloc_inode(ip, &M_RES(mp)->tr_ichange, 0, 0, false,
fs/xfs/xfs_rtalloc.c
746
const struct xfs_mount *mp,
fs/xfs/xfs_rtalloc.c
752
nmp = kmemdup(mp, sizeof(*mp), GFP_KERNEL);
fs/xfs/xfs_rtalloc.c
782
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
786
if (rgno < mp->m_sb.sb_rgcount)
fs/xfs/xfs_rtalloc.c
787
start_rtx = xfs_rtgroup_extents(mp, rgno);
fs/xfs/xfs_rtalloc.c
788
end_rtx = xfs_rtgroup_extents(nargs->mp, rgno);
fs/xfs/xfs_rtalloc.c
794
if (xfs_has_rtsb(nargs->mp) && rgno == 0 && start_rtx == 0)
fs/xfs/xfs_rtalloc.c
807
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
810
step = (bmbno + 1) * mp->m_rtx_per_rbmblock * rextsize;
fs/xfs/xfs_rtalloc.c
811
if (xfs_has_rtgroups(mp)) {
fs/xfs/xfs_rtalloc.c
812
xfs_rfsblock_t rgblocks = mp->m_sb.sb_rgextents * rextsize;
fs/xfs/xfs_rtalloc.c
831
struct xfs_mount *mp = args->mp;
fs/xfs/xfs_rtalloc.c
835
if (!xfs_has_rtsb(nargs->mp))
fs/xfs/xfs_rtalloc.c
839
if (mp->m_sb.sb_rblocks)
fs/xfs/xfs_rtalloc.c
842
error = xfs_buf_get_uncached(mp->m_rtdev_targp, XFS_FSB_TO_BB(mp, 1),
fs/xfs/xfs_rtalloc.c
850
xfs_update_rtsb(rtsb_bp, mp->m_sb_bp);
fs/xfs/xfs_rtalloc.c
851
mp->m_rtsb_bp = rtsb_bp;
fs/xfs/xfs_rtalloc.c
859
error = xfs_rtrmapbt_init_rtsb(nargs->mp, args->rtg, args->tp);
fs/xfs/xfs_rtalloc.c
869
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_rtalloc.c
871
if (nmp->m_sb.sb_rextsize != mp->m_sb.sb_rextsize)
fs/xfs/xfs_rtalloc.c
873
nmp->m_sb.sb_rextsize - mp->m_sb.sb_rextsize);
fs/xfs/xfs_rtalloc.c
874
if (nmp->m_sb.sb_rbmblocks != mp->m_sb.sb_rbmblocks)
fs/xfs/xfs_rtalloc.c
876
nmp->m_sb.sb_rbmblocks - mp->m_sb.sb_rbmblocks);
fs/xfs/xfs_rtalloc.c
877
if (nmp->m_sb.sb_rblocks != mp->m_sb.sb_rblocks)
fs/xfs/xfs_rtalloc.c
879
nmp->m_sb.sb_rblocks - mp->m_sb.sb_rblocks);
fs/xfs/xfs_rtalloc.c
880
if (nmp->m_sb.sb_rextents != mp->m_sb.sb_rextents)
fs/xfs/xfs_rtalloc.c
882
nmp->m_sb.sb_rextents - mp->m_sb.sb_rextents);
fs/xfs/xfs_rtalloc.c
883
if (nmp->m_sb.sb_rextslog != mp->m_sb.sb_rextslog)
fs/xfs/xfs_rtalloc.c
885
nmp->m_sb.sb_rextslog - mp->m_sb.sb_rextslog);
fs/xfs/xfs_rtalloc.c
886
if (nmp->m_sb.sb_rgcount != mp->m_sb.sb_rgcount)
fs/xfs/xfs_rtalloc.c
888
nmp->m_sb.sb_rgcount - mp->m_sb.sb_rgcount);
fs/xfs/xfs_rtalloc.c
896
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
906
nmp = xfs_growfs_rt_alloc_fake_mount(mp, nrblocks,
fs/xfs/xfs_rtalloc.c
907
mp->m_sb.sb_rextsize);
fs/xfs/xfs_rtalloc.c
910
freed_rtx = nmp->m_sb.sb_rextents - mp->m_sb.sb_rextents;
fs/xfs/xfs_rtalloc.c
915
error = xfs_trans_alloc(mp, &M_RES(nmp)->tr_growrtfree, 0, 0, 0, &tp);
fs/xfs/xfs_rtalloc.c
933
mp->m_features |= XFS_FEAT_REALTIME;
fs/xfs/xfs_rtalloc.c
934
xfs_rtrmapbt_compute_maxlevels(mp);
fs/xfs/xfs_rtalloc.c
935
xfs_rtrefcountbt_compute_maxlevels(mp);
fs/xfs/xfs_rtalloc.c
936
xfs_zoned_add_available(mp, freed_rtx);
fs/xfs/xfs_rtalloc.c
949
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_rtalloc.c
953
.mp = mp,
fs/xfs/xfs_rtalloc.c
967
nmp = nargs.mp = xfs_growfs_rt_alloc_fake_mount(mp,
fs/xfs/xfs_rtalloc.c
981
error = xfs_trans_alloc(mp, &M_RES(nmp)->tr_growrtfree, 0, 0, 0,
fs/xfs/xfs_rtalloc.h
16
int xfs_rtmount_readsb(struct xfs_mount *mp);
fs/xfs/xfs_rtalloc.h
17
void xfs_rtmount_freesb(struct xfs_mount *mp);
fs/xfs/xfs_rtalloc.h
24
struct xfs_mount *mp); /* file system mount structure */
fs/xfs/xfs_rtalloc.h
27
struct xfs_mount *mp);
fs/xfs/xfs_rtalloc.h
35
struct xfs_mount *mp); /* file system mount structure */
fs/xfs/xfs_rtalloc.h
42
struct xfs_mount *mp, /* file system mount structure */
fs/xfs/xfs_rtalloc.h
45
int xfs_rtalloc_reinit_frextents(struct xfs_mount *mp);
fs/xfs/xfs_rtalloc.h
46
int xfs_growfs_check_rtgeom(const struct xfs_mount *mp, xfs_rfsblock_t dblocks,
fs/xfs/xfs_rtalloc.h
49
# define xfs_growfs_rt(mp,in) (-ENOSYS)
fs/xfs/xfs_rtalloc.h
51
# define xfs_rtmount_readsb(mp) (0)
fs/xfs/xfs_rtalloc.h
52
# define xfs_rtmount_freesb(mp) ((void)0)
fs/xfs/xfs_rtalloc.h
55
xfs_mount_t *mp) /* file system mount structure */
fs/xfs/xfs_rtalloc.h
57
if (mp->m_sb.sb_rblocks == 0)
fs/xfs/xfs_rtalloc.h
60
xfs_warn(mp, "Not built with CONFIG_XFS_RT");
fs/xfs/xfs_rtalloc.h
63
# define xfs_rtmount_inodes(m) (((mp)->m_sb.sb_rblocks == 0)? 0 : (-ENOSYS))
fs/xfs/xfs_rtalloc.h
67
xfs_growfs_check_rtgeom(const struct xfs_mount *mp,
fs/xfs/xfs_stats.h
175
#define XFS_STATS_INC(mp, v) \
fs/xfs/xfs_stats.h
178
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v++; \
fs/xfs/xfs_stats.h
181
#define XFS_STATS_DEC(mp, v) \
fs/xfs/xfs_stats.h
184
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v--; \
fs/xfs/xfs_stats.h
187
#define XFS_STATS_ADD(mp, v, inc) \
fs/xfs/xfs_stats.h
190
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->s.v += (inc); \
fs/xfs/xfs_stats.h
193
#define XFS_STATS_INC_OFF(mp, off) \
fs/xfs/xfs_stats.h
196
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]++; \
fs/xfs/xfs_stats.h
199
#define XFS_STATS_DEC_OFF(mp, off) \
fs/xfs/xfs_stats.h
202
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off]; \
fs/xfs/xfs_stats.h
205
#define XFS_STATS_ADD_OFF(mp, off, inc) \
fs/xfs/xfs_stats.h
208
per_cpu_ptr(mp->m_stats.xs_stats, current_cpu())->a[off] += (inc); \
fs/xfs/xfs_super.c
1005
xfs_save_resvblks(mp);
fs/xfs/xfs_super.c
1006
ret = xfs_log_quiesce(mp);
fs/xfs/xfs_super.c
1016
if (ret && !xfs_is_readonly(mp)) {
fs/xfs/xfs_super.c
1017
xfs_blockgc_start(mp);
fs/xfs/xfs_super.c
1018
xfs_inodegc_start(mp);
fs/xfs/xfs_super.c
1019
xfs_zone_gc_start(mp);
fs/xfs/xfs_super.c
1029
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
1031
xfs_restore_resvblks(mp);
fs/xfs/xfs_super.c
1032
xfs_log_work_queue(mp);
fs/xfs/xfs_super.c
1040
if (!xfs_is_readonly(mp)) {
fs/xfs/xfs_super.c
1041
xfs_zone_gc_start(mp);
fs/xfs/xfs_super.c
1042
xfs_blockgc_start(mp);
fs/xfs/xfs_super.c
1043
xfs_inodegc_start(mp);
fs/xfs/xfs_super.c
1055
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1058
if (xfs_has_logv2(mp)) {
fs/xfs/xfs_super.c
1059
if (mp->m_logbsize <= 0 &&
fs/xfs/xfs_super.c
1060
mp->m_sb.sb_logsunit > XLOG_BIG_RECORD_BSIZE) {
fs/xfs/xfs_super.c
1061
mp->m_logbsize = mp->m_sb.sb_logsunit;
fs/xfs/xfs_super.c
1062
} else if (mp->m_logbsize > 0 &&
fs/xfs/xfs_super.c
1063
mp->m_logbsize < mp->m_sb.sb_logsunit) {
fs/xfs/xfs_super.c
1064
xfs_warn(mp,
fs/xfs/xfs_super.c
1070
if (mp->m_logbsize > XLOG_BIG_RECORD_BSIZE) {
fs/xfs/xfs_super.c
1071
xfs_warn(mp,
fs/xfs/xfs_super.c
1080
if ((mp->m_sb.sb_flags & XFS_SBF_READONLY) && !xfs_is_readonly(mp)) {
fs/xfs/xfs_super.c
1081
xfs_warn(mp,
fs/xfs/xfs_super.c
1086
if ((mp->m_qflags & XFS_GQUOTA_ACCT) &&
fs/xfs/xfs_super.c
1087
(mp->m_qflags & XFS_PQUOTA_ACCT) &&
fs/xfs/xfs_super.c
1088
!xfs_has_pquotino(mp)) {
fs/xfs/xfs_super.c
1089
xfs_warn(mp,
fs/xfs/xfs_super.c
1094
if (!xfs_has_zoned(mp)) {
fs/xfs/xfs_super.c
1095
if (mp->m_max_open_zones) {
fs/xfs/xfs_super.c
1096
xfs_warn(mp,
fs/xfs/xfs_super.c
1100
if (mp->m_features & XFS_FEAT_NOLIFETIME) {
fs/xfs/xfs_super.c
1101
xfs_warn(mp,
fs/xfs/xfs_super.c
1112
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1117
error = percpu_counter_init(&mp->m_icount, 0, GFP_KERNEL);
fs/xfs/xfs_super.c
1121
error = percpu_counter_init(&mp->m_ifree, 0, GFP_KERNEL);
fs/xfs/xfs_super.c
1125
error = percpu_counter_init(&mp->m_delalloc_blks, 0, GFP_KERNEL);
fs/xfs/xfs_super.c
1129
error = percpu_counter_init(&mp->m_delalloc_rtextents, 0, GFP_KERNEL);
fs/xfs/xfs_super.c
1134
error = percpu_counter_init(&mp->m_free[i].count, 0,
fs/xfs/xfs_super.c
1144
percpu_counter_destroy(&mp->m_free[i].count);
fs/xfs/xfs_super.c
1145
percpu_counter_destroy(&mp->m_delalloc_rtextents);
fs/xfs/xfs_super.c
1147
percpu_counter_destroy(&mp->m_delalloc_blks);
fs/xfs/xfs_super.c
1149
percpu_counter_destroy(&mp->m_ifree);
fs/xfs/xfs_super.c
1151
percpu_counter_destroy(&mp->m_icount);
fs/xfs/xfs_super.c
1157
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1159
percpu_counter_set(&mp->m_icount, mp->m_sb.sb_icount);
fs/xfs/xfs_super.c
1160
percpu_counter_set(&mp->m_ifree, mp->m_sb.sb_ifree);
fs/xfs/xfs_super.c
1161
xfs_set_freecounter(mp, XC_FREE_BLOCKS, mp->m_sb.sb_fdblocks);
fs/xfs/xfs_super.c
1162
if (!xfs_has_zoned(mp))
fs/xfs/xfs_super.c
1163
xfs_set_freecounter(mp, XC_FREE_RTEXTENTS,
fs/xfs/xfs_super.c
1164
mp->m_sb.sb_frextents);
fs/xfs/xfs_super.c
1169
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1174
percpu_counter_destroy(&mp->m_free[i].count);
fs/xfs/xfs_super.c
1175
percpu_counter_destroy(&mp->m_icount);
fs/xfs/xfs_super.c
1176
percpu_counter_destroy(&mp->m_ifree);
fs/xfs/xfs_super.c
1177
ASSERT(xfs_is_shutdown(mp) ||
fs/xfs/xfs_super.c
1178
percpu_counter_sum(&mp->m_delalloc_rtextents) == 0);
fs/xfs/xfs_super.c
1179
percpu_counter_destroy(&mp->m_delalloc_rtextents);
fs/xfs/xfs_super.c
1180
ASSERT(xfs_is_shutdown(mp) ||
fs/xfs/xfs_super.c
1181
percpu_counter_sum(&mp->m_delalloc_blks) == 0);
fs/xfs/xfs_super.c
1182
percpu_counter_destroy(&mp->m_delalloc_blks);
fs/xfs/xfs_super.c
1187
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1192
mp->m_inodegc = alloc_percpu(struct xfs_inodegc);
fs/xfs/xfs_super.c
1193
if (!mp->m_inodegc)
fs/xfs/xfs_super.c
1197
gc = per_cpu_ptr(mp->m_inodegc, cpu);
fs/xfs/xfs_super.c
1199
gc->mp = mp;
fs/xfs/xfs_super.c
1210
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1212
if (!mp->m_inodegc)
fs/xfs/xfs_super.c
1214
free_percpu(mp->m_inodegc);
fs/xfs/xfs_super.c
1221
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
1223
xfs_notice(mp, "Unmounting Filesystem %pU", &mp->m_sb.sb_uuid);
fs/xfs/xfs_super.c
1224
xfs_filestream_unmount(mp);
fs/xfs/xfs_super.c
1225
xfs_unmountfs(mp);
fs/xfs/xfs_super.c
1227
xfs_rtmount_freesb(mp);
fs/xfs/xfs_super.c
1228
xfs_freesb(mp);
fs/xfs/xfs_super.c
1229
xchk_mount_stats_free(mp);
fs/xfs/xfs_super.c
1230
free_percpu(mp->m_stats.xs_stats);
fs/xfs/xfs_super.c
1231
xfs_inodegc_free_percpu(mp);
fs/xfs/xfs_super.c
1232
xfs_destroy_percpu_counters(mp);
fs/xfs/xfs_super.c
1233
xfs_destroy_mount_workqueues(mp);
fs/xfs/xfs_super.c
1234
xfs_shutdown_devices(mp);
fs/xfs/xfs_super.c
1268
struct xfs_mount *mp = XFS_M(root->d_sb);
fs/xfs/xfs_super.c
1270
if (xfs_has_zoned(mp) && IS_ENABLED(CONFIG_XFS_RT))
fs/xfs/xfs_super.c
1271
xfs_zoned_show_stats(m, mp);
fs/xfs/xfs_super.c
1572
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1575
if (xfs_has_norecovery(mp) && !xfs_is_readonly(mp)) {
fs/xfs/xfs_super.c
1576
xfs_warn(mp, "no-recovery mounts must be read-only.");
fs/xfs/xfs_super.c
1580
if (xfs_has_noalign(mp) && (mp->m_dalign || mp->m_swidth)) {
fs/xfs/xfs_super.c
1581
xfs_warn(mp,
fs/xfs/xfs_super.c
1587
(mp->m_qflags & ~XFS_QFLAGS_MNTOPTS)) {
fs/xfs/xfs_super.c
1588
xfs_warn(mp, "quota support not available in this kernel.");
fs/xfs/xfs_super.c
1592
if ((mp->m_dalign && !mp->m_swidth) ||
fs/xfs/xfs_super.c
1593
(!mp->m_dalign && mp->m_swidth)) {
fs/xfs/xfs_super.c
1594
xfs_warn(mp, "sunit and swidth must be specified together");
fs/xfs/xfs_super.c
1598
if (mp->m_dalign && (mp->m_swidth % mp->m_dalign != 0)) {
fs/xfs/xfs_super.c
1599
xfs_warn(mp,
fs/xfs/xfs_super.c
1601
mp->m_swidth, mp->m_dalign);
fs/xfs/xfs_super.c
1605
if (mp->m_logbufs != -1 &&
fs/xfs/xfs_super.c
1606
mp->m_logbufs != 0 &&
fs/xfs/xfs_super.c
1607
(mp->m_logbufs < XLOG_MIN_ICLOGS ||
fs/xfs/xfs_super.c
1608
mp->m_logbufs > XLOG_MAX_ICLOGS)) {
fs/xfs/xfs_super.c
1609
xfs_warn(mp, "invalid logbufs value: %d [not %d-%d]",
fs/xfs/xfs_super.c
1610
mp->m_logbufs, XLOG_MIN_ICLOGS, XLOG_MAX_ICLOGS);
fs/xfs/xfs_super.c
1614
if (mp->m_logbsize != -1 &&
fs/xfs/xfs_super.c
1615
mp->m_logbsize != 0 &&
fs/xfs/xfs_super.c
1616
(mp->m_logbsize < XLOG_MIN_RECORD_BSIZE ||
fs/xfs/xfs_super.c
1617
mp->m_logbsize > XLOG_MAX_RECORD_BSIZE ||
fs/xfs/xfs_super.c
1618
!is_power_of_2(mp->m_logbsize))) {
fs/xfs/xfs_super.c
1619
xfs_warn(mp,
fs/xfs/xfs_super.c
1621
mp->m_logbsize);
fs/xfs/xfs_super.c
1625
if (xfs_has_allocsize(mp) &&
fs/xfs/xfs_super.c
1626
(mp->m_allocsize_log > XFS_MAX_IO_LOG ||
fs/xfs/xfs_super.c
1627
mp->m_allocsize_log < XFS_MIN_IO_LOG)) {
fs/xfs/xfs_super.c
1628
xfs_warn(mp, "invalid log iosize: %d [not %d-%d]",
fs/xfs/xfs_super.c
1629
mp->m_allocsize_log, XFS_MIN_IO_LOG, XFS_MAX_IO_LOG);
fs/xfs/xfs_super.c
1656
struct xfs_mount *mp = sb->s_fs_info;
fs/xfs/xfs_super.c
1660
mp->m_super = sb;
fs/xfs/xfs_super.c
1668
xfs_set_readonly(mp);
fs/xfs/xfs_super.c
1670
mp->m_features |= XFS_FEAT_DIRSYNC;
fs/xfs/xfs_super.c
1672
mp->m_features |= XFS_FEAT_WSYNC;
fs/xfs/xfs_super.c
1674
error = xfs_fs_validate_params(mp);
fs/xfs/xfs_super.c
1679
xfs_err(mp, "unable to set blocksize");
fs/xfs/xfs_super.c
1696
xfs_notice(mp, "Delaying mount for %d seconds.",
fs/xfs/xfs_super.c
1704
error = xfs_open_devices(mp);
fs/xfs/xfs_super.c
1709
mp->m_debugfs = xfs_debugfs_mkdir(mp->m_super->s_id,
fs/xfs/xfs_super.c
1712
mp->m_debugfs = NULL;
fs/xfs/xfs_super.c
1715
error = xfs_init_mount_workqueues(mp);
fs/xfs/xfs_super.c
1719
error = xfs_init_percpu_counters(mp);
fs/xfs/xfs_super.c
1723
error = xfs_inodegc_init_percpu(mp);
fs/xfs/xfs_super.c
1728
mp->m_stats.xs_stats = alloc_percpu(struct xfsstats);
fs/xfs/xfs_super.c
1729
if (!mp->m_stats.xs_stats) {
fs/xfs/xfs_super.c
1734
error = xchk_mount_stats_alloc(mp);
fs/xfs/xfs_super.c
1738
error = xfs_readsb(mp, flags);
fs/xfs/xfs_super.c
1742
error = xfs_finish_flags(mp);
fs/xfs/xfs_super.c
1746
error = xfs_setup_devices(mp);
fs/xfs/xfs_super.c
1756
if (!(mp->m_features & XFS_FEAT_CRC)) {
fs/xfs/xfs_super.c
1758
xfs_warn(mp,
fs/xfs/xfs_super.c
1763
xfs_warn_once(mp,
fs/xfs/xfs_super.c
1768
if (xfs_has_asciici(mp)) {
fs/xfs/xfs_super.c
1770
xfs_warn_once(mp,
fs/xfs/xfs_super.c
1773
xfs_warn(mp,
fs/xfs/xfs_super.c
1785
if (xfs_has_needsrepair(mp) && !xfs_has_norecovery(mp)) {
fs/xfs/xfs_super.c
1786
xfs_warn(mp, "Filesystem needs repair. Please run xfs_repair.");
fs/xfs/xfs_super.c
1796
if (mp->m_sb.sb_inprogress) {
fs/xfs/xfs_super.c
1797
xfs_warn(mp, "Offline file system operation in progress!");
fs/xfs/xfs_super.c
1802
if (mp->m_sb.sb_blocksize > PAGE_SIZE) {
fs/xfs/xfs_super.c
1805
if (!xfs_has_crc(mp)) {
fs/xfs/xfs_super.c
1806
xfs_warn(mp,
fs/xfs/xfs_super.c
1808
mp->m_sb.sb_blocksize, PAGE_SIZE);
fs/xfs/xfs_super.c
1813
if (mp->m_sb.sb_blocksize > max_folio_size) {
fs/xfs/xfs_super.c
1814
xfs_warn(mp,
fs/xfs/xfs_super.c
1816
mp->m_sb.sb_blocksize, max_folio_size);
fs/xfs/xfs_super.c
1823
if (xfs_sb_validate_fsb_count(&mp->m_sb, mp->m_sb.sb_dblocks) ||
fs/xfs/xfs_super.c
1824
xfs_sb_validate_fsb_count(&mp->m_sb, mp->m_sb.sb_rblocks)) {
fs/xfs/xfs_super.c
1825
xfs_warn(mp,
fs/xfs/xfs_super.c
1842
if (!xfs_verify_fileoff(mp, XFS_B_TO_FSBT(mp, MAX_LFS_FILESIZE))) {
fs/xfs/xfs_super.c
1843
xfs_warn(mp,
fs/xfs/xfs_super.c
1845
XFS_B_TO_FSBT(mp, MAX_LFS_FILESIZE),
fs/xfs/xfs_super.c
1851
error = xfs_rtmount_readsb(mp);
fs/xfs/xfs_super.c
1855
error = xfs_filestream_mount(mp);
fs/xfs/xfs_super.c
1864
sb->s_blocksize = mp->m_sb.sb_blocksize;
fs/xfs/xfs_super.c
1869
if (xfs_has_bigtime(mp)) {
fs/xfs/xfs_super.c
1876
trace_xfs_inode_timestamp_range(mp, sb->s_time_min, sb->s_time_max);
fs/xfs/xfs_super.c
1882
if (xfs_has_crc(mp))
fs/xfs/xfs_super.c
1885
if (xfs_has_dax_always(mp)) {
fs/xfs/xfs_super.c
1886
error = xfs_setup_dax_always(mp);
fs/xfs/xfs_super.c
1891
if (xfs_has_discard(mp) && !bdev_max_discard_sectors(sb->s_bdev)) {
fs/xfs/xfs_super.c
1892
xfs_warn(mp,
fs/xfs/xfs_super.c
1894
mp->m_features &= ~XFS_FEAT_DISCARD;
fs/xfs/xfs_super.c
1897
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_super.c
1898
if (!xfs_has_metadir(mp)) {
fs/xfs/xfs_super.c
1899
xfs_alert(mp,
fs/xfs/xfs_super.c
1904
xfs_warn_experimental(mp, XFS_EXPERIMENTAL_ZONED);
fs/xfs/xfs_super.c
1907
if (xfs_has_reflink(mp)) {
fs/xfs/xfs_super.c
1908
if (xfs_has_realtime(mp) &&
fs/xfs/xfs_super.c
1909
!xfs_reflink_supports_rextsize(mp, mp->m_sb.sb_rextsize)) {
fs/xfs/xfs_super.c
1910
xfs_alert(mp,
fs/xfs/xfs_super.c
1912
mp->m_sb.sb_rextsize);
fs/xfs/xfs_super.c
1917
if (xfs_has_zoned(mp)) {
fs/xfs/xfs_super.c
1918
xfs_alert(mp,
fs/xfs/xfs_super.c
1925
xfs_info(mp, "using DEBUG-only always_cow mode.");
fs/xfs/xfs_super.c
1926
mp->m_always_cow = true;
fs/xfs/xfs_super.c
1934
if (!(mp->m_qflags & XFS_QFLAGS_MNTOPTS))
fs/xfs/xfs_super.c
1935
xfs_set_resuming_quotaon(mp);
fs/xfs/xfs_super.c
1936
mp->m_qflags &= ~XFS_QFLAGS_MNTOPTS;
fs/xfs/xfs_super.c
1938
error = xfs_mountfs(mp);
fs/xfs/xfs_super.c
1942
root = igrab(VFS_I(mp->m_rootip));
fs/xfs/xfs_super.c
1956
xfs_filestream_unmount(mp);
fs/xfs/xfs_super.c
1958
xfs_rtmount_freesb(mp);
fs/xfs/xfs_super.c
1960
xfs_freesb(mp);
fs/xfs/xfs_super.c
1962
xchk_mount_stats_free(mp);
fs/xfs/xfs_super.c
1964
free_percpu(mp->m_stats.xs_stats);
fs/xfs/xfs_super.c
1966
xfs_inodegc_free_percpu(mp);
fs/xfs/xfs_super.c
1968
xfs_destroy_percpu_counters(mp);
fs/xfs/xfs_super.c
1970
xfs_destroy_mount_workqueues(mp);
fs/xfs/xfs_super.c
1972
xfs_shutdown_devices(mp);
fs/xfs/xfs_super.c
1976
xfs_filestream_unmount(mp);
fs/xfs/xfs_super.c
1977
xfs_unmountfs(mp);
fs/xfs/xfs_super.c
1990
struct xfs_mount *mp)
fs/xfs/xfs_super.c
1992
struct xfs_sb *sbp = &mp->m_sb;
fs/xfs/xfs_super.c
1995
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp &&
fs/xfs/xfs_super.c
1996
xfs_readonly_buftarg(mp->m_logdev_targp)) {
fs/xfs/xfs_super.c
1997
xfs_warn(mp,
fs/xfs/xfs_super.c
2002
if (mp->m_rtdev_targp && xfs_readonly_buftarg(mp->m_rtdev_targp)) {
fs/xfs/xfs_super.c
2003
xfs_warn(mp,
fs/xfs/xfs_super.c
2008
if (xfs_has_norecovery(mp)) {
fs/xfs/xfs_super.c
2009
xfs_warn(mp,
fs/xfs/xfs_super.c
2016
xfs_warn(mp,
fs/xfs/xfs_super.c
2023
xfs_clear_readonly(mp);
fs/xfs/xfs_super.c
2029
if (mp->m_update_sb) {
fs/xfs/xfs_super.c
2030
error = xfs_sync_sb(mp, false);
fs/xfs/xfs_super.c
2032
xfs_warn(mp, "failed to write sb changes");
fs/xfs/xfs_super.c
2035
mp->m_update_sb = false;
fs/xfs/xfs_super.c
2042
xfs_restore_resvblks(mp);
fs/xfs/xfs_super.c
2043
xfs_log_work_queue(mp);
fs/xfs/xfs_super.c
2044
xfs_blockgc_start(mp);
fs/xfs/xfs_super.c
2047
error = xfs_fs_reserve_ag_blocks(mp);
fs/xfs/xfs_super.c
2052
xfs_inodegc_start(mp);
fs/xfs/xfs_super.c
2055
xfs_zone_gc_start(mp);
fs/xfs/xfs_super.c
2062
struct xfs_mount *mp)
fs/xfs/xfs_super.c
207
struct xfs_mount *mp = XFS_M(root->d_sb);
fs/xfs/xfs_super.c
2070
error = sync_filesystem(mp->m_super);
fs/xfs/xfs_super.c
2078
xfs_blockgc_stop(mp);
fs/xfs/xfs_super.c
2086
error = xfs_blockgc_free_space(mp, &icw);
fs/xfs/xfs_super.c
2088
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_super.c
2099
xfs_inodegc_stop(mp);
fs/xfs/xfs_super.c
2102
xfs_zone_gc_stop(mp);
fs/xfs/xfs_super.c
2105
xfs_fs_unreserve_ag_blocks(mp);
fs/xfs/xfs_super.c
211
if (mp->m_features & xfs_infop->flag)
fs/xfs/xfs_super.c
2114
xfs_save_resvblks(mp);
fs/xfs/xfs_super.c
2116
xfs_log_clean(mp);
fs/xfs/xfs_super.c
2117
xfs_set_readonly(mp);
fs/xfs/xfs_super.c
2138
struct xfs_mount *mp = XFS_M(fc->root->d_sb);
fs/xfs/xfs_super.c
2146
if (xfs_has_crc(mp))
fs/xfs/xfs_super.c
215
seq_printf(m, ",inode%d", xfs_has_small_inums(mp) ? 32 : 64);
fs/xfs/xfs_super.c
2153
xfs_errortag_copy(mp, new_mp);
fs/xfs/xfs_super.c
2156
if (mp->m_awu_max_bytes != new_mp->m_awu_max_bytes) {
fs/xfs/xfs_super.c
2157
error = xfs_set_max_atomic_write_opt(mp,
fs/xfs/xfs_super.c
2164
if (xfs_has_small_inums(mp) && !xfs_has_small_inums(new_mp)) {
fs/xfs/xfs_super.c
2165
mp->m_features &= ~XFS_FEAT_SMALL_INUMS;
fs/xfs/xfs_super.c
2166
mp->m_maxagi = xfs_set_inode_alloc(mp, mp->m_sb.sb_agcount);
fs/xfs/xfs_super.c
217
if (xfs_has_allocsize(mp))
fs/xfs/xfs_super.c
2170
if (!xfs_has_small_inums(mp) && xfs_has_small_inums(new_mp)) {
fs/xfs/xfs_super.c
2171
mp->m_features |= XFS_FEAT_SMALL_INUMS;
fs/xfs/xfs_super.c
2172
mp->m_maxagi = xfs_set_inode_alloc(mp, mp->m_sb.sb_agcount);
fs/xfs/xfs_super.c
2182
error = xfs_finish_flags(mp);
fs/xfs/xfs_super.c
2187
if (xfs_is_readonly(mp) && !(flags & SB_RDONLY)) {
fs/xfs/xfs_super.c
2188
error = xfs_remount_rw(mp);
fs/xfs/xfs_super.c
219
(1 << mp->m_allocsize_log) >> 10);
fs/xfs/xfs_super.c
2194
if (!xfs_is_readonly(mp) && (flags & SB_RDONLY)) {
fs/xfs/xfs_super.c
2195
error = xfs_remount_ro(mp);
fs/xfs/xfs_super.c
2207
struct xfs_mount *mp = fc->s_fs_info;
fs/xfs/xfs_super.c
221
if (mp->m_logbufs > 0)
fs/xfs/xfs_super.c
2215
if (mp)
fs/xfs/xfs_super.c
2216
xfs_mount_free(mp);
fs/xfs/xfs_super.c
222
seq_printf(m, ",logbufs=%d", mp->m_logbufs);
fs/xfs/xfs_super.c
223
if (mp->m_logbsize > 0)
fs/xfs/xfs_super.c
2235
struct xfs_mount *mp;
fs/xfs/xfs_super.c
2238
mp = kzalloc_obj(struct xfs_mount);
fs/xfs/xfs_super.c
2239
if (!mp)
fs/xfs/xfs_super.c
224
seq_printf(m, ",logbsize=%dk", mp->m_logbsize >> 10);
fs/xfs/xfs_super.c
2242
mp->m_errortag = kzalloc_objs(*mp->m_errortag, XFS_ERRTAG_MAX);
fs/xfs/xfs_super.c
2243
if (!mp->m_errortag) {
fs/xfs/xfs_super.c
2244
kfree(mp);
fs/xfs/xfs_super.c
2249
spin_lock_init(&mp->m_sb_lock);
fs/xfs/xfs_super.c
2251
xa_init(&mp->m_groups[i].xa);
fs/xfs/xfs_super.c
2252
mutex_init(&mp->m_growlock);
fs/xfs/xfs_super.c
2253
mutex_init(&mp->m_metafile_resv_lock);
fs/xfs/xfs_super.c
2254
INIT_WORK(&mp->m_flush_inodes_work, xfs_flush_inodes_worker);
fs/xfs/xfs_super.c
2255
INIT_DELAYED_WORK(&mp->m_reclaim_work, xfs_reclaim_worker);
fs/xfs/xfs_super.c
2256
mp->m_kobj.kobject.kset = xfs_kset;
fs/xfs/xfs_super.c
226
if (mp->m_logname)
fs/xfs/xfs_super.c
2263
mp->m_finobt_nores = true;
fs/xfs/xfs_super.c
2268
mp->m_logbufs = -1;
fs/xfs/xfs_super.c
2269
mp->m_logbsize = -1;
fs/xfs/xfs_super.c
227
seq_show_option(m, "logdev", mp->m_logname);
fs/xfs/xfs_super.c
2270
mp->m_allocsize_log = 16; /* 64k */
fs/xfs/xfs_super.c
2272
xfs_hooks_init(&mp->m_dir_update_hooks);
fs/xfs/xfs_super.c
2274
fc->s_fs_info = mp;
fs/xfs/xfs_super.c
228
if (mp->m_rtname)
fs/xfs/xfs_super.c
229
seq_show_option(m, "rtdev", mp->m_rtname);
fs/xfs/xfs_super.c
231
if (mp->m_dalign > 0)
fs/xfs/xfs_super.c
233
(int)XFS_FSB_TO_BB(mp, mp->m_dalign));
fs/xfs/xfs_super.c
234
if (mp->m_swidth > 0)
fs/xfs/xfs_super.c
236
(int)XFS_FSB_TO_BB(mp, mp->m_swidth));
fs/xfs/xfs_super.c
238
if (mp->m_qflags & XFS_UQUOTA_ENFD)
fs/xfs/xfs_super.c
240
else if (mp->m_qflags & XFS_UQUOTA_ACCT)
fs/xfs/xfs_super.c
243
if (mp->m_qflags & XFS_PQUOTA_ENFD)
fs/xfs/xfs_super.c
245
else if (mp->m_qflags & XFS_PQUOTA_ACCT)
fs/xfs/xfs_super.c
248
if (mp->m_qflags & XFS_GQUOTA_ENFD)
fs/xfs/xfs_super.c
250
else if (mp->m_qflags & XFS_GQUOTA_ACCT)
fs/xfs/xfs_super.c
253
if (!(mp->m_qflags & XFS_ALL_QUOTA_ACCT))
fs/xfs/xfs_super.c
256
if (mp->m_max_open_zones)
fs/xfs/xfs_super.c
257
seq_printf(m, ",max_open_zones=%u", mp->m_max_open_zones);
fs/xfs/xfs_super.c
258
if (mp->m_awu_max_bytes)
fs/xfs/xfs_super.c
260
mp->m_awu_max_bytes >> 10);
fs/xfs/xfs_super.c
308
struct xfs_mount *mp,
fs/xfs/xfs_super.c
313
xfs_sb_t *sbp = &mp->m_sb;
fs/xfs/xfs_super.c
322
if (M_IGEO(mp)->maxicount) {
fs/xfs/xfs_super.c
335
agino = XFS_AGB_TO_AGINO(mp, sbp->sb_agblocks - 1);
fs/xfs/xfs_super.c
336
ino = XFS_AGINO_TO_INO(mp, agcount - 1, agino);
fs/xfs/xfs_super.c
343
if (xfs_has_small_inums(mp) && ino > XFS_MAXINUMBER_32)
fs/xfs/xfs_super.c
344
xfs_set_inode32(mp);
fs/xfs/xfs_super.c
346
xfs_clear_inode32(mp);
fs/xfs/xfs_super.c
351
ino = XFS_AGINO_TO_INO(mp, index, agino);
fs/xfs/xfs_super.c
353
pag = xfs_perag_get(mp, index);
fs/xfs/xfs_super.c
359
return xfs_is_inode32(mp) ? maxagi : agcount;
fs/xfs/xfs_super.c
364
struct xfs_mount *mp)
fs/xfs/xfs_super.c
366
if (!mp->m_ddev_targp->bt_daxdev &&
fs/xfs/xfs_super.c
367
(!mp->m_rtdev_targp || !mp->m_rtdev_targp->bt_daxdev)) {
fs/xfs/xfs_super.c
368
xfs_alert(mp,
fs/xfs/xfs_super.c
373
if (mp->m_super->s_blocksize != PAGE_SIZE) {
fs/xfs/xfs_super.c
374
xfs_alert(mp,
fs/xfs/xfs_super.c
379
if (xfs_has_reflink(mp) &&
fs/xfs/xfs_super.c
380
bdev_is_partition(mp->m_ddev_targp->bt_bdev)) {
fs/xfs/xfs_super.c
381
xfs_alert(mp,
fs/xfs/xfs_super.c
389
xfs_mount_set_dax_mode(mp, XFS_DAX_NEVER);
fs/xfs/xfs_super.c
395
xfs_mount_t *mp,
fs/xfs/xfs_super.c
402
mode = sb_open_mode(mp->m_super->s_flags);
fs/xfs/xfs_super.c
404
mp->m_super, &fs_holder_ops);
fs/xfs/xfs_super.c
408
xfs_warn(mp, "Invalid device [%s], error=%d", name, error);
fs/xfs/xfs_super.c
416
struct xfs_mount *mp)
fs/xfs/xfs_super.c
444
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp) {
fs/xfs/xfs_super.c
445
blkdev_issue_flush(mp->m_logdev_targp->bt_bdev);
fs/xfs/xfs_super.c
446
invalidate_bdev(mp->m_logdev_targp->bt_bdev);
fs/xfs/xfs_super.c
448
if (mp->m_rtdev_targp) {
fs/xfs/xfs_super.c
449
blkdev_issue_flush(mp->m_rtdev_targp->bt_bdev);
fs/xfs/xfs_super.c
450
invalidate_bdev(mp->m_rtdev_targp->bt_bdev);
fs/xfs/xfs_super.c
452
blkdev_issue_flush(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_super.c
453
invalidate_bdev(mp->m_ddev_targp->bt_bdev);
fs/xfs/xfs_super.c
468
struct xfs_mount *mp)
fs/xfs/xfs_super.c
470
struct super_block *sb = mp->m_super;
fs/xfs/xfs_super.c
478
if (mp->m_logname) {
fs/xfs/xfs_super.c
479
error = xfs_blkdev_get(mp, mp->m_logname, &logdev_file);
fs/xfs/xfs_super.c
484
if (mp->m_rtname) {
fs/xfs/xfs_super.c
485
error = xfs_blkdev_get(mp, mp->m_rtname, &rtdev_file);
fs/xfs/xfs_super.c
492
xfs_warn(mp,
fs/xfs/xfs_super.c
502
mp->m_ddev_targp = xfs_alloc_buftarg(mp, sb->s_bdev_file);
fs/xfs/xfs_super.c
503
if (IS_ERR(mp->m_ddev_targp)) {
fs/xfs/xfs_super.c
504
error = PTR_ERR(mp->m_ddev_targp);
fs/xfs/xfs_super.c
505
mp->m_ddev_targp = NULL;
fs/xfs/xfs_super.c
510
mp->m_rtdev_targp = xfs_alloc_buftarg(mp, rtdev_file);
fs/xfs/xfs_super.c
511
if (IS_ERR(mp->m_rtdev_targp)) {
fs/xfs/xfs_super.c
512
error = PTR_ERR(mp->m_rtdev_targp);
fs/xfs/xfs_super.c
513
mp->m_rtdev_targp = NULL;
fs/xfs/xfs_super.c
519
mp->m_logdev_targp = xfs_alloc_buftarg(mp, logdev_file);
fs/xfs/xfs_super.c
520
if (IS_ERR(mp->m_logdev_targp)) {
fs/xfs/xfs_super.c
521
error = PTR_ERR(mp->m_logdev_targp);
fs/xfs/xfs_super.c
522
mp->m_logdev_targp = NULL;
fs/xfs/xfs_super.c
526
mp->m_logdev_targp = mp->m_ddev_targp;
fs/xfs/xfs_super.c
535
if (mp->m_rtdev_targp)
fs/xfs/xfs_super.c
536
xfs_free_buftarg(mp->m_rtdev_targp);
fs/xfs/xfs_super.c
538
xfs_free_buftarg(mp->m_ddev_targp);
fs/xfs/xfs_super.c
553
struct xfs_mount *mp)
fs/xfs/xfs_super.c
557
error = xfs_configure_buftarg(mp->m_ddev_targp, mp->m_sb.sb_sectsize,
fs/xfs/xfs_super.c
558
mp->m_sb.sb_dblocks);
fs/xfs/xfs_super.c
562
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp) {
fs/xfs/xfs_super.c
565
if (xfs_has_sector(mp))
fs/xfs/xfs_super.c
566
log_sector_size = mp->m_sb.sb_logsectsize;
fs/xfs/xfs_super.c
567
error = xfs_configure_buftarg(mp->m_logdev_targp,
fs/xfs/xfs_super.c
568
log_sector_size, mp->m_sb.sb_logblocks);
fs/xfs/xfs_super.c
573
if (mp->m_sb.sb_rtstart) {
fs/xfs/xfs_super.c
574
if (mp->m_rtdev_targp) {
fs/xfs/xfs_super.c
575
xfs_warn(mp,
fs/xfs/xfs_super.c
579
mp->m_rtdev_targp = mp->m_ddev_targp;
fs/xfs/xfs_super.c
580
} else if (mp->m_rtname) {
fs/xfs/xfs_super.c
581
error = xfs_configure_buftarg(mp->m_rtdev_targp,
fs/xfs/xfs_super.c
582
mp->m_sb.sb_sectsize, mp->m_sb.sb_rblocks);
fs/xfs/xfs_super.c
592
struct xfs_mount *mp)
fs/xfs/xfs_super.c
594
mp->m_buf_workqueue = alloc_workqueue("xfs-buf/%s",
fs/xfs/xfs_super.c
596
1, mp->m_super->s_id);
fs/xfs/xfs_super.c
597
if (!mp->m_buf_workqueue)
fs/xfs/xfs_super.c
600
mp->m_unwritten_workqueue = alloc_workqueue("xfs-conv/%s",
fs/xfs/xfs_super.c
602
0, mp->m_super->s_id);
fs/xfs/xfs_super.c
603
if (!mp->m_unwritten_workqueue)
fs/xfs/xfs_super.c
606
mp->m_reclaim_workqueue = alloc_workqueue("xfs-reclaim/%s",
fs/xfs/xfs_super.c
608
0, mp->m_super->s_id);
fs/xfs/xfs_super.c
609
if (!mp->m_reclaim_workqueue)
fs/xfs/xfs_super.c
612
mp->m_blockgc_wq = alloc_workqueue("xfs-blockgc/%s",
fs/xfs/xfs_super.c
614
0, mp->m_super->s_id);
fs/xfs/xfs_super.c
615
if (!mp->m_blockgc_wq)
fs/xfs/xfs_super.c
618
mp->m_inodegc_wq = alloc_workqueue("xfs-inodegc/%s",
fs/xfs/xfs_super.c
620
1, mp->m_super->s_id);
fs/xfs/xfs_super.c
621
if (!mp->m_inodegc_wq)
fs/xfs/xfs_super.c
624
mp->m_sync_workqueue = alloc_workqueue("xfs-sync/%s",
fs/xfs/xfs_super.c
626
mp->m_super->s_id);
fs/xfs/xfs_super.c
627
if (!mp->m_sync_workqueue)
fs/xfs/xfs_super.c
633
destroy_workqueue(mp->m_inodegc_wq);
fs/xfs/xfs_super.c
635
destroy_workqueue(mp->m_blockgc_wq);
fs/xfs/xfs_super.c
637
destroy_workqueue(mp->m_reclaim_workqueue);
fs/xfs/xfs_super.c
639
destroy_workqueue(mp->m_unwritten_workqueue);
fs/xfs/xfs_super.c
641
destroy_workqueue(mp->m_buf_workqueue);
fs/xfs/xfs_super.c
648
struct xfs_mount *mp)
fs/xfs/xfs_super.c
650
destroy_workqueue(mp->m_sync_workqueue);
fs/xfs/xfs_super.c
651
destroy_workqueue(mp->m_blockgc_wq);
fs/xfs/xfs_super.c
652
destroy_workqueue(mp->m_inodegc_wq);
fs/xfs/xfs_super.c
653
destroy_workqueue(mp->m_reclaim_workqueue);
fs/xfs/xfs_super.c
654
destroy_workqueue(mp->m_unwritten_workqueue);
fs/xfs/xfs_super.c
655
destroy_workqueue(mp->m_buf_workqueue);
fs/xfs/xfs_super.c
662
struct xfs_mount *mp = container_of(work, struct xfs_mount,
fs/xfs/xfs_super.c
664
struct super_block *sb = mp->m_super;
fs/xfs/xfs_super.c
680
struct xfs_mount *mp)
fs/xfs/xfs_super.c
686
if (flush_work(&mp->m_flush_inodes_work))
fs/xfs/xfs_super.c
689
queue_work(mp->m_sync_workqueue, &mp->m_flush_inodes_work);
fs/xfs/xfs_super.c
690
flush_work(&mp->m_flush_inodes_work);
fs/xfs/xfs_super.c
790
struct xfs_mount *mp)
fs/xfs/xfs_super.c
792
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_super.c
793
xfs_free_buftarg(mp->m_logdev_targp);
fs/xfs/xfs_super.c
794
if (mp->m_rtdev_targp && mp->m_rtdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_super.c
795
xfs_free_buftarg(mp->m_rtdev_targp);
fs/xfs/xfs_super.c
796
if (mp->m_ddev_targp)
fs/xfs/xfs_super.c
797
xfs_free_buftarg(mp->m_ddev_targp);
fs/xfs/xfs_super.c
799
debugfs_remove(mp->m_debugfs);
fs/xfs/xfs_super.c
80
struct xfs_mount *mp,
fs/xfs/xfs_super.c
800
kfree(mp->m_rtname);
fs/xfs/xfs_super.c
801
kfree(mp->m_logname);
fs/xfs/xfs_super.c
803
kfree(mp->m_errortag);
fs/xfs/xfs_super.c
805
kfree(mp);
fs/xfs/xfs_super.c
813
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
816
trace_xfs_fs_sync_fs(mp, __return_address);
fs/xfs/xfs_super.c
824
error = xfs_log_force(mp, XFS_LOG_SYNC);
fs/xfs/xfs_super.c
843
xfs_inodegc_stop(mp);
fs/xfs/xfs_super.c
844
xfs_blockgc_stop(mp);
fs/xfs/xfs_super.c
845
xfs_zone_gc_stop(mp);
fs/xfs/xfs_super.c
85
mp->m_features &= ~(XFS_FEAT_DAX_ALWAYS | XFS_FEAT_DAX_NEVER);
fs/xfs/xfs_super.c
853
struct xfs_mount *mp)
fs/xfs/xfs_super.c
855
if (!mp->m_sb.sb_logstart)
fs/xfs/xfs_super.c
857
return mp->m_sb.sb_logblocks;
fs/xfs/xfs_super.c
862
struct xfs_mount *mp,
fs/xfs/xfs_super.c
866
xfs_sum_freecounter(mp, XC_FREE_BLOCKS);
fs/xfs/xfs_super.c
870
fdblocks - xfs_freecounter_unavailable(mp, XC_FREE_BLOCKS));
fs/xfs/xfs_super.c
876
st->f_blocks = mp->m_sb.sb_dblocks - xfs_internal_log_size(mp);
fs/xfs/xfs_super.c
88
mp->m_features |= XFS_FEAT_DAX_ALWAYS;
fs/xfs/xfs_super.c
886
struct xfs_mount *mp,
fs/xfs/xfs_super.c
889
st->f_bfree = xfs_rtbxlen_to_blen(mp,
fs/xfs/xfs_super.c
89
mp->m_features &= ~XFS_FEAT_DAX_NEVER;
fs/xfs/xfs_super.c
890
xfs_sum_freecounter(mp, XC_FREE_RTEXTENTS));
fs/xfs/xfs_super.c
891
st->f_blocks = mp->m_sb.sb_rblocks - xfs_rtbxlen_to_blen(mp,
fs/xfs/xfs_super.c
892
mp->m_free[XC_FREE_RTEXTENTS].res_total);
fs/xfs/xfs_super.c
897
struct xfs_mount *mp,
fs/xfs/xfs_super.c
900
uint64_t icount = percpu_counter_sum(&mp->m_icount);
fs/xfs/xfs_super.c
901
uint64_t ifree = percpu_counter_sum(&mp->m_ifree);
fs/xfs/xfs_super.c
902
uint64_t fakeinos = XFS_FSB_TO_INO(mp, st->f_bfree);
fs/xfs/xfs_super.c
905
if (M_IGEO(mp)->maxicount)
fs/xfs/xfs_super.c
907
M_IGEO(mp)->maxicount);
fs/xfs/xfs_super.c
911
mp->m_sb.sb_icount);
fs/xfs/xfs_super.c
92
mp->m_features |= XFS_FEAT_DAX_NEVER;
fs/xfs/xfs_super.c
922
struct xfs_mount *mp = XFS_M(dentry->d_sb);
fs/xfs/xfs_super.c
929
xfs_inodegc_push(mp);
fs/xfs/xfs_super.c
93
mp->m_features &= ~XFS_FEAT_DAX_ALWAYS;
fs/xfs/xfs_super.c
933
st->f_bsize = mp->m_sb.sb_blocksize;
fs/xfs/xfs_super.c
934
st->f_fsid = u64_to_fsid(huge_encode_dev(mp->m_ddev_targp->bt_dev));
fs/xfs/xfs_super.c
936
xfs_statfs_data(mp, st);
fs/xfs/xfs_super.c
937
xfs_statfs_inodes(mp, st);
fs/xfs/xfs_super.c
939
if (XFS_IS_REALTIME_MOUNT(mp) &&
fs/xfs/xfs_super.c
941
xfs_statfs_rt(mp, st);
fs/xfs/xfs_super.c
944
((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) ==
fs/xfs/xfs_super.c
958
struct xfs_mount *mp)
fs/xfs/xfs_super.c
963
mp->m_free[i].res_saved = mp->m_free[i].res_total;
fs/xfs/xfs_super.c
964
xfs_reserve_blocks(mp, i, 0);
fs/xfs/xfs_super.c
970
struct xfs_mount *mp)
fs/xfs/xfs_super.c
976
if (mp->m_free[i].res_saved) {
fs/xfs/xfs_super.c
977
resblks = mp->m_free[i].res_saved;
fs/xfs/xfs_super.c
978
mp->m_free[i].res_saved = 0;
fs/xfs/xfs_super.c
980
resblks = xfs_default_resblks(mp, i);
fs/xfs/xfs_super.c
981
xfs_reserve_blocks(mp, i, resblks);
fs/xfs/xfs_super.c
995
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.h
89
extern void xfs_flush_inodes(struct xfs_mount *mp);
fs/xfs/xfs_super.h
96
extern void xfs_reinit_percpu_counters(struct xfs_mount *mp);
fs/xfs/xfs_symlink.c
117
if (xfs_is_shutdown(mp))
fs/xfs/xfs_symlink.c
139
if (pathlen <= XFS_LITINO(mp) && !xfs_has_parent(mp))
fs/xfs/xfs_symlink.c
142
fs_blocks = xfs_symlink_blocks(mp, pathlen);
fs/xfs/xfs_symlink.c
143
resblks = xfs_symlink_space_res(mp, link_name->len, fs_blocks);
fs/xfs/xfs_symlink.c
145
error = xfs_parent_start(mp, &du.ppargs);
fs/xfs/xfs_symlink.c
149
error = xfs_trans_alloc_icreate(mp, &M_RES(mp)->tr_symlink, udqp, gdqp,
fs/xfs/xfs_symlink.c
188
resblks -= XFS_IALLOC_SPACE_RES(mp);
fs/xfs/xfs_symlink.c
208
if (xfs_has_wsync(mp) || xfs_has_dirsync(mp))
fs/xfs/xfs_symlink.c
222
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_symlink.c
239
xfs_parent_finish(mp, du.ppargs);
fs/xfs/xfs_symlink.c
263
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
277
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_itruncate, 0, 0, 0, &tp);
fs/xfs/xfs_symlink.c
300
ASSERT(xfs_is_shutdown(mp));
fs/xfs/xfs_symlink.c
328
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
333
if (xfs_is_shutdown(mp))
fs/xfs/xfs_symlink.c
341
xfs_alert(mp, "%s: inode (0x%llx) bad symlink length (%d)",
fs/xfs/xfs_symlink.c
36
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_symlink.c
42
if (xfs_is_shutdown(mp))
fs/xfs/xfs_symlink.c
54
xfs_alert(mp, "%s: inode (%llu) bad symlink length (%lld)",
fs/xfs/xfs_symlink.c
92
struct xfs_mount *mp = dp->i_mount;
fs/xfs/xfs_sysfs.c
584
struct xfs_mount *mp = err_to_mp(kobject);
fs/xfs/xfs_sysfs.c
586
return sysfs_emit(buf, "%d\n", mp->m_fail_unmount);
fs/xfs/xfs_sysfs.c
595
struct xfs_mount *mp = err_to_mp(kobject);
fs/xfs/xfs_sysfs.c
606
mp->m_fail_unmount = val;
fs/xfs/xfs_sysfs.c
662
struct xfs_mount *mp,
fs/xfs/xfs_sysfs.c
675
&mp->m_error_kobj, parent_name);
fs/xfs/xfs_sysfs.c
680
cfg = &mp->m_error_cfg[class][i];
fs/xfs/xfs_sysfs.c
698
cfg = &mp->m_error_cfg[class][i];
fs/xfs/xfs_sysfs.c
767
struct xfs_mount *mp)
fs/xfs/xfs_sysfs.c
771
super_set_sysfs_name_id(mp->m_super);
fs/xfs/xfs_sysfs.c
774
error = xfs_sysfs_init(&mp->m_kobj, &xfs_mp_ktype,
fs/xfs/xfs_sysfs.c
775
NULL, mp->m_super->s_id);
fs/xfs/xfs_sysfs.c
780
error = xfs_sysfs_init(&mp->m_stats.xs_kobj, &xfs_stats_ktype,
fs/xfs/xfs_sysfs.c
781
&mp->m_kobj, "stats");
fs/xfs/xfs_sysfs.c
786
error = xfs_sysfs_init(&mp->m_error_kobj, &xfs_error_ktype,
fs/xfs/xfs_sysfs.c
787
&mp->m_kobj, "error");
fs/xfs/xfs_sysfs.c
792
error = sysfs_create_file(&mp->m_error_kobj.kobject,
fs/xfs/xfs_sysfs.c
799
error = xfs_error_sysfs_init_class(mp, XFS_ERR_METADATA,
fs/xfs/xfs_sysfs.c
800
"metadata", &mp->m_error_meta_kobj,
fs/xfs/xfs_sysfs.c
805
if (IS_ENABLED(CONFIG_XFS_RT) && xfs_has_zoned(mp)) {
fs/xfs/xfs_sysfs.c
807
error = xfs_sysfs_init(&mp->m_zoned_kobj, &xfs_zoned_ktype,
fs/xfs/xfs_sysfs.c
808
&mp->m_kobj, "zoned");
fs/xfs/xfs_sysfs.c
816
xfs_sysfs_del(&mp->m_error_kobj);
fs/xfs/xfs_sysfs.c
818
xfs_sysfs_del(&mp->m_stats.xs_kobj);
fs/xfs/xfs_sysfs.c
820
xfs_sysfs_del(&mp->m_kobj);
fs/xfs/xfs_sysfs.c
826
struct xfs_mount *mp)
fs/xfs/xfs_sysfs.c
831
if (IS_ENABLED(CONFIG_XFS_RT) && xfs_has_zoned(mp))
fs/xfs/xfs_sysfs.c
832
xfs_sysfs_del(&mp->m_zoned_kobj);
fs/xfs/xfs_sysfs.c
836
cfg = &mp->m_error_cfg[i][j];
fs/xfs/xfs_sysfs.c
841
xfs_sysfs_del(&mp->m_error_meta_kobj);
fs/xfs/xfs_sysfs.c
842
xfs_sysfs_del(&mp->m_error_kobj);
fs/xfs/xfs_sysfs.c
843
xfs_sysfs_del(&mp->m_stats.xs_kobj);
fs/xfs/xfs_sysfs.c
844
xfs_sysfs_del(&mp->m_kobj);
fs/xfs/xfs_sysfs.c
849
struct xfs_mount *mp,
fs/xfs/xfs_sysfs.c
860
cfg = &mp->m_error_cfg[error_class][XFS_ERR_EIO];
fs/xfs/xfs_sysfs.c
863
cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENOSPC];
fs/xfs/xfs_sysfs.c
866
cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENODEV];
fs/xfs/xfs_sysfs.c
869
cfg = &mp->m_error_cfg[error_class][XFS_ERR_DEFAULT];
fs/xfs/xfs_sysfs.h
55
int xfs_mount_sysfs_init(struct xfs_mount *mp);
fs/xfs/xfs_sysfs.h
56
void xfs_mount_sysfs_del(struct xfs_mount *mp);
fs/xfs/xfs_trace.h
1633
TP_PROTO(struct xfs_mount *mp, xfs_lsn_t lsn, unsigned long caller_ip),
fs/xfs/xfs_trace.h
1634
TP_ARGS(mp, lsn, caller_ip),
fs/xfs/xfs_trace.h
1641
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
178
TP_PROTO(struct xfs_mount *mp, enum xfs_group_type type,
fs/xfs/xfs_trace.h
181
TP_ARGS(mp, type, max_write, max_ioend, max_gsize, awu_max),
fs/xfs/xfs_trace.h
191
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
208
TP_PROTO(struct xfs_mount *mp, unsigned int per_intent,
fs/xfs/xfs_trace.h
2105
TP_PROTO(struct xfs_mount *mp, struct xfs_agf *agf, int flags,
fs/xfs/xfs_trace.h
2107
TP_ARGS(mp, agf, flags, caller_ip),
fs/xfs/xfs_trace.h
211
TP_ARGS(mp, per_intent, step_size, logres, blockcount),
fs/xfs/xfs_trace.h
2125
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
2160
TP_PROTO(struct xfs_mount *mp, struct xfs_agf *agf, int flags, \
fs/xfs/xfs_trace.h
2162
TP_ARGS(mp, agf, flags, caller_ip))
fs/xfs/xfs_trace.h
220
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
2224
__entry->dev = args->mp->m_super->s_dev;
fs/xfs/xfs_trace.h
235
TP_PROTO(struct xfs_mount *mp, unsigned int per_intent,
fs/xfs/xfs_trace.h
238
TP_ARGS(mp, per_intent, step_size, blockcount, min_logblocks, logres),
fs/xfs/xfs_trace.h
249
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
254
__entry->cur_logblocks = mp->m_sb.sb_logblocks;
fs/xfs/xfs_trace.h
268
TP_PROTO(struct xfs_mount *mp, const struct xfs_defer_op_type *ops,
fs/xfs/xfs_trace.h
270
TP_ARGS(mp, ops, error),
fs/xfs/xfs_trace.h
277
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
2866
TP_PROTO(struct xfs_mount *mp,
fs/xfs/xfs_trace.h
2868
TP_ARGS(mp, rtbno, len),
fs/xfs/xfs_trace.h
2875
__entry->dev = mp->m_rtdev_targp->bt_dev;
fs/xfs/xfs_trace.h
2887
TP_PROTO(struct xfs_mount *mp, \
fs/xfs/xfs_trace.h
2889
TP_ARGS(mp, rtbno, len))
fs/xfs/xfs_trace.h
3064
TP_PROTO(struct xfs_mount *mp, struct xfs_defer_pending *dfp),
fs/xfs/xfs_trace.h
3065
TP_ARGS(mp, dfp),
fs/xfs/xfs_trace.h
3075
__entry->dev = mp ? mp->m_super->s_dev : 0;
fs/xfs/xfs_trace.h
3092
TP_PROTO(struct xfs_mount *mp, struct xfs_defer_pending *dfp), \
fs/xfs/xfs_trace.h
3093
TP_ARGS(mp, dfp))
fs/xfs/xfs_trace.h
3114
TP_PROTO(struct xfs_mount *mp, struct xfs_extent_free_item *free),
fs/xfs/xfs_trace.h
3115
TP_ARGS(mp, free),
fs/xfs/xfs_trace.h
3125
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
3128
__entry->agbno = xfs_fsb_to_gbno(mp, free->xefi_startblock,
fs/xfs/xfs_trace.h
3143
TP_PROTO(struct xfs_mount *mp, struct xfs_extent_free_item *free), \
fs/xfs/xfs_trace.h
3144
TP_ARGS(mp, free))
fs/xfs/xfs_trace.h
3150
TP_PROTO(struct xfs_mount *mp, struct xfs_defer_pending *dfp,
fs/xfs/xfs_trace.h
3152
TP_ARGS(mp, dfp, item),
fs/xfs/xfs_trace.h
3163
__entry->dev = mp ? mp->m_super->s_dev : 0;
fs/xfs/xfs_trace.h
3182
TP_PROTO(struct xfs_mount *mp, struct xfs_defer_pending *dfp, \
fs/xfs/xfs_trace.h
3184
TP_ARGS(mp, dfp, item))
fs/xfs/xfs_trace.h
3368
TP_PROTO(struct xfs_mount *mp, struct xfs_rmap_intent *ri),
fs/xfs/xfs_trace.h
3369
TP_ARGS(mp, ri),
fs/xfs/xfs_trace.h
3383
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
3386
__entry->gbno = xfs_fsb_to_gbno(mp,
fs/xfs/xfs_trace.h
3410
TP_PROTO(struct xfs_mount *mp, struct xfs_rmap_intent *ri), \
fs/xfs/xfs_trace.h
3411
TP_ARGS(mp, ri))
fs/xfs/xfs_trace.h
3451
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trace.h
3453
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
3457
!xfs_has_rtgroups(mp)) {
fs/xfs/xfs_trace.h
3467
__entry->gbno = xfs_fsb_to_gbno(mp,
fs/xfs/xfs_trace.h
374
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_trace.h
376
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
379
__entry->nr_open = mp->m_zone_info->zi_nr_open_zones;
fs/xfs/xfs_trace.h
3928
TP_PROTO(struct xfs_mount *mp, struct xfs_refcount_intent *refc),
fs/xfs/xfs_trace.h
3929
TP_ARGS(mp, refc),
fs/xfs/xfs_trace.h
3939
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
3943
__entry->gbno = xfs_fsb_to_gbno(mp, refc->ri_startblock,
fs/xfs/xfs_trace.h
3957
TP_PROTO(struct xfs_mount *mp, struct xfs_refcount_intent *refc), \
fs/xfs/xfs_trace.h
3958
TP_ARGS(mp, refc))
fs/xfs/xfs_trace.h
4236
TP_PROTO(struct xfs_mount *mp, u32 keydev, xfs_agnumber_t agno,
fs/xfs/xfs_trace.h
4238
TP_ARGS(mp, keydev, agno, frec),
fs/xfs/xfs_trace.h
4251
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4274
TP_PROTO(struct xfs_mount *mp, u32 keydev, xfs_agnumber_t agno,
fs/xfs/xfs_trace.h
4276
TP_ARGS(mp, keydev, agno, rmap),
fs/xfs/xfs_trace.h
4287
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4306
TP_PROTO(struct xfs_mount *mp, u32 keydev, xfs_agnumber_t agno, \
fs/xfs/xfs_trace.h
4308
TP_ARGS(mp, keydev, agno, rmap))
fs/xfs/xfs_trace.h
4313
TP_PROTO(struct xfs_mount *mp, u32 keydev, xfs_fsblock_t bno),
fs/xfs/xfs_trace.h
4314
TP_ARGS(mp, keydev, bno),
fs/xfs/xfs_trace.h
4321
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4332
TP_PROTO(struct xfs_mount *mp, u32 keydev, uint64_t bno), \
fs/xfs/xfs_trace.h
4333
TP_ARGS(mp, keydev, bno))
fs/xfs/xfs_trace.h
4338
TP_PROTO(struct xfs_mount *mp, struct xfs_fsmap *fsmap),
fs/xfs/xfs_trace.h
4339
TP_ARGS(mp, fsmap),
fs/xfs/xfs_trace.h
4350
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4369
TP_PROTO(struct xfs_mount *mp, struct xfs_fsmap *fsmap), \
fs/xfs/xfs_trace.h
4370
TP_ARGS(mp, fsmap))
fs/xfs/xfs_trace.h
4376
TP_PROTO(struct xfs_mount *mp, unsigned int type,
fs/xfs/xfs_trace.h
4378
TP_ARGS(mp, type, res),
fs/xfs/xfs_trace.h
4387
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4403
TP_PROTO(struct xfs_mount *mp, unsigned int type, \
fs/xfs/xfs_trace.h
4405
TP_ARGS(mp, type, res))
fs/xfs/xfs_trace.h
4410
TP_PROTO(struct xfs_mount *mp, const struct xfs_trans_res *res),
fs/xfs/xfs_trace.h
4411
TP_ARGS(mp, res),
fs/xfs/xfs_trace.h
4418
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4589
TP_PROTO(struct xfs_mount *mp, unsigned int flags),
fs/xfs/xfs_trace.h
4590
TP_ARGS(mp, flags),
fs/xfs/xfs_trace.h
4596
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4605
TP_PROTO(struct xfs_mount *mp, unsigned int flags), \
fs/xfs/xfs_trace.h
4606
TP_ARGS(mp, flags))
fs/xfs/xfs_trace.h
4689
TP_PROTO(struct xfs_mount *mp, unsigned int nr_threads, pid_t pid),
fs/xfs/xfs_trace.h
4690
TP_ARGS(mp, nr_threads, pid),
fs/xfs/xfs_trace.h
4697
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4707
TP_PROTO(struct xfs_mount *mp, int new_dalign, xfs_ino_t calc_rootino),
fs/xfs/xfs_trace.h
4708
TP_ARGS(mp, new_dalign, calc_rootino),
fs/xfs/xfs_trace.h
4716
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4718
__entry->sb_rootino = mp->m_sb.sb_rootino;
fs/xfs/xfs_trace.h
4873
TP_PROTO(struct xfs_mount *mp, time64_t min, time64_t max),
fs/xfs/xfs_trace.h
4874
TP_ARGS(mp, min, max),
fs/xfs/xfs_trace.h
488
TP_PROTO(struct xfs_mount *mp),
fs/xfs/xfs_trace.h
4881
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
489
TP_ARGS(mp),
fs/xfs/xfs_trace.h
4893
TP_PROTO(struct xfs_mount *mp, long long min, long long max), \
fs/xfs/xfs_trace.h
4894
TP_ARGS(mp, min, max))
fs/xfs/xfs_trace.h
4899
TP_PROTO(struct xfs_mount *mp, struct xfs_icwalk *icw,
fs/xfs/xfs_trace.h
4901
TP_ARGS(mp, icw, caller_ip),
fs/xfs/xfs_trace.h
4913
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
4915
__entry->uid = icw ? from_kuid(mp->m_super->s_user_ns,
fs/xfs/xfs_trace.h
4917
__entry->gid = icw ? from_kgid(mp->m_super->s_user_ns,
fs/xfs/xfs_trace.h
4936
TP_PROTO(struct xfs_mount *mp, struct xfs_icwalk *icw, \
fs/xfs/xfs_trace.h
4938
TP_ARGS(mp, icw, caller_ip))
fs/xfs/xfs_trace.h
497
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
498
__entry->rgcount = mp->m_sb.sb_rgcount;
fs/xfs/xfs_trace.h
499
__entry->blocks = mp->m_groups[XG_TYPE_RTG].blocks;
fs/xfs/xfs_trace.h
500
__entry->max_open_zones = mp->m_max_open_zones;
fs/xfs/xfs_trace.h
5053
TP_PROTO(struct xfs_mount *mp, int ptag, int flags, const char *fname,
fs/xfs/xfs_trace.h
5055
TP_ARGS(mp, ptag, flags, fname, line_num),
fs/xfs/xfs_trace.h
5064
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
511
TP_PROTO(struct xfs_mount *mp, unsigned int shrinker_hits),
fs/xfs/xfs_trace.h
512
TP_ARGS(mp, shrinker_hits),
fs/xfs/xfs_trace.h
5162
TP_PROTO(struct xfs_mount *mp, struct xfbtree *xfbt,
fs/xfs/xfs_trace.h
5164
TP_ARGS(mp, xfbt, ops),
fs/xfs/xfs_trace.h
518
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
527
TP_PROTO(struct xfs_mount *mp, void *caller_ip),
fs/xfs/xfs_trace.h
528
TP_ARGS(mp, caller_ip),
fs/xfs/xfs_trace.h
537
if (mp) {
fs/xfs/xfs_trace.h
538
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
539
__entry->mflags = mp->m_features;
fs/xfs/xfs_trace.h
540
__entry->opstate = mp->m_opstate;
fs/xfs/xfs_trace.h
541
__entry->sbflags = mp->m_super->s_flags;
fs/xfs/xfs_trace.h
5442
TP_PROTO(struct xfs_mount *mp, unsigned long long bmbt_blocks,
fs/xfs/xfs_trace.h
5444
TP_ARGS(mp, bmbt_blocks, rmapbt_blocks),
fs/xfs/xfs_trace.h
5451
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
5516
TP_PROTO(struct xfs_mount *mp, const struct xfs_exchmaps_intent *xmi),
fs/xfs/xfs_trace.h
5517
TP_ARGS(mp, xmi),
fs/xfs/xfs_trace.h
5532
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
555
TP_PROTO(struct xfs_mount *mp, void *caller_ip), \
fs/xfs/xfs_trace.h
5556
TP_PROTO(struct xfs_mount *mp, const struct xfs_exchmaps_intent *xmi), \
fs/xfs/xfs_trace.h
5557
TP_ARGS(mp, xmi))
fs/xfs/xfs_trace.h
556
TP_ARGS(mp, caller_ip))
fs/xfs/xfs_trace.h
5562
TP_PROTO(struct xfs_mount *mp,
fs/xfs/xfs_trace.h
5568
TP_ARGS(mp, left, curr, new, right, delta, state),
fs/xfs/xfs_trace.h
5587
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
570
TP_PROTO(struct xfs_mount *mp, struct shrink_control *sc,
fs/xfs/xfs_trace.h
572
TP_ARGS(mp, sc, caller_ip),
fs/xfs/xfs_trace.h
579
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
5836
TP_PROTO(struct xfs_mount *mp, xfs_filblks_t len),
fs/xfs/xfs_trace.h
5837
TP_ARGS(mp, len),
fs/xfs/xfs_trace.h
5847
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
5848
__entry->freeblks = xfs_sum_freecounter_raw(mp, XC_FREE_BLOCKS);
fs/xfs/xfs_trace.h
5849
__entry->reserved = mp->m_metafile_resv_avail;
fs/xfs/xfs_trace.h
5850
__entry->asked = mp->m_metafile_resv_target;
fs/xfs/xfs_trace.h
5851
__entry->used = mp->m_metafile_resv_used;
fs/xfs/xfs_trace.h
5864
TP_PROTO(struct xfs_mount *mp, xfs_filblks_t len), \
fs/xfs/xfs_trace.h
5865
TP_ARGS(mp, len))
fs/xfs/xfs_trace.h
5875
TP_PROTO(const struct xfs_mount *mp, unsigned int min_logfsbs),
fs/xfs/xfs_trace.h
5876
TP_ARGS(mp, min_logfsbs),
fs/xfs/xfs_trace.h
5883
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
5884
__entry->logblocks = mp->m_sb.sb_logblocks;
fs/xfs/xfs_trace.h
5899
TP_PROTO(struct xfs_mount *mp, enum xfs_free_counter ctr,
fs/xfs/xfs_trace.h
5901
TP_ARGS(mp, ctr, delta, caller_ip),
fs/xfs/xfs_trace.h
5911
__entry->dev = mp->m_super->s_dev;
fs/xfs/xfs_trace.h
5914
__entry->avail = mp->m_free[ctr].res_avail;
fs/xfs/xfs_trace.h
5915
__entry->total = mp->m_free[ctr].res_total;
fs/xfs/xfs_trace.h
5928
TP_PROTO(struct xfs_mount *mp, enum xfs_free_counter ctr, \
fs/xfs/xfs_trace.h
5930
TP_ARGS(mp, ctr, delta, caller_ip))
fs/xfs/xfs_trace.h
6346
TP_PROTO(const struct xfs_mount *mp, const struct xfs_verify_media *me,
fs/xfs/xfs_trace.h
6349
TP_ARGS(mp, me, fdev, daddr, bbcount, folio),
fs/xfs/xfs_trace.h
6361
__entry->dev = mp->m_ddev_targp->bt_dev;
fs/xfs/xfs_trace.h
6382
TP_PROTO(const struct xfs_mount *mp, const struct xfs_verify_media *me,
fs/xfs/xfs_trace.h
6384
TP_ARGS(mp, me, fdev),
fs/xfs/xfs_trace.h
6393
__entry->dev = mp->m_ddev_targp->bt_dev;
fs/xfs/xfs_trace.h
6408
TP_PROTO(const struct xfs_mount *mp, const struct xfs_verify_media *me,
fs/xfs/xfs_trace.h
6411
TP_ARGS(mp, me, fdev, daddr, bbcount, status),
fs/xfs/xfs_trace.h
6423
__entry->dev = mp->m_ddev_targp->bt_dev;
fs/xfs/xfs_trans.c
1075
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1080
error = xfs_trans_alloc(mp, resv, dblocks,
fs/xfs/xfs_trans.c
1081
xfs_extlen_to_rtxlen(mp, rblocks),
fs/xfs/xfs_trans.c
1157
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1158
unsigned int rtx = xfs_extlen_to_rtxlen(mp, rblocks);
fs/xfs/xfs_trans.c
1167
if (!XFS_IS_QUOTA_ON(mp) || xfs_is_quota_inode(&mp->m_sb, ip->i_ino))
fs/xfs/xfs_trans.c
1179
xfs_add_fdblocks(mp, dblocks);
fs/xfs/xfs_trans.c
1181
xfs_add_frextents(mp, rtx);
fs/xfs/xfs_trans.c
1192
struct xfs_mount *mp,
fs/xfs/xfs_trans.c
1205
error = xfs_trans_alloc(mp, resv, dblocks, 0, 0, &tp);
fs/xfs/xfs_trans.c
1212
xfs_blockgc_free_dquots(mp, udqp, gdqp, pdqp, 0);
fs/xfs/xfs_trans.c
1244
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1252
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_ichange, 0, 0, 0, &tp);
fs/xfs/xfs_trans.c
1305
error = xfs_trans_reserve_quota_bydquots(tp, mp, udqp, gdqp,
fs/xfs/xfs_trans.c
1309
xfs_blockgc_free_dquots(mp, udqp, gdqp, pdqp, 0);
fs/xfs/xfs_trans.c
1318
error = xfs_trans_reserve_quota_bydquots(tp, mp, udqp, gdqp,
fs/xfs/xfs_trans.c
1322
xfs_blockgc_free_dquots(mp, udqp, gdqp, pdqp, 0);
fs/xfs/xfs_trans.c
1362
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans.c
1370
error = xfs_trans_alloc(mp, resv, resblks, 0, 0, &tp);
fs/xfs/xfs_trans.c
1374
error = xfs_trans_alloc(mp, resv, resblks, 0, 0, &tp);
fs/xfs/xfs_trans.c
151
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
163
error = xfs_dec_fdblocks(mp, blocks, rsvd);
fs/xfs/xfs_trans.c
174
error = xfs_log_reserve(mp, resp->tr_logres, resp->tr_logcount,
fs/xfs/xfs_trans.c
188
error = xfs_dec_frextents(mp, rtextents);
fs/xfs/xfs_trans.c
199
xfs_log_ticket_ungrant(mp->m_log, tp->t_ticket);
fs/xfs/xfs_trans.c
205
xfs_add_fdblocks(mp, blocks);
fs/xfs/xfs_trans.c
213
struct xfs_mount *mp,
fs/xfs/xfs_trans.c
218
ASSERT(!(flags & XFS_TRANS_RES_FDBLKS) || xfs_has_lazysbcount(mp));
fs/xfs/xfs_trans.c
222
sb_start_intwrite(mp->m_super);
fs/xfs/xfs_trans.c
225
tp->t_mountp = mp;
fs/xfs/xfs_trans.c
235
struct xfs_mount *mp,
fs/xfs/xfs_trans.c
254
tp = __xfs_trans_alloc(mp, flags);
fs/xfs/xfs_trans.c
255
WARN_ON(mp->m_super->s_writers.frozen == SB_FREEZE_COMPLETE);
fs/xfs/xfs_trans.c
266
error = xfs_blockgc_flush_all(mp);
fs/xfs/xfs_trans.c
301
struct xfs_mount *mp)
fs/xfs/xfs_trans.c
303
return __xfs_trans_alloc(mp, XFS_TRANS_NO_WRITECOUNT);
fs/xfs/xfs_trans.c
329
xfs_mount_t *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
334
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_trans.c
339
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_trans.c
351
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_trans.c
36
struct xfs_mount *mp)
fs/xfs/xfs_trans.c
368
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_trans.c
378
if (xfs_has_lazysbcount(mp))
fs/xfs/xfs_trans.c
392
if (xfs_has_rtgroups(mp))
fs/xfs/xfs_trans.c
403
if (xfs_has_rtgroups(mp))
fs/xfs/xfs_trans.c
42
res = (struct xfs_trans_res *)M_RES(mp);
fs/xfs/xfs_trans.c
43
end_res = (struct xfs_trans_res *)(M_RES(mp) + 1);
fs/xfs/xfs_trans.c
45
trace_xfs_trans_resv_calc(mp, i, res);
fs/xfs/xfs_trans.c
455
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
463
if (!xfs_has_lazysbcount(mp)) {
fs/xfs/xfs_trans.c
48
# define xfs_trans_trace_reservations(mp)
fs/xfs/xfs_trans.c
490
!xfs_has_rtgroups(mp)) {
fs/xfs/xfs_trans.c
495
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_trans.c
497
mp->m_sb.sb_frextents += rtxdelta;
fs/xfs/xfs_trans.c
498
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_trans.c
503
mp->m_ddev_targp->bt_nr_sectors +=
fs/xfs/xfs_trans.c
504
XFS_FSB_TO_BB(mp, tp->t_dblocks_delta);
fs/xfs/xfs_trans.c
524
if (xfs_has_rtgroups(mp)) {
fs/xfs/xfs_trans.c
537
mp->m_rtdev_targp->bt_nr_sectors +=
fs/xfs/xfs_trans.c
538
XFS_FSB_TO_BB(mp, tp->t_rblocks_delta);
fs/xfs/xfs_trans.c
57
struct xfs_mount *mp)
fs/xfs/xfs_trans.c
59
xfs_trans_resv_calc(mp, M_RES(mp));
fs/xfs/xfs_trans.c
591
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
60
xfs_trans_trace_reservations(mp);
fs/xfs/xfs_trans.c
611
if (xfs_has_lazysbcount(mp) || (tp->t_flags & XFS_TRANS_SB_DIRTY)) {
fs/xfs/xfs_trans.c
617
if (xfs_has_rtgroups(mp) || (tp->t_flags & XFS_TRANS_SB_DIRTY)) {
fs/xfs/xfs_trans.c
622
if (xfs_has_lazysbcount(mp) || (tp->t_flags & XFS_TRANS_SB_DIRTY)) {
fs/xfs/xfs_trans.c
629
xfs_add_fdblocks(mp, blkdelta);
fs/xfs/xfs_trans.c
632
percpu_counter_add_batch(&mp->m_icount, idelta,
fs/xfs/xfs_trans.c
636
percpu_counter_add(&mp->m_ifree, ifreedelta);
fs/xfs/xfs_trans.c
639
xfs_add_frextents(mp, rtxdelta);
fs/xfs/xfs_trans.c
645
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_trans.c
646
mp->m_sb.sb_fdblocks += tp->t_fdblocks_delta + tp->t_res_fdblocks_delta;
fs/xfs/xfs_trans.c
647
mp->m_sb.sb_icount += idelta;
fs/xfs/xfs_trans.c
648
mp->m_sb.sb_ifree += ifreedelta;
fs/xfs/xfs_trans.c
654
mp->m_sb.sb_dblocks += tp->t_dblocks_delta;
fs/xfs/xfs_trans.c
655
mp->m_sb.sb_agcount += tp->t_agcount_delta;
fs/xfs/xfs_trans.c
656
mp->m_sb.sb_imax_pct += tp->t_imaxpct_delta;
fs/xfs/xfs_trans.c
658
xfs_mount_sb_set_rextsize(mp, &mp->m_sb,
fs/xfs/xfs_trans.c
659
mp->m_sb.sb_rextsize + tp->t_rextsize_delta);
fs/xfs/xfs_trans.c
660
mp->m_sb.sb_rbmblocks += tp->t_rbmblocks_delta;
fs/xfs/xfs_trans.c
661
mp->m_sb.sb_rblocks += tp->t_rblocks_delta;
fs/xfs/xfs_trans.c
662
mp->m_sb.sb_rextents += tp->t_rextents_delta;
fs/xfs/xfs_trans.c
663
mp->m_sb.sb_rextslog += tp->t_rextslog_delta;
fs/xfs/xfs_trans.c
664
mp->m_sb.sb_rgcount += tp->t_rgcount_delta;
fs/xfs/xfs_trans.c
665
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_trans.c
671
ASSERT(mp->m_sb.sb_imax_pct >= 0);
fs/xfs/xfs_trans.c
672
ASSERT(mp->m_sb.sb_rextslog >= 0);
fs/xfs/xfs_trans.c
777
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
804
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_trans.c
825
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
826
struct xlog *log = mp->m_log;
fs/xfs/xfs_trans.c
877
error = xfs_log_force_seq(mp, commit_seq, XFS_LOG_SYNC, NULL);
fs/xfs/xfs_trans.c
878
XFS_STATS_INC(mp, xs_trans_sync);
fs/xfs/xfs_trans.c
880
XFS_STATS_INC(mp, xs_trans_async);
fs/xfs/xfs_trans.c
904
XFS_STATS_INC(mp, xs_trans_empty);
fs/xfs/xfs_trans.c
949
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans.c
950
struct xlog *log = mp->m_log;
fs/xfs/xfs_trans.c
973
if (dirty && !xfs_is_shutdown(mp)) {
fs/xfs/xfs_trans.c
974
XFS_ERROR_REPORT("xfs_trans_cancel", XFS_ERRLEVEL_LOW, mp);
fs/xfs/xfs_trans.c
975
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_trans.h
111
void xfs_log_item_init(struct xfs_mount *mp, struct xfs_log_item *item,
fs/xfs/xfs_trans.h
167
int xfs_trans_alloc(struct xfs_mount *mp, struct xfs_trans_res *resp,
fs/xfs/xfs_trans.h
172
struct xfs_trans *xfs_trans_alloc_empty(struct xfs_mount *mp);
fs/xfs/xfs_trans.h
192
int xfs_trans_read_buf_map(struct xfs_mount *mp,
fs/xfs/xfs_trans.h
202
struct xfs_mount *mp,
fs/xfs/xfs_trans.h
212
return xfs_trans_read_buf_map(mp, tp, target, &map, 1,
fs/xfs/xfs_trans.h
258
int xfs_trans_alloc_icreate(struct xfs_mount *mp, struct xfs_trans_res *resv,
fs/xfs/xfs_trans_ail.c
474
struct xfs_mount *mp = ailp->ail_log->l_mp;
fs/xfs/xfs_trans_ail.c
491
XFS_STATS_INC(mp, xs_push_ail_success);
fs/xfs/xfs_trans_ail.c
509
XFS_STATS_INC(mp, xs_push_ail_flushing);
fs/xfs/xfs_trans_ail.c
517
XFS_STATS_INC(mp, xs_push_ail_pinned);
fs/xfs/xfs_trans_ail.c
524
XFS_STATS_INC(mp, xs_push_ail_locked);
fs/xfs/xfs_trans_ail.c
539
struct xfs_mount *mp = ailp->ail_log->l_mp;
fs/xfs/xfs_trans_ail.c
562
XFS_STATS_INC(mp, xs_push_ail_flush);
fs/xfs/xfs_trans_ail.c
576
XFS_STATS_INC(mp, xs_push_ail);
fs/xfs/xfs_trans_ail.c
944
xfs_mount_t *mp)
fs/xfs/xfs_trans_ail.c
952
ailp->ail_log = mp->m_log;
fs/xfs/xfs_trans_ail.c
960
mp->m_super->s_id);
fs/xfs/xfs_trans_ail.c
964
mp->m_ail = ailp;
fs/xfs/xfs_trans_ail.c
974
xfs_mount_t *mp)
fs/xfs/xfs_trans_ail.c
976
struct xfs_ail *ailp = mp->m_ail;
fs/xfs/xfs_trans_buf.c
227
struct xfs_mount *mp,
fs/xfs/xfs_trans_buf.c
262
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_trans_buf.c
318
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_trans_dquot.c
1000
error = xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1007
error = xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1011
xfs_trans_reserve_quota_bydquots(tp, mp, ip->i_udquot,
fs/xfs/xfs_trans_dquot.c
1029
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_trans_dquot.c
1031
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_trans_dquot.c
1034
return xfs_trans_reserve_quota_bydquots(tp, mp, udqp, gdqp, pdqp,
fs/xfs/xfs_trans_dquot.c
247
xfs_mount_t *mp = tp->t_mountp;
fs/xfs/xfs_trans_dquot.c
249
if (!XFS_IS_QUOTA_ON(mp) ||
fs/xfs/xfs_trans_dquot.c
250
xfs_is_quota_inode(&mp->m_sb, ip->i_ino) ||
fs/xfs/xfs_trans_dquot.c
254
if (XFS_IS_UQUOTA_ON(mp) && ip->i_udquot)
fs/xfs/xfs_trans_dquot.c
256
if (XFS_IS_GQUOTA_ON(mp) && ip->i_gdquot)
fs/xfs/xfs_trans_dquot.c
258
if (XFS_IS_PQUOTA_ON(mp) && ip->i_pdquot)
fs/xfs/xfs_trans_dquot.c
728
struct xfs_mount *mp,
fs/xfs/xfs_trans_dquot.c
749
mp->m_super->s_dev, type);
fs/xfs/xfs_trans_dquot.c
812
struct xfs_mount *mp,
fs/xfs/xfs_trans_dquot.c
818
struct xfs_quotainfo *q = mp->m_quotainfo;
fs/xfs/xfs_trans_dquot.c
850
xfs_quota_warn(mp, dqp, quota_nl + 3);
fs/xfs/xfs_trans_dquot.c
858
xfs_quota_warn(mp, dqp, quota_nl);
fs/xfs/xfs_trans_dquot.c
885
if (XFS_IS_CORRUPT(mp, dqp->q_blk.reserved < dqp->q_blk.count) ||
fs/xfs/xfs_trans_dquot.c
886
XFS_IS_CORRUPT(mp, dqp->q_rtb.reserved < dqp->q_rtb.count) ||
fs/xfs/xfs_trans_dquot.c
887
XFS_IS_CORRUPT(mp, dqp->q_ino.reserved < dqp->q_ino.count))
fs/xfs/xfs_trans_dquot.c
900
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_trans_dquot.c
901
xfs_fs_mark_sick(mp, XFS_SICK_FS_QUOTACHECK);
fs/xfs/xfs_trans_dquot.c
921
struct xfs_mount *mp,
fs/xfs/xfs_trans_dquot.c
931
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_trans_dquot.c
937
error = xfs_trans_dqresv(tp, mp, udqp, nblks, ninos, flags);
fs/xfs/xfs_trans_dquot.c
943
error = xfs_trans_dqresv(tp, mp, gdqp, nblks, ninos, flags);
fs/xfs/xfs_trans_dquot.c
949
error = xfs_trans_dqresv(tp, mp, pdqp, nblks, ninos, flags);
fs/xfs/xfs_trans_dquot.c
962
xfs_trans_dqresv(tp, mp, gdqp, -nblks, -ninos, flags);
fs/xfs/xfs_trans_dquot.c
966
xfs_trans_dqresv(tp, mp, udqp, -nblks, -ninos, flags);
fs/xfs/xfs_trans_dquot.c
984
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_trans_dquot.c
988
if (!XFS_IS_QUOTA_ON(mp))
fs/xfs/xfs_trans_dquot.c
993
ASSERT(!xfs_is_quota_inode(&mp->m_sb, ip->i_ino));
fs/xfs/xfs_verify_media.c
102
start_bno = xfs_daddr_to_rtb(mp, daddr);
fs/xfs/xfs_verify_media.c
103
end_bno = xfs_daddr_to_rtb(mp, daddr + bblen - 1);
fs/xfs/xfs_verify_media.c
105
start_bno = XFS_DADDR_TO_FSB(mp, daddr);
fs/xfs/xfs_verify_media.c
106
end_bno = XFS_DADDR_TO_FSB(mp, daddr + bblen - 1);
fs/xfs/xfs_verify_media.c
109
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_verify_media.c
110
start_gno = xfs_fsb_to_gno(mp, start_bno, type);
fs/xfs/xfs_verify_media.c
111
end_gno = xfs_fsb_to_gno(mp, end_bno, type);
fs/xfs/xfs_verify_media.c
112
while ((xg = xfs_group_next_range(mp, xg, start_gno, end_gno, type))) {
fs/xfs/xfs_verify_media.c
129
cur = xfs_rmapbt_init_cursor(mp, tp, agf_bp, pag);
fs/xfs/xfs_verify_media.c
143
xfs_fsb_to_gbno(mp, start_bno, type);
fs/xfs/xfs_verify_media.c
146
xfs_fsb_to_gbno(mp, end_bno, type);
fs/xfs/xfs_verify_media.c
213
struct xfs_mount *mp,
fs/xfs/xfs_verify_media.c
220
trace_xfs_verify_media_error(mp, me, btp->bt_dev, daddr, bio_bbcount,
fs/xfs/xfs_verify_media.c
247
xfs_healthmon_report_media(mp, me->me_dev, daddr, bio_bbcount);
fs/xfs/xfs_verify_media.c
249
if (!xfs_has_rmapbt(mp))
fs/xfs/xfs_verify_media.c
254
xfs_verify_report_losses(mp, XG_TYPE_AG, daddr, bio_bbcount);
fs/xfs/xfs_verify_media.c
257
xfs_verify_report_losses(mp, XG_TYPE_RTG, daddr, bio_bbcount);
fs/xfs/xfs_verify_media.c
265
struct xfs_mount *mp,
fs/xfs/xfs_verify_media.c
279
btp = mp->m_ddev_targp;
fs/xfs/xfs_verify_media.c
282
if (mp->m_logdev_targp != mp->m_ddev_targp)
fs/xfs/xfs_verify_media.c
283
btp = mp->m_logdev_targp;
fs/xfs/xfs_verify_media.c
286
btp = mp->m_rtdev_targp;
fs/xfs/xfs_verify_media.c
333
trace_xfs_verify_media(mp, me, btp->bt_dev, daddr, bbcount, folio);
fs/xfs/xfs_verify_media.c
360
xfs_verify_media_error(mp, me, btp, daddr, bio_bbcount,
fs/xfs/xfs_verify_media.c
401
trace_xfs_verify_media_end(mp, me, btp->bt_dev);
fs/xfs/xfs_verify_media.c
412
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_verify_media.c
42
struct xfs_mount *mp = cur->bc_mp;
fs/xfs/xfs_verify_media.c
435
error = xfs_verify_media(mp, &me);
fs/xfs/xfs_verify_media.c
58
error = xfs_iget(mp, cur->bc_tp, rec->rm_owner, 0, 0, &ip);
fs/xfs/xfs_verify_media.c
79
fserror_report_data_lost(VFS_I(ip), XFS_FSB_TO_B(mp, fileoff),
fs/xfs/xfs_verify_media.c
80
XFS_FSB_TO_B(mp, blocks), GFP_NOFS);
fs/xfs/xfs_verify_media.c
90
struct xfs_mount *mp,
fs/xfs/xfs_xattr.c
100
error = xfs_attr_grab_log_assist(mp);
fs/xfs/xfs_xattr.c
108
args->geo = mp->m_attr_geo;
fs/xfs/xfs_xattr.c
30
struct xfs_mount *mp)
fs/xfs/xfs_xattr.c
35
if (xfs_is_using_logged_xattrs(mp))
fs/xfs/xfs_xattr.c
44
if (!xfs_has_reflink(mp) && !xfs_has_rmapbt(mp))
fs/xfs/xfs_xattr.c
48
error = xfs_add_incompat_log_feature(mp,
fs/xfs/xfs_xattr.c
52
xfs_set_using_logged_xattrs(mp);
fs/xfs/xfs_xattr.c
54
xfs_warn_experimental(mp, XFS_EXPERIMENTAL_LARP);
fs/xfs/xfs_xattr.c
61
struct xfs_mount *mp)
fs/xfs/xfs_xattr.c
65
return xfs_has_crc(mp) && xfs_globals.larp;
fs/xfs/xfs_xattr.c
82
struct xfs_mount *mp = args->dp->i_mount;
fs/xfs/xfs_xattr.c
85
if (xfs_is_shutdown(mp))
fs/xfs/xfs_xattr.c
99
if (xfs_attr_want_log_assist(mp)) {
fs/xfs/xfs_zone_alloc.c
1001
xfs_warn(mp, "mismatched zone start: 0x%llx/0x%llx.",
fs/xfs/xfs_zone_alloc.c
1006
if (!xfs_validate_blk_zone(mp, &zone, rtg_rgno(rtg),
fs/xfs/xfs_zone_alloc.c
1037
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_alloc.c
1038
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
1043
xfs_warn(mp, "zone %u has invalid write pointer (0x%x).",
fs/xfs/xfs_zone_alloc.c
1049
xfs_warn(mp,
fs/xfs/xfs_zone_alloc.c
1056
xfs_warn(mp,
fs/xfs/xfs_zone_alloc.c
1063
xfs_warn(mp, "empty zone %u has non-zero used counter (0x%x).",
fs/xfs/xfs_zone_alloc.c
1112
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1121
max_open_data_zones = (mp->m_sb.sb_rgcount - XFS_MIN_ZONES) / 2 + 1;
fs/xfs/xfs_zone_alloc.c
1129
max_open = min(max_open, mp->m_sb.sb_rgcount / 4);
fs/xfs/xfs_zone_alloc.c
1150
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1152
struct block_device *bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_zone_alloc.c
1155
if (!mp->m_max_open_zones) {
fs/xfs/xfs_zone_alloc.c
1157
mp->m_max_open_zones = bdev_open_zones;
fs/xfs/xfs_zone_alloc.c
1159
mp->m_max_open_zones = XFS_DEFAULT_MAX_OPEN_ZONES;
fs/xfs/xfs_zone_alloc.c
1162
if (mp->m_max_open_zones < XFS_MIN_OPEN_ZONES) {
fs/xfs/xfs_zone_alloc.c
1163
xfs_notice(mp, "need at least %u open zones.",
fs/xfs/xfs_zone_alloc.c
1168
if (bdev_open_zones && bdev_open_zones < mp->m_max_open_zones) {
fs/xfs/xfs_zone_alloc.c
1169
mp->m_max_open_zones = bdev_open_zones;
fs/xfs/xfs_zone_alloc.c
1170
xfs_info(mp, "limiting open zones to %u due to hardware limit.\n",
fs/xfs/xfs_zone_alloc.c
1174
if (mp->m_max_open_zones > xfs_max_open_zones(mp)) {
fs/xfs/xfs_zone_alloc.c
1175
mp->m_max_open_zones = xfs_max_open_zones(mp);
fs/xfs/xfs_zone_alloc.c
1176
xfs_info(mp,
fs/xfs/xfs_zone_alloc.c
1178
mp->m_max_open_zones, mp->m_sb.sb_rgcount);
fs/xfs/xfs_zone_alloc.c
1186
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1188
return kvmalloc_array(BITS_TO_LONGS(mp->m_sb.sb_rgcount),
fs/xfs/xfs_zone_alloc.c
1194
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1210
zi->zi_used_bucket_bitmap[i] = xfs_alloc_bucket_bitmap(mp);
fs/xfs/xfs_zone_alloc.c
1237
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1240
.zone_capacity = mp->m_groups[XG_TYPE_RTG].blocks,
fs/xfs/xfs_zone_alloc.c
1241
.zone_size = xfs_rtgroup_raw_size(mp),
fs/xfs/xfs_zone_alloc.c
1246
if (!mp->m_rtdev_targp) {
fs/xfs/xfs_zone_alloc.c
1247
xfs_notice(mp, "RT device missing.");
fs/xfs/xfs_zone_alloc.c
1251
if (!xfs_has_rtgroups(mp) || !xfs_has_rmapbt(mp)) {
fs/xfs/xfs_zone_alloc.c
1252
xfs_notice(mp, "invalid flag combination.");
fs/xfs/xfs_zone_alloc.c
1255
if (mp->m_sb.sb_rextsize != 1) {
fs/xfs/xfs_zone_alloc.c
1256
xfs_notice(mp, "zoned file systems do not support rextsize.");
fs/xfs/xfs_zone_alloc.c
1259
if (mp->m_sb.sb_rgcount < XFS_MIN_ZONES) {
fs/xfs/xfs_zone_alloc.c
1260
xfs_notice(mp,
fs/xfs/xfs_zone_alloc.c
1265
error = xfs_calc_open_zones(mp);
fs/xfs/xfs_zone_alloc.c
1269
mp->m_zone_info = xfs_alloc_zone_info(mp);
fs/xfs/xfs_zone_alloc.c
127
if (zi->zi_gc_thread && xfs_zoned_need_gc(mp))
fs/xfs/xfs_zone_alloc.c
1270
if (!mp->m_zone_info)
fs/xfs/xfs_zone_alloc.c
1273
xfs_info(mp, "%u zones of %u blocks (%u max open zones)",
fs/xfs/xfs_zone_alloc.c
1274
mp->m_sb.sb_rgcount, iz.zone_capacity, mp->m_max_open_zones);
fs/xfs/xfs_zone_alloc.c
1275
trace_xfs_zones_mount(mp);
fs/xfs/xfs_zone_alloc.c
1297
mp->m_super->s_min_writeback_pages =
fs/xfs/xfs_zone_alloc.c
1298
XFS_FSB_TO_B(mp, min(iz.zone_capacity, XFS_MAX_BMBT_EXTLEN)) >>
fs/xfs/xfs_zone_alloc.c
1301
while ((rtg = xfs_rtgroup_next(mp, rtg))) {
fs/xfs/xfs_zone_alloc.c
1313
xfs_set_freecounter(mp, XC_FREE_RTAVAILABLE, iz.available);
fs/xfs/xfs_zone_alloc.c
1314
xfs_set_freecounter(mp, XC_FREE_RTEXTENTS,
fs/xfs/xfs_zone_alloc.c
1322
mp->m_zonegc_low_space = 0;
fs/xfs/xfs_zone_alloc.c
1324
error = xfs_zone_gc_mount(mp);
fs/xfs/xfs_zone_alloc.c
1330
xfs_free_zone_info(mp->m_zone_info);
fs/xfs/xfs_zone_alloc.c
1336
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
1338
xfs_zone_gc_unmount(mp);
fs/xfs/xfs_zone_alloc.c
1339
xfs_free_zone_info(mp->m_zone_info);
fs/xfs/xfs_zone_alloc.c
168
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_alloc.c
169
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
199
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_zone_alloc.c
203
trace_xfs_zone_record_blocks(oz, xfs_rtb_to_rgbno(mp, fsbno), len);
fs/xfs/xfs_zone_alloc.c
326
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
327
xfs_fileoff_t end_fsb = XFS_B_TO_FSB(mp, offset + count);
fs/xfs/xfs_zone_alloc.c
329
.br_startoff = XFS_B_TO_FSBT(mp, offset),
fs/xfs/xfs_zone_alloc.c
330
.br_startblock = xfs_daddr_to_rtb(mp, daddr),
fs/xfs/xfs_zone_alloc.c
334
XFS_EXTENTADD_SPACE_RES(mp, XFS_DATA_FORK);
fs/xfs/xfs_zone_alloc.c
338
if (xfs_is_shutdown(mp))
fs/xfs/xfs_zone_alloc.c
344
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write, resblks, 0,
fs/xfs/xfs_zone_alloc.c
381
struct xfs_mount *mp = tp->t_mountp;
fs/xfs/xfs_zone_alloc.c
387
xfs_err(mp,
fs/xfs/xfs_zone_alloc.c
392
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
fs/xfs/xfs_zone_alloc.c
396
trace_xfs_zone_free_blocks(rtg, xfs_rtb_to_rgbno(mp, fsbno), len);
fs/xfs/xfs_zone_alloc.c
406
xfs_add_frextents(mp, len);
fs/xfs/xfs_zone_alloc.c
445
struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.c
449
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
450
XA_STATE (xas, &mp->m_groups[XG_TYPE_RTG].xa, 0);
fs/xfs/xfs_zone_alloc.c
476
struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.c
479
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
482
if (zi->zi_nr_open_zones >= mp->m_max_open_zones - XFS_OPEN_GC_ZONES)
fs/xfs/xfs_zone_alloc.c
49
struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.c
494
oz = xfs_open_zone(mp, write_hint, false);
fs/xfs/xfs_zone_alloc.c
510
if (xfs_zoned_need_gc(mp))
fs/xfs/xfs_zone_alloc.c
53
mp->m_groups[XG_TYPE_RTG].blocks;
fs/xfs/xfs_zone_alloc.c
646
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
648
XFS_FSB_TO_B(mp, mp->m_groups[XG_TYPE_RTG].blocks);
fs/xfs/xfs_zone_alloc.c
663
struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.c
667
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
670
if (xfs_is_shutdown(mp))
fs/xfs/xfs_zone_alloc.c
693
oz = xfs_try_open_zone(mp, write_hint);
fs/xfs/xfs_zone_alloc.c
715
struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.c
719
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
723
oz = xfs_select_zone_nowait(mp, write_hint, pack_tight);
fs/xfs/xfs_zone_alloc.c
729
oz = xfs_select_zone_nowait(mp, write_hint, pack_tight);
fs/xfs/xfs_zone_alloc.c
730
if (oz || xfs_is_shutdown(mp))
fs/xfs/xfs_zone_alloc.c
746
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_alloc.c
763
*is_seq = bdev_zone_is_seq(mp->m_rtdev_targp->bt_bdev, *sector);
fs/xfs/xfs_zone_alloc.c
765
*sector += XFS_FSB_TO_BB(mp, allocated);
fs/xfs/xfs_zone_alloc.c
766
return XFS_FSB_TO_B(mp, count_fsb);
fs/xfs/xfs_zone_alloc.c
773
struct xfs_mount *mp = XFS_I(ioend->io_inode)->i_mount;
fs/xfs/xfs_zone_alloc.c
776
if (xfs_rtb_to_rgbno(mp, xfs_daddr_to_rtb(mp, sector)) == 0)
fs/xfs/xfs_zone_alloc.c
82
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_alloc.c
83
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_alloc.c
857
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_alloc.c
86
uint32_t from_bucket = xfs_zone_bucket(mp, used + freed);
fs/xfs/xfs_zone_alloc.c
864
if (xfs_is_shutdown(mp))
fs/xfs/xfs_zone_alloc.c
87
uint32_t to_bucket = xfs_zone_bucket(mp, used);
fs/xfs/xfs_zone_alloc.c
876
*oz = xfs_select_zone(mp, write_hint, pack_tight);
fs/xfs/xfs_zone_alloc.c
882
alloc_len = xfs_zone_alloc_blocks(*oz, XFS_B_TO_FSB(mp, ioend->io_size),
fs/xfs/xfs_zone_alloc.c
915
struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.c
923
if ((mp->m_super->s_flags & SB_ACTIVE) && mp->m_zone_info)
fs/xfs/xfs_zone_alloc.c
924
wake_up_all(&mp->m_zone_info->zi_zone_wait);
fs/xfs/xfs_zone_alloc.c
989
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_alloc.c
990
struct block_device *bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_zone_alloc.h
26
int xfs_zoned_space_reserve(struct xfs_mount *mp, xfs_filblks_t count_fsb,
fs/xfs/xfs_zone_alloc.h
28
void xfs_zoned_space_unreserve(struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.h
30
void xfs_zoned_add_available(struct xfs_mount *mp, xfs_filblks_t count_fsb);
fs/xfs/xfs_zone_alloc.h
41
void xfs_zoned_wake_all(struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
45
uint64_t xfs_zoned_default_resblks(struct xfs_mount *mp,
fs/xfs/xfs_zone_alloc.h
47
void xfs_zoned_show_stats(struct seq_file *m, struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
50
int xfs_mount_zones(struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
51
void xfs_unmount_zones(struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
52
void xfs_zone_gc_start(struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
53
void xfs_zone_gc_stop(struct xfs_mount *mp);
fs/xfs/xfs_zone_alloc.h
55
static inline int xfs_mount_zones(struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.h
59
static inline void xfs_unmount_zones(struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.h
62
static inline void xfs_zone_gc_start(struct xfs_mount *mp)
fs/xfs/xfs_zone_alloc.h
65
static inline void xfs_zone_gc_stop(struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
1002
if (xfs_is_shutdown(data->mp))
fs/xfs/xfs_zone_gc.c
1007
oz = xfs_zone_gc_ensure_target(data->mp);
fs/xfs/xfs_zone_gc.c
1014
if (!xfs_zoned_need_gc(data->mp))
fs/xfs/xfs_zone_gc.c
1034
struct xfs_zone_info *zi = data->mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
1092
struct xfs_mount *mp = data->mp;
fs/xfs/xfs_zone_gc.c
1093
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
1101
xfs_set_zonegc_running(mp);
fs/xfs/xfs_zone_gc.c
1119
xfs_clear_zonegc_running(mp);
fs/xfs/xfs_zone_gc.c
1120
xfs_zoned_resv_wake_all(mp);
fs/xfs/xfs_zone_gc.c
1136
xfs_clear_zonegc_running(mp);
fs/xfs/xfs_zone_gc.c
1148
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
1150
if (xfs_has_zoned(mp))
fs/xfs/xfs_zone_gc.c
1151
kthread_unpark(mp->m_zone_info->zi_gc_thread);
fs/xfs/xfs_zone_gc.c
1156
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
1158
if (xfs_has_zoned(mp))
fs/xfs/xfs_zone_gc.c
1159
kthread_park(mp->m_zone_info->zi_gc_thread);
fs/xfs/xfs_zone_gc.c
1164
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
1166
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
1179
if (!xfs_group_marked(mp, XG_TYPE_RTG, XFS_RTG_FREE) ||
fs/xfs/xfs_zone_gc.c
1180
zi->zi_nr_open_zones == mp->m_max_open_zones)
fs/xfs/xfs_zone_gc.c
1183
oz = xfs_open_zone(mp, WRITE_LIFE_NOT_SET, true);
fs/xfs/xfs_zone_gc.c
1185
xfs_warn(mp, "unable to allocate a zone for gc");
fs/xfs/xfs_zone_gc.c
1193
data = xfs_zone_gc_data_alloc(mp);
fs/xfs/xfs_zone_gc.c
1200
"xfs-zone-gc/%s", mp->m_super->s_id);
fs/xfs/xfs_zone_gc.c
1202
xfs_warn(mp, "unable to create zone gc thread");
fs/xfs/xfs_zone_gc.c
1221
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
1223
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
127
struct xfs_mount *mp;
fs/xfs/xfs_zone_gc.c
168
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
173
if (!xfs_zoned_have_reclaimable(mp->m_zone_info))
fs/xfs/xfs_zone_gc.c
176
available = xfs_estimate_freecounter(mp, XC_FREE_RTAVAILABLE);
fs/xfs/xfs_zone_gc.c
179
xfs_rtgs_to_rfsbs(mp, mp->m_max_open_zones - XFS_OPEN_GC_ZONES))
fs/xfs/xfs_zone_gc.c
182
free = xfs_estimate_freecounter(mp, XC_FREE_RTEXTENTS);
fs/xfs/xfs_zone_gc.c
185
threshold = threshold * mp->m_zonegc_low_space +
fs/xfs/xfs_zone_gc.c
186
remainder * div_s64(mp->m_zonegc_low_space, 100);
fs/xfs/xfs_zone_gc.c
196
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
222
data->mp = mp;
fs/xfs/xfs_zone_gc.c
303
struct xfs_mount *mp,
fs/xfs/xfs_zone_gc.c
324
tp = xfs_trans_alloc_empty(mp);
fs/xfs/xfs_zone_gc.c
366
struct xfs_mount *mp,
fs/xfs/xfs_zone_gc.c
379
error = xfs_zone_gc_query(mp, iter);
fs/xfs/xfs_zone_gc.c
387
error = xfs_iget(mp, NULL, irec->rm_owner,
fs/xfs/xfs_zone_gc.c
410
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_gc.c
430
struct xfs_mount *mp,
fs/xfs/xfs_zone_gc.c
433
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
442
mp->m_sb.sb_rgcount) {
fs/xfs/xfs_zone_gc.c
443
struct xfs_rtgroup *rtg = xfs_rtgroup_grab(mp, bit);
fs/xfs/xfs_zone_gc.c
497
struct xfs_mount *mp = data->mp;
fs/xfs/xfs_zone_gc.c
498
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
504
victim_rtg = xfs_zone_gc_pick_victim_from(mp, bucket);
fs/xfs/xfs_zone_gc.c
542
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
544
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
554
mp->m_max_open_zones - XFS_OPEN_GC_ZONES);
fs/xfs/xfs_zone_gc.c
555
oz = xfs_open_zone(mp, WRITE_LIFE_NOT_SET, true);
fs/xfs/xfs_zone_gc.c
572
struct xfs_mount *mp)
fs/xfs/xfs_zone_gc.c
574
struct xfs_open_zone *oz = mp->m_zone_info->zi_open_gc_zone;
fs/xfs/xfs_zone_gc.c
577
return xfs_zone_gc_select_target(mp);
fs/xfs/xfs_zone_gc.c
590
wake_up_process(data->mp->m_zone_info->zi_gc_thread);
fs/xfs/xfs_zone_gc.c
600
struct xfs_mount *mp = data->mp;
fs/xfs/xfs_zone_gc.c
603
oz = xfs_zone_gc_ensure_target(mp);
fs/xfs/xfs_zone_gc.c
607
*count_fsb = min(*count_fsb, XFS_B_TO_FSB(mp, data->scratch_available));
fs/xfs/xfs_zone_gc.c
616
spin_lock(&mp->m_sb_lock);
fs/xfs/xfs_zone_gc.c
620
mp->m_free[XC_FREE_RTEXTENTS].res_avail,
fs/xfs/xfs_zone_gc.c
621
mp->m_free[XC_FREE_RTAVAILABLE].res_avail);
fs/xfs/xfs_zone_gc.c
622
mp->m_free[XC_FREE_RTEXTENTS].res_avail -= *count_fsb;
fs/xfs/xfs_zone_gc.c
623
mp->m_free[XC_FREE_RTAVAILABLE].res_avail -= *count_fsb;
fs/xfs/xfs_zone_gc.c
624
spin_unlock(&mp->m_sb_lock);
fs/xfs/xfs_zone_gc.c
630
*is_seq = bdev_zone_is_seq(mp->m_rtdev_targp->bt_bdev, *daddr);
fs/xfs/xfs_zone_gc.c
632
*daddr += XFS_FSB_TO_BB(mp, oz->oz_allocated);
fs/xfs/xfs_zone_gc.c
665
struct xfs_mount *mp = data->mp;
fs/xfs/xfs_zone_gc.c
666
struct block_device *bdev = mp->m_rtdev_targp->bt_bdev;
fs/xfs/xfs_zone_gc.c
676
if (xfs_is_shutdown(mp))
fs/xfs/xfs_zone_gc.c
679
if (!xfs_zone_gc_iter_next(mp, iter, &irec, &ip))
fs/xfs/xfs_zone_gc.c
688
len = XFS_FSB_TO_B(mp, irec.rm_blockcount);
fs/xfs/xfs_zone_gc.c
695
chunk->offset = XFS_FSB_TO_B(mp, irec.rm_offset);
fs/xfs/xfs_zone_gc.c
707
bio->bi_iter.bi_sector = xfs_rtb_to_daddr(mp, chunk->old_startblock);
fs/xfs/xfs_zone_gc.c
713
XFS_STATS_INC(mp, xs_gc_read_calls);
fs/xfs/xfs_zone_gc.c
772
data->mp->m_sb.sb_blocksize) >> SECTOR_SHIFT;
fs/xfs/xfs_zone_gc.c
794
chunk->old_startblock += XFS_B_TO_FSB(data->mp, split_len);
fs/xfs/xfs_zone_gc.c
807
struct xfs_mount *mp = chunk->ip->i_mount;
fs/xfs/xfs_zone_gc.c
811
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_gc.c
812
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_zone_gc.c
817
XFS_STATS_INC(mp, xs_gc_write_calls);
fs/xfs/xfs_zone_gc.c
818
XFS_STATS_ADD(mp, xs_gc_bytes, chunk->len);
fs/xfs/xfs_zone_gc.c
836
struct xfs_mount *mp = ip->i_mount;
fs/xfs/xfs_zone_gc.c
840
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_gc.c
841
if (xfs_is_shutdown(mp)) {
fs/xfs/xfs_zone_gc.c
872
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_gc.c
881
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_gc.c
882
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_gc.c
885
xfs_force_shutdown(mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_gc.c
892
xfs_zoned_add_available(mp, rtg_blocks(rtg));
fs/xfs/xfs_zone_gc.c
905
struct xfs_mount *mp = rtg_mount(rtg);
fs/xfs/xfs_zone_gc.c
911
if (XFS_TEST_ERROR(mp, XFS_ERRTAG_ZONE_RESET)) {
fs/xfs/xfs_zone_gc.c
916
XFS_STATS_INC(mp, xs_gc_zone_reset_calls);
fs/xfs/xfs_zone_gc.c
930
bio->bi_iter.bi_size = XFS_FSB_TO_B(mp, rtg_blocks(rtg));
fs/xfs/xfs_zone_gc.c
968
if (blkdev_issue_flush(data->mp->m_rtdev_targp->bt_bdev) < 0) {
fs/xfs/xfs_zone_gc.c
969
xfs_force_shutdown(data->mp, SHUTDOWN_META_IO_ERROR);
fs/xfs/xfs_zone_info.c
104
xfs_show_full_zone_used_distribution(m, mp);
fs/xfs/xfs_zone_info.c
43
struct xfs_mount *mp)
fs/xfs/xfs_zone_info.c
45
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_info.c
60
full = mp->m_sb.sb_rgcount;
fs/xfs/xfs_zone_info.c
73
struct xfs_mount *mp)
fs/xfs/xfs_zone_info.c
75
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_info.c
81
xfs_sum_freecounter(mp, XC_FREE_RTEXTENTS));
fs/xfs/xfs_zone_info.c
83
mp->m_free[XC_FREE_RTEXTENTS].res_avail);
fs/xfs/xfs_zone_info.c
85
xfs_sum_freecounter(mp, XC_FREE_RTAVAILABLE));
fs/xfs/xfs_zone_info.c
87
mp->m_free[XC_FREE_RTAVAILABLE].res_avail);
fs/xfs/xfs_zone_info.c
91
xfs_zoned_need_gc(mp));
fs/xfs/xfs_zone_priv.h
110
struct xfs_open_zone *xfs_open_zone(struct xfs_mount *mp,
fs/xfs/xfs_zone_priv.h
114
bool xfs_zoned_need_gc(struct xfs_mount *mp);
fs/xfs/xfs_zone_priv.h
116
int xfs_zone_gc_mount(struct xfs_mount *mp);
fs/xfs/xfs_zone_priv.h
117
void xfs_zone_gc_unmount(struct xfs_mount *mp);
fs/xfs/xfs_zone_priv.h
119
void xfs_zoned_resv_wake_all(struct xfs_mount *mp);
fs/xfs/xfs_zone_space_resv.c
108
struct xfs_mount *mp)
fs/xfs/xfs_zone_space_resv.c
110
if (xfs_is_shutdown(mp))
fs/xfs/xfs_zone_space_resv.c
119
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
123
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_space_resv.c
141
error = xfs_dec_freecounter(mp, XC_FREE_RTAVAILABLE, count_fsb,
fs/xfs/xfs_zone_space_resv.c
152
while ((error = xfs_zoned_space_wait_error(mp)) == 0) {
fs/xfs/xfs_zone_space_resv.c
155
error = xfs_dec_freecounter(mp, XC_FREE_RTAVAILABLE, count_fsb,
fs/xfs/xfs_zone_space_resv.c
167
if (!xfs_is_zonegc_running(mp))
fs/xfs/xfs_zone_space_resv.c
175
if (!xfs_zoned_have_reclaimable(mp->m_zone_info) &&
fs/xfs/xfs_zone_space_resv.c
176
!xfs_is_zonegc_running(mp))
fs/xfs/xfs_zone_space_resv.c
199
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
203
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_space_resv.c
208
len = min(len, xfs_sum_freecounter(mp, XC_FREE_RTEXTENTS));
fs/xfs/xfs_zone_space_resv.c
211
error = xfs_dec_freecounter(mp, XC_FREE_RTEXTENTS, *count_fsb,
fs/xfs/xfs_zone_space_resv.c
220
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
230
error = xfs_dec_freecounter(mp, XC_FREE_RTEXTENTS, count_fsb,
fs/xfs/xfs_zone_space_resv.c
233
xfs_inodegc_flush(mp);
fs/xfs/xfs_zone_space_resv.c
234
error = xfs_dec_freecounter(mp, XC_FREE_RTEXTENTS, count_fsb,
fs/xfs/xfs_zone_space_resv.c
238
error = xfs_zoned_reserve_extents_greedy(mp, &count_fsb, flags);
fs/xfs/xfs_zone_space_resv.c
242
error = xfs_zoned_reserve_available(mp, count_fsb, flags);
fs/xfs/xfs_zone_space_resv.c
244
xfs_add_freecounter(mp, XC_FREE_RTEXTENTS, count_fsb);
fs/xfs/xfs_zone_space_resv.c
253
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
257
xfs_zoned_add_available(mp, ac->reserved_blocks);
fs/xfs/xfs_zone_space_resv.c
258
xfs_add_freecounter(mp, XC_FREE_RTEXTENTS, ac->reserved_blocks);
fs/xfs/xfs_zone_space_resv.c
52
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
57
return xfs_rtgs_to_rfsbs(mp, XFS_RESERVED_ZONES) +
fs/xfs/xfs_zone_space_resv.c
58
mp->m_sb.sb_rtreserved;
fs/xfs/xfs_zone_space_resv.c
60
return xfs_rtgs_to_rfsbs(mp, XFS_GC_ZONES);
fs/xfs/xfs_zone_space_resv.c
69
struct xfs_mount *mp)
fs/xfs/xfs_zone_space_resv.c
71
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_space_resv.c
82
struct xfs_mount *mp,
fs/xfs/xfs_zone_space_resv.c
85
struct xfs_zone_info *zi = mp->m_zone_info;
fs/xfs/xfs_zone_space_resv.c
89
xfs_add_freecounter(mp, XC_FREE_RTAVAILABLE, count_fsb);
fs/xfs/xfs_zone_space_resv.c
94
xfs_add_freecounter(mp, XC_FREE_RTAVAILABLE, count_fsb);
fs/xfs/xfs_zone_space_resv.c
95
count_fsb = xfs_sum_freecounter(mp, XC_FREE_RTAVAILABLE);
include/linux/if_vlan.h
211
struct vlan_priority_tci_mapping *mp;
include/linux/if_vlan.h
215
mp = vlan_dev_priv(dev)->egress_priority_map[(skprio & 0xF)];
include/linux/if_vlan.h
216
while (mp) {
include/linux/if_vlan.h
217
if (mp->priority == skprio) {
include/linux/if_vlan.h
218
return mp->vlan_qos; /* This should already be shifted
include/linux/if_vlan.h
222
mp = mp->next;
include/linux/module.h
50
struct module_param_attrs *mp;
include/media/drv-intf/exynos-fimc.h
142
struct media_pipeline mp;
include/uapi/linux/hdlcdrv.h
54
struct hdlcdrv_params mp;
io_uring/mock_file.c
281
struct io_uring_mock_probe mp, __user *uarg;
io_uring/mock_file.c
288
uarg_size != sizeof(mp))
io_uring/mock_file.c
291
memset(&mp, 0, sizeof(mp));
io_uring/mock_file.c
292
if (copy_from_user(&mp, uarg, uarg_size))
io_uring/mock_file.c
294
if (!mem_is_zero(&mp, sizeof(mp)))
io_uring/mock_file.c
297
mp.features = IORING_MOCK_FEAT_END;
io_uring/mock_file.c
299
if (copy_to_user(uarg, &mp, uarg_size))
kernel/debug/kdb/kdb_main.c
679
kdbtab_t *mp;
kernel/debug/kdb/kdb_main.c
714
mp = &kdb_macro->cmd;
kernel/debug/kdb/kdb_main.c
715
mp->func = kdb_exec_defcmd;
kernel/debug/kdb/kdb_main.c
716
mp->minlen = 0;
kernel/debug/kdb/kdb_main.c
717
mp->flags = KDB_ENABLE_ALWAYS_SAFE;
kernel/debug/kdb/kdb_main.c
718
mp->name = kdb_strdup(argv[1], GFP_KDB);
kernel/debug/kdb/kdb_main.c
719
if (!mp->name)
kernel/debug/kdb/kdb_main.c
721
mp->usage = kdb_strdup_dequote(argv[2], GFP_KDB);
kernel/debug/kdb/kdb_main.c
722
if (!mp->usage)
kernel/debug/kdb/kdb_main.c
724
mp->help = kdb_strdup_dequote(argv[3], GFP_KDB);
kernel/debug/kdb/kdb_main.c
725
if (!mp->help)
kernel/debug/kdb/kdb_main.c
732
kfree(mp->usage);
kernel/debug/kdb/kdb_main.c
734
kfree(mp->name);
kernel/params.c
634
if (!mk->mp) {
kernel/params.c
636
mk->mp = kzalloc_obj(*mk->mp);
kernel/params.c
637
if (!mk->mp)
kernel/params.c
639
mk->mp->grp.name = "parameters";
kernel/params.c
641
mk->mp->grp.attrs = kzalloc_obj(mk->mp->grp.attrs[0]);
kernel/params.c
643
if (!mk->mp->grp.attrs)
kernel/params.c
648
new_mp = krealloc(mk->mp, struct_size(mk->mp, attrs, mk->mp->num + 1),
kernel/params.c
652
mk->mp = new_mp;
kernel/params.c
653
mk->mp->num++;
kernel/params.c
656
new_attrs = krealloc_array(mk->mp->grp.attrs, mk->mp->num + 1,
kernel/params.c
657
sizeof(mk->mp->grp.attrs[0]), GFP_KERNEL);
kernel/params.c
660
mk->mp->grp.attrs = new_attrs;
kernel/params.c
663
memset(&mk->mp->attrs[mk->mp->num - 1], 0, sizeof(mk->mp->attrs[0]));
kernel/params.c
664
sysfs_attr_init(&mk->mp->attrs[mk->mp->num - 1].mattr.attr);
kernel/params.c
665
mk->mp->attrs[mk->mp->num - 1].param = kp;
kernel/params.c
666
mk->mp->attrs[mk->mp->num - 1].mattr.show = param_attr_show;
kernel/params.c
669
mk->mp->attrs[mk->mp->num - 1].mattr.store = param_attr_store;
kernel/params.c
671
mk->mp->attrs[mk->mp->num - 1].mattr.store = NULL;
kernel/params.c
672
mk->mp->attrs[mk->mp->num - 1].mattr.attr.name = (char *)name;
kernel/params.c
673
mk->mp->attrs[mk->mp->num - 1].mattr.attr.mode = kp->perm;
kernel/params.c
676
for (i = 0; i < mk->mp->num; i++)
kernel/params.c
677
mk->mp->grp.attrs[i] = &mk->mp->attrs[i].mattr.attr;
kernel/params.c
678
mk->mp->grp.attrs[mk->mp->num] = NULL;
kernel/params.c
685
if (mk->mp)
kernel/params.c
686
kfree(mk->mp->grp.attrs);
kernel/params.c
687
kfree(mk->mp);
kernel/params.c
688
mk->mp = NULL;
kernel/params.c
722
err = sysfs_create_group(&mod->mkobj.kobj, &mod->mkobj.mp->grp);
kernel/params.c
737
if (mod->mkobj.mp) {
kernel/params.c
738
sysfs_remove_group(&mod->mkobj.kobj, &mod->mkobj.mp->grp);
kernel/params.c
802
if (mk->mp)
kernel/params.c
803
sysfs_remove_group(&mk->kobj, &mk->mp->grp);
kernel/params.c
808
err = sysfs_create_group(&mk->kobj, &mk->mp->grp);
lib/crypto/mpi/mpi-pow.c
110
if (rp == ep || rp == mp || rp == bp) {
lib/crypto/mpi/mpi-pow.c
136
if (rp == mp) {
lib/crypto/mpi/mpi-pow.c
139
mp = mp_marker = mpi_alloc_limb_space(msize);
lib/crypto/mpi/mpi-pow.c
140
if (!mp)
lib/crypto/mpi/mpi-pow.c
142
MPN_COPY(mp, rp, msize);
lib/crypto/mpi/mpi-pow.c
207
mp, msize);
lib/crypto/mpi/mpi-pow.c
232
xp, xsize, mp,
lib/crypto/mpi/mpi-pow.c
272
mpihelp_divrem(rp + msize, 0, rp, rsize, mp, msize);
lib/crypto/mpi/mpi-pow.c
284
mpihelp_rshift(mp, mp, msize, mod_shift_cnt);
lib/crypto/mpi/mpi-pow.c
285
mpihelp_sub(rp, mp, msize, rp, rsize);
lib/crypto/mpi/mpi-pow.c
32
mpi_ptr_t rp, ep, mp, bp;
lib/crypto/mpi/mpi-pow.c
72
mp = mp_marker = mpi_alloc_limb_space(msize);
lib/crypto/mpi/mpi-pow.c
73
if (!mp)
lib/crypto/mpi/mpi-pow.c
77
mpihelp_lshift(mp, mod->d, msize, mod_shift_cnt);
lib/crypto/mpi/mpi-pow.c
79
MPN_COPY(mp, mod->d, msize);
lib/crypto/mpi/mpi-pow.c
92
mpihelp_divrem(bp + msize, 0, bp, bsize, mp, msize);
mm/cma.c
342
struct list_head *mp;
mm/cma.c
348
list_for_each(mp, ranges) {
mm/cma.c
349
mlp = list_entry(mp, struct cma_init_memrange, list);
mm/cma.c
538
struct list_head *mp, *next;
mm/cma.c
647
list_for_each_safe(mp, next, &ranges) {
mm/cma.c
648
mlp = list_entry(mp, struct cma_init_memrange, list);
mm/cma.c
649
list_del(mp);
mm/cma.c
662
list_for_each(mp, &final_ranges) {
mm/cma.c
663
mlp = list_entry(mp, struct cma_init_memrange, list);
mm/cma.c
688
list_for_each(mp, &final_ranges) {
mm/cma.c
689
mlp = list_entry(mp, struct cma_init_memrange, list);
net/8021q/vlan_dev.c
175
struct vlan_priority_tci_mapping *mp = NULL;
net/8021q/vlan_dev.c
180
mp = vlan->egress_priority_map[skb_prio & 0xF];
net/8021q/vlan_dev.c
181
while (mp) {
net/8021q/vlan_dev.c
182
if (mp->priority == skb_prio) {
net/8021q/vlan_dev.c
183
if (mp->vlan_qos && !vlan_qos)
net/8021q/vlan_dev.c
185
else if (!mp->vlan_qos && vlan_qos)
net/8021q/vlan_dev.c
187
mp->vlan_qos = vlan_qos;
net/8021q/vlan_dev.c
190
mp = mp->next;
net/8021q/vlan_dev.c
194
mp = vlan->egress_priority_map[skb_prio & 0xF];
net/8021q/vlan_dev.c
199
np->next = mp;
net/8021q/vlanproc.c
266
const struct vlan_priority_tci_mapping *mp
net/8021q/vlanproc.c
268
while (mp) {
net/8021q/vlanproc.c
270
mp->priority, ((mp->vlan_qos >> 13) & 0x7));
net/8021q/vlanproc.c
271
mp = mp->next;
net/bridge/br_mdb.c
1003
return br_mdb_replace_group_star_g(cfg, mp, p, brmctx,
net/bridge/br_mdb.c
1025
br_mdb_notify(cfg->br->dev, mp, p, RTM_NEWMDB);
net/bridge/br_mdb.c
1045
struct net_bridge_mdb_entry *mp;
net/bridge/br_mdb.c
1055
mp = br_multicast_new_group(br, &group);
net/bridge/br_mdb.c
1056
if (IS_ERR(mp))
net/bridge/br_mdb.c
1057
return PTR_ERR(mp);
net/bridge/br_mdb.c
1061
if (mp->host_joined && !(cfg->nlflags & NLM_F_REPLACE)) {
net/bridge/br_mdb.c
1066
br_multicast_host_join(brmctx, mp, false);
net/bridge/br_mdb.c
1067
br_mdb_notify(br->dev, mp, NULL, RTM_NEWMDB);
net/bridge/br_mdb.c
1076
return br_mdb_add_group_star_g(cfg, mp, brmctx, flags, extack);
net/bridge/br_mdb.c
1078
return br_mdb_add_group_sg(cfg, mp, brmctx, flags, extack);
net/bridge/br_mdb.c
1362
struct net_bridge_mdb_entry *mp;
net/bridge/br_mdb.c
1369
mp = br_mdb_ip_get(br, &ip);
net/bridge/br_mdb.c
1370
if (!mp)
net/bridge/br_mdb.c
1374
if (entry->ifindex == mp->br->dev->ifindex && mp->host_joined) {
net/bridge/br_mdb.c
1375
br_multicast_host_leave(mp, false);
net/bridge/br_mdb.c
1377
br_mdb_notify(br->dev, mp, NULL, RTM_DELMDB);
net/bridge/br_mdb.c
1378
if (!mp->ports && netif_running(br->dev))
net/bridge/br_mdb.c
1379
mod_timer(&mp->timer, jiffies);
net/bridge/br_mdb.c
1383
for (pp = &mp->ports;
net/bridge/br_mdb.c
1389
br_multicast_del_pg(mp, p, pp);
net/bridge/br_mdb.c
1477
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
1488
state = br_group_is_l2(&mp->addr) ? MDB_PERMANENT : 0;
net/bridge/br_mdb.c
1492
br_multicast_host_leave(mp, true);
net/bridge/br_mdb.c
1493
if (!mp->ports && netif_running(br->dev))
net/bridge/br_mdb.c
1494
mod_timer(&mp->timer, jiffies);
net/bridge/br_mdb.c
1498
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
1504
for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;) {
net/bridge/br_mdb.c
1525
br_multicast_del_pg(mp, p, pp);
net/bridge/br_mdb.c
1532
struct net_bridge_mdb_entry *mp;
net/bridge/br_mdb.c
1539
hlist_for_each_entry(mp, &br->mdb_list, mdb_node) {
net/bridge/br_mdb.c
1540
if (desc->vid && desc->vid != mp->addr.vid)
net/bridge/br_mdb.c
1543
br_mdb_flush_host(br, mp, desc);
net/bridge/br_mdb.c
1544
br_mdb_flush_pgs(br, mp, desc);
net/bridge/br_mdb.c
1601
br_mdb_get_reply_alloc(const struct net_bridge_mdb_entry *mp)
net/bridge/br_mdb.c
1612
if (mp->host_joined)
net/bridge/br_mdb.c
1615
for (pg = mlock_dereference(mp->ports, mp->br); pg;
net/bridge/br_mdb.c
1616
pg = mlock_dereference(pg->next, mp->br))
net/bridge/br_mdb.c
1623
struct net_bridge_mdb_entry *mp, u32 portid,
net/bridge/br_mdb.c
1639
bpm->ifindex = mp->br->dev->ifindex;
net/bridge/br_mdb.c
1651
if (mp->host_joined) {
net/bridge/br_mdb.c
1652
err = __mdb_fill_info(skb, mp, NULL);
net/bridge/br_mdb.c
1657
for (pg = mlock_dereference(mp->ports, mp->br); pg;
net/bridge/br_mdb.c
1658
pg = mlock_dereference(pg->next, mp->br)) {
net/bridge/br_mdb.c
1659
err = __mdb_fill_info(skb, mp, pg);
net/bridge/br_mdb.c
1679
struct net_bridge_mdb_entry *mp;
net/bridge/br_mdb.c
1694
mp = br_mdb_ip_get(br, &group);
net/bridge/br_mdb.c
1695
if (!mp || (!mp->ports && !mp->host_joined)) {
net/bridge/br_mdb.c
1701
skb = br_mdb_get_reply_alloc(mp);
net/bridge/br_mdb.c
1707
err = br_mdb_get_reply_fill(skb, mp, portid, seq);
net/bridge/br_mdb.c
233
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
249
ifindex = mp->br->dev->ifindex;
net/bridge/br_mdb.c
250
mtimer = &mp->timer;
net/bridge/br_mdb.c
255
e.vid = mp->addr.vid;
net/bridge/br_mdb.c
256
if (mp->addr.proto == htons(ETH_P_IP)) {
net/bridge/br_mdb.c
257
e.addr.u.ip4 = mp->addr.dst.ip4;
net/bridge/br_mdb.c
259
} else if (mp->addr.proto == htons(ETH_P_IPV6)) {
net/bridge/br_mdb.c
260
e.addr.u.ip6 = mp->addr.dst.ip6;
net/bridge/br_mdb.c
263
ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
net/bridge/br_mdb.c
266
e.addr.proto = mp->addr.proto;
net/bridge/br_mdb.c
278
switch (mp->addr.proto) {
net/bridge/br_mdb.c
280
dump_srcs_mode = !!(mp->br->multicast_ctx.multicast_igmp_version == 3);
net/bridge/br_mdb.c
281
if (mp->addr.src.ip4) {
net/bridge/br_mdb.c
283
mp->addr.src.ip4))
net/bridge/br_mdb.c
290
dump_srcs_mode = !!(mp->br->multicast_ctx.multicast_mld_version == 2);
net/bridge/br_mdb.c
291
if (!ipv6_addr_any(&mp->addr.src.ip6)) {
net/bridge/br_mdb.c
293
&mp->addr.src.ip6))
net/bridge/br_mdb.c
300
ether_addr_copy(e.addr.u.mac_addr, mp->addr.dst.mac_addr);
net/bridge/br_mdb.c
325
struct net_bridge_mdb_entry *mp;
net/bridge/br_mdb.c
332
hlist_for_each_entry_rcu(mp, &br->mdb_list, mdb_node) {
net/bridge/br_mdb.c
345
if (!s_pidx && mp->host_joined) {
net/bridge/br_mdb.c
346
err = __mdb_fill_info(skb, mp, NULL);
net/bridge/br_mdb.c
353
for (pp = &mp->ports; (p = rcu_dereference(*pp)) != NULL;
net/bridge/br_mdb.c
360
err = __mdb_fill_info(skb, mp, p);
net/bridge/br_mdb.c
417
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
440
if (__mdb_fill_info(skb, mp, pg))
net/bridge/br_mdb.c
523
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
532
br_switchdev_mdb_notify(dev, mp, pg, type);
net/bridge/br_mdb.c
538
err = nlmsg_populate_mdb_fill(skb, dev, mp, pg, type);
net/bridge/br_mdb.c
551
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
555
__br_mdb_notify(dev, mp, pg, type, true);
net/bridge/br_mdb.c
559
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
562
__br_mdb_notify(dev, mp, pg, RTM_NEWMDB, false);
net/bridge/br_mdb.c
740
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
755
br_mdb_notify(cfg->br->dev, mp, pg, RTM_NEWMDB);
net/bridge/br_mdb.c
761
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
770
for (pp = &mp->ports;
net/bridge/br_mdb.c
778
return br_mdb_replace_group_sg(cfg, mp, p, brmctx,
net/bridge/br_mdb.c
794
br_mdb_notify(cfg->br->dev, mp, p, RTM_NEWMDB);
net/bridge/br_mdb.c
953
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
976
br_mdb_notify(cfg->br->dev, mp, pg, RTM_NEWMDB);
net/bridge/br_mdb.c
985
struct net_bridge_mdb_entry *mp,
net/bridge/br_mdb.c
995
for (pp = &mp->ports;
net/bridge/br_multicast.c
1279
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
1282
mp = br_mdb_ip_get(br, group);
net/bridge/br_multicast.c
1283
if (mp)
net/bridge/br_multicast.c
1284
return mp;
net/bridge/br_multicast.c
1293
mp = kzalloc_obj(*mp, GFP_ATOMIC);
net/bridge/br_multicast.c
1294
if (unlikely(!mp))
net/bridge/br_multicast.c
1297
mp->br = br;
net/bridge/br_multicast.c
1298
mp->addr = *group;
net/bridge/br_multicast.c
1299
mp->mcast_gc.destroy = br_multicast_destroy_mdb_entry;
net/bridge/br_multicast.c
1300
timer_setup(&mp->timer, br_multicast_group_expired, 0);
net/bridge/br_multicast.c
1301
err = rhashtable_lookup_insert_fast(&br->mdb_hash_tbl, &mp->rhnode,
net/bridge/br_multicast.c
1304
kfree(mp);
net/bridge/br_multicast.c
1305
mp = ERR_PTR(err);
net/bridge/br_multicast.c
1307
hlist_add_head_rcu(&mp->mdb_node, &br->mdb_list);
net/bridge/br_multicast.c
1310
return mp;
net/bridge/br_multicast.c
1473
struct net_bridge_mdb_entry *mp, bool notify)
net/bridge/br_multicast.c
1475
if (!mp->host_joined) {
net/bridge/br_multicast.c
1476
mp->host_joined = true;
net/bridge/br_multicast.c
1477
if (br_multicast_is_star_g(&mp->addr))
net/bridge/br_multicast.c
1478
br_multicast_star_g_host_state(mp);
net/bridge/br_multicast.c
1480
br_mdb_notify(mp->br->dev, mp, NULL, RTM_NEWMDB);
net/bridge/br_multicast.c
1483
if (br_group_is_l2(&mp->addr))
net/bridge/br_multicast.c
1486
mod_timer(&mp->timer, jiffies + brmctx->multicast_membership_interval);
net/bridge/br_multicast.c
1489
void br_multicast_host_leave(struct net_bridge_mdb_entry *mp, bool notify)
net/bridge/br_multicast.c
1491
if (!mp->host_joined)
net/bridge/br_multicast.c
1494
mp->host_joined = false;
net/bridge/br_multicast.c
1495
if (br_multicast_is_star_g(&mp->addr))
net/bridge/br_multicast.c
1496
br_multicast_star_g_host_state(mp);
net/bridge/br_multicast.c
1498
br_mdb_notify(mp->br->dev, mp, NULL, RTM_DELMDB);
net/bridge/br_multicast.c
1512
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
1518
mp = br_multicast_new_group(brmctx->br, group);
net/bridge/br_multicast.c
1519
if (IS_ERR(mp))
net/bridge/br_multicast.c
1520
return ERR_CAST(mp);
net/bridge/br_multicast.c
1523
br_multicast_host_join(brmctx, mp, true);
net/bridge/br_multicast.c
1527
for (pp = &mp->ports;
net/bridge/br_multicast.c
1545
br_mdb_notify(brmctx->br->dev, mp, p, RTM_NEWMDB);
net/bridge/br_multicast.c
2827
br_multicast_find_port(struct net_bridge_mdb_entry *mp,
net/bridge/br_multicast.c
2831
struct net_bridge *br __maybe_unused = mp->br;
net/bridge/br_multicast.c
2834
for (pg = mlock_dereference(mp->ports, br);
net/bridge/br_multicast.c
347
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
3491
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
353
mp = br_mdb_ip_get(br, &pg->key.addr);
net/bridge/br_multicast.c
3536
mp = br_mdb_ip4_get(brmctx->br, group, vid);
net/bridge/br_multicast.c
3537
if (!mp)
net/bridge/br_multicast.c
354
if (!mp)
net/bridge/br_multicast.c
3542
if (mp->host_joined &&
net/bridge/br_multicast.c
3543
(timer_pending(&mp->timer) ?
net/bridge/br_multicast.c
3544
time_after(mp->timer.expires, now + max_delay) :
net/bridge/br_multicast.c
3545
timer_delete_sync_try(&mp->timer) >= 0))
net/bridge/br_multicast.c
3546
mod_timer(&mp->timer, now + max_delay);
net/bridge/br_multicast.c
3548
for (pp = &mp->ports;
net/bridge/br_multicast.c
3571
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
3626
mp = br_mdb_ip6_get(brmctx->br, group, vid);
net/bridge/br_multicast.c
3627
if (!mp)
net/bridge/br_multicast.c
363
for (pg_lst = mlock_dereference(mp->ports, br);
net/bridge/br_multicast.c
3631
if (mp->host_joined &&
net/bridge/br_multicast.c
3632
(timer_pending(&mp->timer) ?
net/bridge/br_multicast.c
3633
time_after(mp->timer.expires, now + max_delay) :
net/bridge/br_multicast.c
3634
timer_delete_sync_try(&mp->timer) >= 0))
net/bridge/br_multicast.c
3635
mod_timer(&mp->timer, now + max_delay);
net/bridge/br_multicast.c
3637
for (pp = &mp->ports;
net/bridge/br_multicast.c
3662
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
3671
mp = br_mdb_ip_get(brmctx->br, group);
net/bridge/br_multicast.c
3672
if (!mp)
net/bridge/br_multicast.c
3678
for (pp = &mp->ports;
net/bridge/br_multicast.c
3688
br_multicast_del_pg(mp, p, pp);
net/bridge/br_multicast.c
3697
__br_multicast_send_query(brmctx, pmctx, NULL, NULL, &mp->addr,
net/bridge/br_multicast.c
3705
for (p = mlock_dereference(mp->ports, brmctx->br);
net/bridge/br_multicast.c
3727
if (mp->host_joined &&
net/bridge/br_multicast.c
3728
(timer_pending(&mp->timer) ?
net/bridge/br_multicast.c
3729
time_after(mp->timer.expires, time) :
net/bridge/br_multicast.c
3730
timer_delete_sync_try(&mp->timer) >= 0)) {
net/bridge/br_multicast.c
3731
mod_timer(&mp->timer, time);
net/bridge/br_multicast.c
3737
for (p = mlock_dereference(mp->ports, brmctx->br);
net/bridge/br_multicast.c
4445
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
4450
hlist_for_each_entry_safe(mp, tmp, &br->mdb_list, mdb_node)
net/bridge/br_multicast.c
4451
br_multicast_del_mdb_entry(mp);
net/bridge/br_multicast.c
559
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
566
mp = br_mdb_ip_get(src->br, &sg_ip);
net/bridge/br_multicast.c
567
if (!mp)
net/bridge/br_multicast.c
570
for (pp = &mp->ports;
net/bridge/br_multicast.c
583
br_multicast_del_pg(mp, p, pp);
net/bridge/br_multicast.c
625
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
627
mp = container_of(gc, struct net_bridge_mdb_entry, mcast_gc);
net/bridge/br_multicast.c
628
WARN_ON(!hlist_unhashed(&mp->mdb_node));
net/bridge/br_multicast.c
629
WARN_ON(mp->ports);
net/bridge/br_multicast.c
631
timer_shutdown_sync(&mp->timer);
net/bridge/br_multicast.c
632
kfree_rcu(mp, rcu);
net/bridge/br_multicast.c
635
static void br_multicast_del_mdb_entry(struct net_bridge_mdb_entry *mp)
net/bridge/br_multicast.c
637
struct net_bridge *br = mp->br;
net/bridge/br_multicast.c
639
rhashtable_remove_fast(&br->mdb_hash_tbl, &mp->rhnode,
net/bridge/br_multicast.c
641
hlist_del_init_rcu(&mp->mdb_node);
net/bridge/br_multicast.c
642
hlist_add_head(&mp->mcast_gc.gc_node, &br->mcast_gc_list);
net/bridge/br_multicast.c
648
struct net_bridge_mdb_entry *mp = timer_container_of(mp, t, timer);
net/bridge/br_multicast.c
649
struct net_bridge *br = mp->br;
net/bridge/br_multicast.c
652
if (hlist_unhashed(&mp->mdb_node) || !netif_running(br->dev) ||
net/bridge/br_multicast.c
653
timer_pending(&mp->timer))
net/bridge/br_multicast.c
656
br_multicast_host_leave(mp, true);
net/bridge/br_multicast.c
658
if (mp->ports)
net/bridge/br_multicast.c
660
br_multicast_del_mdb_entry(mp);
net/bridge/br_multicast.c
802
void br_multicast_del_pg(struct net_bridge_mdb_entry *mp,
net/bridge/br_multicast.c
815
br_mdb_notify(br->dev, mp, pg, RTM_DELMDB);
net/bridge/br_multicast.c
816
if (!br_multicast_is_star_g(&mp->addr)) {
net/bridge/br_multicast.c
819
br_multicast_sg_del_exclude_ports(mp);
net/bridge/br_multicast.c
827
if (!mp->ports && !mp->host_joined && netif_running(br->dev))
net/bridge/br_multicast.c
828
mod_timer(&mp->timer, jiffies);
net/bridge/br_multicast.c
835
struct net_bridge_mdb_entry *mp;
net/bridge/br_multicast.c
838
mp = br_mdb_ip_get(br, &pg->key.addr);
net/bridge/br_multicast.c
839
if (WARN_ON(!mp))
net/bridge/br_multicast.c
842
for (pp = &mp->ports;
net/bridge/br_multicast.c
848
br_multicast_del_pg(mp, pg, pp);
net/bridge/br_multicast.c
880
struct net_bridge_mdb_entry *mp = br_mdb_ip_get(br, &pg->key.addr);
net/bridge/br_multicast.c
885
if (WARN_ON(!mp))
net/bridge/br_multicast.c
887
br_mdb_notify(br->dev, mp, pg, RTM_NEWMDB);
net/bridge/br_private.h
1012
void br_mdb_notify(struct net_device *dev, struct net_bridge_mdb_entry *mp,
net/bridge/br_private.h
1014
void br_mdb_flag_change_notify(struct net_device *dev, struct net_bridge_mdb_entry *mp,
net/bridge/br_private.h
1018
void br_multicast_del_pg(struct net_bridge_mdb_entry *mp,
net/bridge/br_private.h
1043
struct net_bridge_mdb_entry *mp, bool notify);
net/bridge/br_private.h
1044
void br_multicast_host_leave(struct net_bridge_mdb_entry *mp, bool notify);
net/bridge/br_private.h
2232
struct net_bridge_mdb_entry *mp,
net/bridge/br_private.h
2332
struct net_bridge_mdb_entry *mp,
net/bridge/br_switchdev.c
507
struct net_bridge_mdb_entry *mp;
net/bridge/br_switchdev.c
516
mp = br_mdb_ip_get(br, &data->ip);
net/bridge/br_switchdev.c
517
if (!mp)
net/bridge/br_switchdev.c
519
for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
net/bridge/br_switchdev.c
527
br_mdb_flag_change_notify(br->dev, mp, p);
net/bridge/br_switchdev.c
536
const struct net_bridge_mdb_entry *mp)
net/bridge/br_switchdev.c
538
if (mp->addr.proto == htons(ETH_P_IP))
net/bridge/br_switchdev.c
539
ip_eth_mc_map(mp->addr.dst.ip4, mdb->addr);
net/bridge/br_switchdev.c
541
else if (mp->addr.proto == htons(ETH_P_IPV6))
net/bridge/br_switchdev.c
542
ipv6_eth_mc_map(&mp->addr.dst.ip6, mdb->addr);
net/bridge/br_switchdev.c
545
ether_addr_copy(mdb->addr, mp->addr.dst.mac_addr);
net/bridge/br_switchdev.c
547
mdb->vid = mp->addr.vid;
net/bridge/br_switchdev.c
552
struct net_bridge_mdb_entry *mp,
net/bridge/br_switchdev.c
563
br_switchdev_mdb_populate(&mdb, mp);
net/bridge/br_switchdev.c
576
struct net_bridge_mdb_entry *mp, int type)
net/bridge/br_switchdev.c
582
br_switchdev_host_mdb_one(dev, lower_dev, mp, type);
net/bridge/br_switchdev.c
609
const struct net_bridge_mdb_entry *mp,
net/bridge/br_switchdev.c
620
br_switchdev_mdb_populate(&mdb, mp);
net/bridge/br_switchdev.c
644
struct net_bridge_mdb_entry *mp,
net/bridge/br_switchdev.c
657
return br_switchdev_host_mdb(dev, mp, type);
net/bridge/br_switchdev.c
659
br_switchdev_mdb_populate(&mdb, mp);
net/bridge/br_switchdev.c
668
complete_info->ip = mp->addr;
net/bridge/br_switchdev.c
687
const struct net_bridge_mdb_entry *mp;
net/bridge/br_switchdev.c
721
hlist_for_each_entry(mp, &br->mdb_list, mdb_node) {
net/bridge/br_switchdev.c
725
if (mp->host_joined) {
net/bridge/br_switchdev.c
728
mp, br_dev);
net/bridge/br_switchdev.c
735
for (pp = &mp->ports; (p = mlock_dereference(*pp, br)) != NULL;
net/bridge/br_switchdev.c
742
mp, dev);
net/bridge/netfilter/ebt_arp.c
69
const unsigned char *mp;
net/bridge/netfilter/ebt_arp.c
75
mp = skb_header_pointer(skb, sizeof(struct arphdr),
net/bridge/netfilter/ebt_arp.c
77
if (mp == NULL)
net/bridge/netfilter/ebt_arp.c
80
!ether_addr_equal_masked(mp, info->smaddr,
net/bridge/netfilter/ebt_arp.c
86
mp = skb_header_pointer(skb, sizeof(struct arphdr) +
net/bridge/netfilter/ebt_arp.c
89
if (mp == NULL)
net/bridge/netfilter/ebt_arp.c
92
!ether_addr_equal_masked(mp, info->dmaddr,
net/ipv4/fib_semantics.c
1685
struct nlattr *mp;
net/ipv4/fib_semantics.c
1687
mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
net/ipv4/fib_semantics.c
1688
if (!mp)
net/ipv4/fib_semantics.c
1708
nla_nest_end(skb, mp);
net/ipv6/route.c
5730
struct nlattr *mp;
net/ipv6/route.c
5732
mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
net/ipv6/route.c
5733
if (!mp)
net/ipv6/route.c
5739
nla_nest_end(skb, mp);
net/ipv6/route.c
5872
struct nlattr *mp;
net/ipv6/route.c
5874
mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
net/ipv6/route.c
5875
if (!mp)
net/ipv6/route.c
5898
nla_nest_end(skb, mp);
net/key/af_key.c
3577
const struct xfrm_migrate *mp;
net/key/af_key.c
3601
for (i = 0, mp = m; i < num_bundles; i++, mp++) {
net/key/af_key.c
3604
pair_size = pfkey_sockaddr_pair_size(mp->old_family);
net/key/af_key.c
3609
pair_size = pfkey_sockaddr_pair_size(mp->new_family);
net/key/af_key.c
3652
for (i = 0, mp = m; i < num_bundles; i++, mp++) {
net/key/af_key.c
3654
int mode = pfkey_mode_from_xfrm(mp->mode);
net/key/af_key.c
3657
if (set_ipsecrequest(skb, mp->proto, mode,
net/key/af_key.c
3658
(mp->reqid ? IPSEC_LEVEL_UNIQUE : IPSEC_LEVEL_REQUIRE),
net/key/af_key.c
3659
mp->reqid, mp->old_family,
net/key/af_key.c
3660
&mp->old_saddr, &mp->old_daddr) < 0)
net/key/af_key.c
3664
if (set_ipsecrequest(skb, mp->proto, mode,
net/key/af_key.c
3665
(mp->reqid ? IPSEC_LEVEL_UNIQUE : IPSEC_LEVEL_REQUIRE),
net/key/af_key.c
3666
mp->reqid, mp->new_family,
net/key/af_key.c
3667
&mp->new_saddr, &mp->new_daddr) < 0)
net/mac80211/rc80211_minstrel_ht.c
1059
minstrel_ht_update_stats(struct minstrel_priv *mp, struct minstrel_ht_sta *mi)
net/mac80211/rc80211_minstrel_ht.c
1070
if (!ieee80211_hw_check(mp->hw, TX_STATUS_NO_AMPDU_LEN))
net/mac80211/rc80211_minstrel_ht.c
1129
minstrel_ht_calc_rate_stats(mp, mrs);
net/mac80211/rc80211_minstrel_ht.c
1184
if (mp->fixed_rate_idx != -1) {
net/mac80211/rc80211_minstrel_ht.c
1186
mi->max_tp_rate[i] = mp->fixed_rate_idx;
net/mac80211/rc80211_minstrel_ht.c
1187
mi->max_prob_rate = mp->fixed_rate_idx;
net/mac80211/rc80211_minstrel_ht.c
1197
minstrel_ht_txstat_valid(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
1212
for (i = 0; i < ARRAY_SIZE(mp->cck_rates); i++)
net/mac80211/rc80211_minstrel_ht.c
1213
if (rate->idx == mp->cck_rates[i])
net/mac80211/rc80211_minstrel_ht.c
1216
for (i = 0; i < ARRAY_SIZE(mp->ofdm_rates[0]); i++)
net/mac80211/rc80211_minstrel_ht.c
1217
if (rate->idx == mp->ofdm_rates[mi->band][i])
net/mac80211/rc80211_minstrel_ht.c
1227
minstrel_ht_ri_txstat_valid(struct minstrel_priv *mp,
net/mac80211/rc80211_minstrel_ht.c
1242
for (i = 0; i < ARRAY_SIZE(mp->cck_rates); i++) {
net/mac80211/rc80211_minstrel_ht.c
1244
minstrel_cck_bitrates[ mp->cck_rates[i] ])
net/mac80211/rc80211_minstrel_ht.c
1248
for (i = 0; i < ARRAY_SIZE(mp->ofdm_rates); i++) {
net/mac80211/rc80211_minstrel_ht.c
1250
minstrel_ofdm_bitrates[ mp->ofdm_rates[mi->band][i] ])
net/mac80211/rc80211_minstrel_ht.c
1289
struct minstrel_priv *mp = priv;
net/mac80211/rc80211_minstrel_ht.c
1290
u32 update_interval = mp->update_interval;
net/mac80211/rc80211_minstrel_ht.c
1323
last = !minstrel_ht_ri_txstat_valid(mp, mi, &(st->rates[0]));
net/mac80211/rc80211_minstrel_ht.c
1326
!minstrel_ht_ri_txstat_valid(mp, mi,
net/mac80211/rc80211_minstrel_ht.c
1329
rate = minstrel_ht_ri_get_stats(mp, mi,
net/mac80211/rc80211_minstrel_ht.c
1339
last = !minstrel_ht_txstat_valid(mp, mi, &ar[0]);
net/mac80211/rc80211_minstrel_ht.c
1342
!minstrel_ht_txstat_valid(mp, mi, &ar[i + 1]);
net/mac80211/rc80211_minstrel_ht.c
1344
rate = minstrel_ht_get_stats(mp, mi, &ar[i]);
net/mac80211/rc80211_minstrel_ht.c
1352
if (mp->hw->max_rates > 1) {
net/mac80211/rc80211_minstrel_ht.c
1374
minstrel_ht_update_stats(mp, mi);
net/mac80211/rc80211_minstrel_ht.c
1378
minstrel_ht_update_rates(mp, mi);
net/mac80211/rc80211_minstrel_ht.c
1382
minstrel_calc_retransmit(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
1387
unsigned int cw = mp->cw_min;
net/mac80211/rc80211_minstrel_ht.c
1408
cw = min((cw << 1) | 1, mp->cw_max);
net/mac80211/rc80211_minstrel_ht.c
1410
cw = min((cw << 1) | 1, mp->cw_max);
net/mac80211/rc80211_minstrel_ht.c
1428
cw = min((cw << 1) | 1, mp->cw_max);
net/mac80211/rc80211_minstrel_ht.c
1434
if (tx_time_rtscts < mp->segment_size)
net/mac80211/rc80211_minstrel_ht.c
1436
} while ((tx_time < mp->segment_size) &&
net/mac80211/rc80211_minstrel_ht.c
1437
(++mrs->retry_count < mp->max_retry));
net/mac80211/rc80211_minstrel_ht.c
1442
minstrel_ht_set_rate(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
1453
minstrel_calc_retransmit(mp, mi, index);
net/mac80211/rc80211_minstrel_ht.c
1467
idx = mp->cck_rates[index % ARRAY_SIZE(mp->cck_rates)];
net/mac80211/rc80211_minstrel_ht.c
1469
idx = mp->ofdm_rates[mi->band][index %
net/mac80211/rc80211_minstrel_ht.c
1470
ARRAY_SIZE(mp->ofdm_rates[0])];
net/mac80211/rc80211_minstrel_ht.c
1550
minstrel_ht_update_rates(struct minstrel_priv *mp, struct minstrel_ht_sta *mi)
net/mac80211/rc80211_minstrel_ht.c
1554
int max_rates = min_t(int, mp->hw->max_rates, IEEE80211_TX_RATE_TABLE_SIZE);
net/mac80211/rc80211_minstrel_ht.c
1561
minstrel_ht_set_rate(mp, mi, rates, i++, mi->max_tp_rate[0]);
net/mac80211/rc80211_minstrel_ht.c
1565
minstrel_ht_set_rate(mp, mi, rates, i, mi->max_tp_rate[i]);
net/mac80211/rc80211_minstrel_ht.c
1568
minstrel_ht_set_rate(mp, mi, rates, i++, mi->max_prob_rate);
net/mac80211/rc80211_minstrel_ht.c
1575
rate_control_set_rates(mp->hw, mi->sta, rates);
net/mac80211/rc80211_minstrel_ht.c
1579
minstrel_ht_get_sample_rate(struct minstrel_priv *mp, struct minstrel_ht_sta *mi)
net/mac80211/rc80211_minstrel_ht.c
1583
if (mp->hw->max_rates > 1) {
net/mac80211/rc80211_minstrel_ht.c
1602
struct minstrel_priv *mp = priv;
net/mac80211/rc80211_minstrel_ht.c
1608
if (mp->fixed_rate_idx != -1)
net/mac80211/rc80211_minstrel_ht.c
1613
if (mp->hw->max_rates == 1 &&
net/mac80211/rc80211_minstrel_ht.c
1621
sample_idx = minstrel_ht_get_sample_rate(mp, mi);
net/mac80211/rc80211_minstrel_ht.c
1636
int idx = sample_idx % ARRAY_SIZE(mp->cck_rates);
net/mac80211/rc80211_minstrel_ht.c
1637
rate->idx = mp->cck_rates[idx];
net/mac80211/rc80211_minstrel_ht.c
1639
int idx = sample_idx % ARRAY_SIZE(mp->ofdm_rates[0]);
net/mac80211/rc80211_minstrel_ht.c
1640
rate->idx = mp->ofdm_rates[mi->band][idx];
net/mac80211/rc80211_minstrel_ht.c
1652
minstrel_ht_update_cck(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
1662
!ieee80211_hw_check(mp->hw, SUPPORTS_HT_CCK_RATES))
net/mac80211/rc80211_minstrel_ht.c
1666
if (mp->cck_rates[i] == 0xff ||
net/mac80211/rc80211_minstrel_ht.c
1667
!rate_supported(sta, sband->band, mp->cck_rates[i]))
net/mac80211/rc80211_minstrel_ht.c
1677
minstrel_ht_update_ofdm(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
1687
rates = mp->ofdm_rates[sband->band];
net/mac80211/rc80211_minstrel_ht.c
1688
for (i = 0; i < ARRAY_SIZE(mp->ofdm_rates[0]); i++) {
net/mac80211/rc80211_minstrel_ht.c
1702
struct minstrel_priv *mp = priv;
net/mac80211/rc80211_minstrel_ht.c
1824
minstrel_ht_update_cck(mp, mi, sband, sta);
net/mac80211/rc80211_minstrel_ht.c
1825
minstrel_ht_update_ofdm(mp, mi, sband, sta);
net/mac80211/rc80211_minstrel_ht.c
1828
minstrel_ht_update_stats(mp, mi);
net/mac80211/rc80211_minstrel_ht.c
1829
minstrel_ht_update_rates(mp, mi);
net/mac80211/rc80211_minstrel_ht.c
1854
struct minstrel_priv *mp = priv;
net/mac80211/rc80211_minstrel_ht.c
1855
struct ieee80211_hw *hw = mp->hw;
net/mac80211/rc80211_minstrel_ht.c
1894
minstrel_ht_init_cck_rates(struct minstrel_priv *mp)
net/mac80211/rc80211_minstrel_ht.c
1899
memset(mp->cck_rates, 0xff, sizeof(mp->cck_rates));
net/mac80211/rc80211_minstrel_ht.c
1900
sband = mp->hw->wiphy->bands[NL80211_BAND_2GHZ];
net/mac80211/rc80211_minstrel_ht.c
1904
BUILD_BUG_ON(ARRAY_SIZE(mp->cck_rates) != ARRAY_SIZE(bitrates));
net/mac80211/rc80211_minstrel_ht.c
1905
minstrel_ht_fill_rate_array(mp->cck_rates, sband,
net/mac80211/rc80211_minstrel_ht.c
1911
minstrel_ht_init_ofdm_rates(struct minstrel_priv *mp, enum nl80211_band band)
net/mac80211/rc80211_minstrel_ht.c
1916
memset(mp->ofdm_rates[band], 0xff, sizeof(mp->ofdm_rates[band]));
net/mac80211/rc80211_minstrel_ht.c
1917
sband = mp->hw->wiphy->bands[band];
net/mac80211/rc80211_minstrel_ht.c
1921
BUILD_BUG_ON(ARRAY_SIZE(mp->ofdm_rates[band]) != ARRAY_SIZE(bitrates));
net/mac80211/rc80211_minstrel_ht.c
1922
minstrel_ht_fill_rate_array(mp->ofdm_rates[band], sband,
net/mac80211/rc80211_minstrel_ht.c
1930
struct minstrel_priv *mp;
net/mac80211/rc80211_minstrel_ht.c
1933
mp = kzalloc_obj(struct minstrel_priv, GFP_ATOMIC);
net/mac80211/rc80211_minstrel_ht.c
1934
if (!mp)
net/mac80211/rc80211_minstrel_ht.c
1940
mp->cw_min = 15;
net/mac80211/rc80211_minstrel_ht.c
1941
mp->cw_max = 1023;
net/mac80211/rc80211_minstrel_ht.c
1944
mp->segment_size = 6000;
net/mac80211/rc80211_minstrel_ht.c
1947
mp->max_retry = hw->max_rate_tries;
net/mac80211/rc80211_minstrel_ht.c
1950
mp->max_retry = 7;
net/mac80211/rc80211_minstrel_ht.c
1952
mp->hw = hw;
net/mac80211/rc80211_minstrel_ht.c
1953
mp->update_interval = HZ / 20;
net/mac80211/rc80211_minstrel_ht.c
1955
minstrel_ht_init_cck_rates(mp);
net/mac80211/rc80211_minstrel_ht.c
1956
for (i = 0; i < ARRAY_SIZE(mp->hw->wiphy->bands); i++)
net/mac80211/rc80211_minstrel_ht.c
1957
minstrel_ht_init_ofdm_rates(mp, i);
net/mac80211/rc80211_minstrel_ht.c
1959
return mp;
net/mac80211/rc80211_minstrel_ht.c
1966
struct minstrel_priv *mp = priv;
net/mac80211/rc80211_minstrel_ht.c
1968
mp->fixed_rate_idx = (u32) -1;
net/mac80211/rc80211_minstrel_ht.c
1970
&mp->fixed_rate_idx);
net/mac80211/rc80211_minstrel_ht.c
278
minstrel_ht_update_rates(struct minstrel_priv *mp, struct minstrel_ht_sta *mi);
net/mac80211/rc80211_minstrel_ht.c
370
minstrel_ht_get_stats(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
388
for (idx = 0; idx < ARRAY_SIZE(mp->cck_rates); idx++) {
net/mac80211/rc80211_minstrel_ht.c
392
if (rate->idx != mp->cck_rates[idx])
net/mac80211/rc80211_minstrel_ht.c
403
for (idx = 0; idx < ARRAY_SIZE(mp->ofdm_rates[0]); idx++)
net/mac80211/rc80211_minstrel_ht.c
404
if (rate->idx == mp->ofdm_rates[mi->band][idx])
net/mac80211/rc80211_minstrel_ht.c
416
minstrel_ht_ri_get_stats(struct minstrel_priv *mp, struct minstrel_ht_sta *mi,
net/mac80211/rc80211_minstrel_ht.c
435
for (idx = 0; idx < ARRAY_SIZE(mp->cck_rates); idx++) {
net/mac80211/rc80211_minstrel_ht.c
436
if (rate->legacy != minstrel_cck_bitrates[ mp->cck_rates[idx] ])
net/mac80211/rc80211_minstrel_ht.c
447
for (idx = 0; idx < ARRAY_SIZE(mp->ofdm_rates[0]); idx++)
net/mac80211/rc80211_minstrel_ht.c
448
if (rate->legacy == minstrel_ofdm_bitrates[ mp->ofdm_rates[mi->band][idx] ])
net/mac80211/rc80211_minstrel_ht.c
766
minstrel_ht_calc_rate_stats(struct minstrel_priv *mp,
net/mpls/af_mpls.c
2091
struct nlattr *mp;
net/mpls/af_mpls.c
2095
mp = nla_nest_start_noflag(skb, RTA_MULTIPATH);
net/mpls/af_mpls.c
2096
if (!mp)
net/mpls/af_mpls.c
2137
nla_nest_end(skb, mp);
net/xdp/xskmap.c
123
const int ret = BPF_REG_0, mp = BPF_REG_1, index = BPF_REG_2;
net/xdp/xskmap.c
129
*insn++ = BPF_ALU64_IMM(BPF_ADD, mp, offsetof(struct xsk_map, xsk_map));
net/xdp/xskmap.c
130
*insn++ = BPF_ALU64_REG(BPF_ADD, ret, mp);
net/xfrm/xfrm_policy.c
4571
struct xfrm_migrate *mp;
net/xfrm/xfrm_policy.c
4583
for (j = 0, mp = m; j < num_migrate; j++, mp++) {
net/xfrm/xfrm_policy.c
4584
if (!migrate_tmpl_match(mp, &pol->xfrm_vec[i]))
net/xfrm/xfrm_policy.c
4592
memcpy(&pol->xfrm_vec[i].id.daddr, &mp->new_daddr,
net/xfrm/xfrm_policy.c
4594
memcpy(&pol->xfrm_vec[i].saddr, &mp->new_saddr,
net/xfrm/xfrm_policy.c
4596
pol->xfrm_vec[i].encap_family = mp->new_family;
net/xfrm/xfrm_policy.c
4657
struct xfrm_migrate *mp;
net/xfrm/xfrm_policy.c
4679
for (i = 0, mp = m; i < num_migrate; i++, mp++) {
net/xfrm/xfrm_policy.c
4680
if ((x = xfrm_migrate_state_find(mp, net, if_id))) {
net/xfrm/xfrm_policy.c
4683
xc = xfrm_state_migrate(x, mp, encap, net, xuo, extack);
net/xfrm/xfrm_user.c
3230
const struct xfrm_migrate *mp;
net/xfrm/xfrm_user.c
3258
for (i = 0, mp = m ; i < num_migrate; i++, mp++) {
net/xfrm/xfrm_user.c
3259
err = copy_to_user_migrate(mp, skb);
scripts/dtc/data.c
132
struct marker **mp = &d.markers;
scripts/dtc/data.c
135
while (*mp)
scripts/dtc/data.c
136
mp = &((*mp)->next);
scripts/dtc/data.c
137
*mp = m;
security/security.c
284
static int lsm_msg_msg_alloc(struct msg_msg *mp)
security/security.c
286
return lsm_blob_alloc(&mp->security, blob_sizes.lbs_msg_msg,
tools/lib/perf/evlist.c
524
perf_evlist__mmap_cb_mmap(struct perf_mmap *map, struct perf_mmap_param *mp,
tools/lib/perf/evlist.c
527
return perf_mmap__mmap(map, mp, output, cpu);
tools/lib/perf/evlist.c
541
int idx, struct perf_mmap_param *mp, int cpu_idx,
tools/lib/perf/evlist.c
566
mp->prot = PROT_READ;
tools/lib/perf/evlist.c
569
mp->prot = PROT_READ | PROT_WRITE;
tools/lib/perf/evlist.c
594
ops->idx(evlist, evsel, mp, idx);
tools/lib/perf/evlist.c
598
if (ops->mmap(map, mp, *output, evlist_cpu) < 0)
tools/lib/perf/evlist.c
635
struct perf_mmap_param *mp)
tools/lib/perf/evlist.c
650
if (mmap_per_evsel(evlist, ops, idx, mp, 0, thread, &output,
tools/lib/perf/evlist.c
660
if (mmap_per_evsel(evlist, ops, idx, mp, cpu, 0, &output,
tools/lib/perf/evlist.c
677
struct perf_mmap_param *mp)
tools/lib/perf/evlist.c
691
if (mmap_per_evsel(evlist, ops, cpu, mp, cpu,
tools/lib/perf/evlist.c
725
struct perf_mmap_param *mp)
tools/lib/perf/evlist.c
733
mp->mask = evlist->mmap_len - page_size - 1;
tools/lib/perf/evlist.c
748
return mmap_per_thread(evlist, ops, mp);
tools/lib/perf/evlist.c
750
return mmap_per_cpu(evlist, ops, mp);
tools/lib/perf/evlist.c
755
struct perf_mmap_param mp;
tools/lib/perf/evlist.c
763
return perf_evlist__mmap_ops(evlist, &ops, &mp);
tools/lib/perf/evsel.c
261
struct perf_mmap_param mp = {
tools/lib/perf/evsel.c
284
ret = perf_mmap__mmap(map, &mp, *fd, cpu);
tools/lib/perf/include/internal/evlist.h
60
struct perf_mmap_param *mp);
tools/lib/perf/include/internal/mmap.h
50
int perf_mmap__mmap(struct perf_mmap *map, struct perf_mmap_param *mp,
tools/lib/perf/mmap.c
36
int perf_mmap__mmap(struct perf_mmap *map, struct perf_mmap_param *mp,
tools/lib/perf/mmap.c
40
map->mask = mp->mask;
tools/lib/perf/mmap.c
41
map->base = mmap(NULL, perf_mmap__mmap_len(map), mp->prot,
tools/perf/util/auxtrace.c
133
struct auxtrace_mmap_params *mp,
tools/perf/util/auxtrace.c
141
mm->mask = mp->mask;
tools/perf/util/auxtrace.c
142
mm->len = mp->len;
tools/perf/util/auxtrace.c
144
mm->idx = mp->idx;
tools/perf/util/auxtrace.c
145
mm->tid = mp->tid;
tools/perf/util/auxtrace.c
146
mm->cpu = mp->cpu.cpu;
tools/perf/util/auxtrace.c
148
if (!mp->len || !mp->mmap_needed) {
tools/perf/util/auxtrace.c
153
pc->aux_offset = mp->offset;
tools/perf/util/auxtrace.c
154
pc->aux_size = mp->len;
tools/perf/util/auxtrace.c
156
mm->base = mmap(NULL, mp->len, mp->prot, MAP_SHARED, fd, mp->offset);
tools/perf/util/auxtrace.c
174
void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
tools/perf/util/auxtrace.c
180
mp->offset = auxtrace_offset;
tools/perf/util/auxtrace.c
181
mp->len = auxtrace_pages * (size_t)page_size;
tools/perf/util/auxtrace.c
182
mp->mask = is_power_of_2(mp->len) ? mp->len - 1 : 0;
tools/perf/util/auxtrace.c
183
mp->prot = PROT_READ | (auxtrace_overwrite ? 0 : PROT_WRITE);
tools/perf/util/auxtrace.c
184
pr_debug2("AUX area mmap length %zu\n", mp->len);
tools/perf/util/auxtrace.c
186
mp->len = 0;
tools/perf/util/auxtrace.c
190
void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
tools/perf/util/auxtrace.c
196
mp->mmap_needed = evsel->needs_auxtrace_mmap;
tools/perf/util/auxtrace.c
198
if (!mp->mmap_needed)
tools/perf/util/auxtrace.c
201
mp->idx = idx;
tools/perf/util/auxtrace.c
204
mp->cpu = perf_cpu_map__cpu(evlist->core.all_cpus, idx);
tools/perf/util/auxtrace.c
205
mp->tid = perf_thread_map__pid(evlist->core.threads, 0);
tools/perf/util/auxtrace.c
207
mp->cpu.cpu = -1;
tools/perf/util/auxtrace.c
208
mp->tid = perf_thread_map__pid(evlist->core.threads, idx);
tools/perf/util/auxtrace.h
493
struct auxtrace_mmap_params *mp,
tools/perf/util/auxtrace.h
496
void auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp,
tools/perf/util/auxtrace.h
500
void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp,
tools/perf/util/evlist.c
1021
struct mmap_params mp = {
tools/perf/util/evlist.c
1036
auxtrace_mmap_params__init(&mp.auxtrace_mp, evlist->core.mmap_len,
tools/perf/util/evlist.c
1039
return perf_evlist__mmap_ops(&evlist->core, &ops, &mp.core);
tools/perf/util/evlist.c
849
struct mmap_params *mp = container_of(_mp, struct mmap_params, core);
tools/perf/util/evlist.c
852
auxtrace_mmap_params__set_idx(&mp->auxtrace_mp, evlist, evsel, idx);
tools/perf/util/evlist.c
885
struct mmap_params *mp = container_of(_mp, struct mmap_params, core);
tools/perf/util/evlist.c
887
return mmap__mmap(map, mp, output, cpu);
tools/perf/util/mmap.c
148
static int perf_mmap__aio_mmap(struct mmap *map, struct mmap_params *mp)
tools/perf/util/mmap.c
152
map->aio.nr_cblocks = mp->nr_cblocks;
tools/perf/util/mmap.c
176
ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity);
tools/perf/util/mmap.c
220
struct mmap_params *mp __maybe_unused)
tools/perf/util/mmap.c
262
static int perf_mmap__setup_affinity_mask(struct mmap *map, struct mmap_params *mp)
tools/perf/util/mmap.c
269
if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1)
tools/perf/util/mmap.c
271
else if (mp->affinity == PERF_AFFINITY_CPU)
tools/perf/util/mmap.c
277
int mmap__mmap(struct mmap *map, struct mmap_params *mp, int fd, struct perf_cpu cpu)
tools/perf/util/mmap.c
279
if (perf_mmap__mmap(&map->core, &mp->core, fd, cpu)) {
tools/perf/util/mmap.c
285
if (mp->affinity != PERF_AFFINITY_SYS &&
tools/perf/util/mmap.c
286
perf_mmap__setup_affinity_mask(map, mp)) {
tools/perf/util/mmap.c
295
map->core.flush = mp->flush;
tools/perf/util/mmap.c
297
if (zstd_init(&map->zstd_data, mp->comp_level)) {
tools/perf/util/mmap.c
302
if (mp->comp_level && !perf_mmap__aio_enabled(map)) {
tools/perf/util/mmap.c
314
&mp->auxtrace_mp, map->core.base, fd))
tools/perf/util/mmap.c
317
return perf_mmap__aio_mmap(map, mp);
tools/perf/util/mmap.c
46
struct auxtrace_mmap_params *mp __maybe_unused,
tools/perf/util/mmap.c
57
void __weak auxtrace_mmap_params__init(struct auxtrace_mmap_params *mp __maybe_unused,
tools/perf/util/mmap.c
64
void __weak auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp __maybe_unused,
tools/perf/util/mmap.h
52
int mmap__mmap(struct mmap *map, struct mmap_params *mp, int fd, struct perf_cpu cpu);
tools/perf/util/sort.c
1476
unsigned char mp, p;
tools/perf/util/sort.c
1481
mp = left->branch_info->flags.mispred != right->branch_info->flags.mispred;
tools/perf/util/sort.c
1483
return mp || p;
tools/power/x86/turbostat/turbostat.c
10563
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
10565
for (mp = head; mp; mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
10567
fprintf(stderr, "%s: %s %s\n", __func__, name, mp->name);
tools/power/x86/turbostat/turbostat.c
10568
if (!strcmp(name, mp->name))
tools/power/x86/turbostat/turbostat.c
10569
return mp;
tools/power/x86/turbostat/turbostat.c
2868
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
2933
for (mp = sys.tp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
2934
outp += print_name(mp->width, &printed, delim, mp->name, mp->type, mp->format);
tools/power/x86/turbostat/turbostat.c
2981
for (mp = sys.cp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
2982
outp += print_name(mp->width, &printed, delim, mp->name, mp->type, mp->format);
tools/power/x86/turbostat/turbostat.c
3084
for (mp = sys.pp; mp; mp = mp->next)
tools/power/x86/turbostat/turbostat.c
3085
outp += print_name(mp->width, &printed, delim, mp->name, mp->type, mp->format);
tools/power/x86/turbostat/turbostat.c
3144
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
3174
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3175
outp += sprintf(outp, "tADDED [%d] %8s msr0x%x: %08llX %s\n", i, mp->name, mp->msr_num, t->counter[i], mp->sp->path);
tools/power/x86/turbostat/turbostat.c
3193
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3194
outp += sprintf(outp, "cADDED [%d] %8s msr0x%x: %08llX %s\n", i, mp->name, mp->msr_num, c->counter[i], mp->sp->path);
tools/power/x86/turbostat/turbostat.c
3226
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3227
outp += sprintf(outp, "pADDED [%d] %8s msr0x%x: %08llX %s\n", i, mp->name, mp->msr_num, p->counter[i], mp->sp->path);
tools/power/x86/turbostat/turbostat.c
3305
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
3455
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3456
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
3457
outp += print_hex_value(mp->width, &printed, delim, t->counter[i]);
tools/power/x86/turbostat/turbostat.c
3458
else if (mp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3459
outp += print_decimal_value(mp->width, &printed, delim, t->counter[i]);
tools/power/x86/turbostat/turbostat.c
3460
else if (mp->format == FORMAT_PERCENT) {
tools/power/x86/turbostat/turbostat.c
3461
if (mp->type == COUNTER_USEC)
tools/power/x86/turbostat/turbostat.c
3472
else if (pp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3529
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3530
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
3531
outp += print_hex_value(mp->width, &printed, delim, c->counter[i]);
tools/power/x86/turbostat/turbostat.c
3532
else if (mp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3533
outp += print_decimal_value(mp->width, &printed, delim, c->counter[i]);
tools/power/x86/turbostat/turbostat.c
3534
else if (mp->format == FORMAT_PERCENT)
tools/power/x86/turbostat/turbostat.c
3542
else if (pp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3681
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3682
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
3683
outp += print_hex_value(mp->width, &printed, delim, p->counter[i]);
tools/power/x86/turbostat/turbostat.c
3684
else if (mp->type == COUNTER_K2M)
tools/power/x86/turbostat/turbostat.c
3686
else if (mp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3687
outp += print_decimal_value(mp->width, &printed, delim, p->counter[i]);
tools/power/x86/turbostat/turbostat.c
3688
else if (mp->format == FORMAT_PERCENT)
tools/power/x86/turbostat/turbostat.c
3698
else if (pp->format == FORMAT_DELTA || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3781
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
3835
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3836
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
3838
else if (mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3866
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
3879
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
3880
if (mp->format == FORMAT_RAW || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
3915
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
4003
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4004
if (mp->format == FORMAT_RAW || mp->format == FORMAT_AVERAGE)
tools/power/x86/turbostat/turbostat.c
4060
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
4128
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4131
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4134
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next)
tools/power/x86/turbostat/turbostat.c
4163
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
4197
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4198
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
4227
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4228
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
4289
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4290
if ((mp->format == FORMAT_RAW) && (topo.num_packages == 0))
tools/power/x86/turbostat/turbostat.c
4317
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
4366
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4367
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
4369
if (mp->type == COUNTER_ITEMS) {
tools/power/x86/turbostat/turbostat.c
4376
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4377
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
4379
if (mp->type == COUNTER_ITEMS) {
tools/power/x86/turbostat/turbostat.c
4385
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
4386
if (mp->format == FORMAT_RAW)
tools/power/x86/turbostat/turbostat.c
4388
if (mp->type == COUNTER_ITEMS) {
tools/power/x86/turbostat/turbostat.c
4478
int get_mp(int cpu, struct msr_counter *mp, unsigned long long *counterp, char *counter_path)
tools/power/x86/turbostat/turbostat.c
4480
if (mp->msr_num != 0) {
tools/power/x86/turbostat/turbostat.c
4482
if (get_msr(cpu, mp->msr_num, counterp))
tools/power/x86/turbostat/turbostat.c
4487
if (mp->flags & SYSFS_PERCPU) {
tools/power/x86/turbostat/turbostat.c
4488
sprintf(path, "/sys/devices/system/cpu/cpu%d/%s", cpu, mp->sp->path);
tools/power/x86/turbostat/turbostat.c
5179
struct msr_counter *mp;
tools/power/x86/turbostat/turbostat.c
5215
for (i = 0, mp = sys.tp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5216
if (get_mp(cpu, mp, &t->counter[i], mp->sp->path))
tools/power/x86/turbostat/turbostat.c
5259
for (i = 0, mp = sys.cp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5260
if (get_mp(cpu, mp, &c->counter[i], mp->sp->path))
tools/power/x86/turbostat/turbostat.c
5329
for (i = 0, mp = sys.pp; mp; i++, mp = mp->next) {
tools/power/x86/turbostat/turbostat.c
5332
if (mp->msr_num == 0) {
tools/power/x86/turbostat/turbostat.c
5333
path = find_sysfs_path_by_id(mp->sp, cpus[t->cpu_id].package_id);
tools/power/x86/turbostat/turbostat.c
5339
if (get_mp(cpu, mp, &p->counter[i], path))
tools/tracing/rtla/src/utils.c
701
static const int find_mount(const char *fs, char *mp, int sizeof_mp)
tools/tracing/rtla/src/utils.c
723
memset(mp, 0, sizeof_mp);
tools/tracing/rtla/src/utils.c
724
strncpy(mp, mount_point, sizeof_mp - 1);
tools/tracing/rtla/src/utils.c
726
debug_msg("Fs %s found at %s\n", fs, mp);