arch/arm/include/asm/uaccess-asm.h
51
.if \isb
arch/arm/include/asm/uaccess-asm.h
63
.if \isb
arch/arm/include/asm/uaccess-asm.h
80
.if \isb
arch/arm/include/asm/uaccess-asm.h
94
.if \isb
arch/arm64/include/asm/assembler.h
104
.macro sb
arch/arm64/include/asm/assembler.h
842
sb
arch/arm64/include/asm/insn.h
434
__AARCH64_INSN_FUNCS(sb, 0xFFFFFFFF, 0xD50330FF)
arch/mips/boot/compressed/calc_vmlinuz_load_addr.c
17
struct stat sb;
arch/mips/boot/compressed/calc_vmlinuz_load_addr.c
25
if (stat(argv[1], &sb) == -1) {
arch/mips/boot/compressed/calc_vmlinuz_load_addr.c
41
vmlinux_size = (uint64_t)sb.st_size;
arch/mips/include/asm/asm-eva.h
118
#define kernel_sb(reg, addr) sb reg, addr
arch/mips/include/asm/sibyte/board.h
28
sb t1, 0x18(t0); \
arch/mips/include/asm/sibyte/board.h
30
sb t1, 0x10(t0); \
arch/mips/include/asm/sibyte/board.h
32
sb t1, 0x08(t0); \
arch/mips/include/asm/sibyte/board.h
34
sb t1, 0x00(t0)
arch/mips/net/bpf_jit_comp32.c
695
emit(ctx, sb, lo(src), off, dst);
arch/mips/net/bpf_jit_comp64.c
363
emit(ctx, sb, src, off, dst);
arch/powerpc/kvm/book3s_xive.c
1084
struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive.c
1097
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1104
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1128
static void xive_lock_for_unmask(struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive.c
1135
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1138
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1143
struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive.c
1271
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1295
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1296
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1298
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1315
xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
1317
xive_lock_for_unmask(sb, state);
arch/powerpc/kvm/book3s_xive.c
1356
xive_finish_unmask(xive, sb, state, priority);
arch/powerpc/kvm/book3s_xive.c
1364
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1372
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1379
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1380
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1382
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1383
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1386
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1394
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1401
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1402
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1404
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1423
xive_lock_for_unmask(sb, state);
arch/powerpc/kvm/book3s_xive.c
1424
xive_finish_unmask(xive, sb, state, state->saved_priority);
arch/powerpc/kvm/book3s_xive.c
1425
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1433
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1440
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1441
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1443
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1450
state->saved_priority = xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
1451
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1458
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1462
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1463
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1465
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1554
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1569
sb = kvmppc_xive_find_source(xive, guest_irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1570
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1572
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1598
prio = xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
1642
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1652
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
1663
sb = kvmppc_xive_find_source(xive, guest_irq, &idx);
arch/powerpc/kvm/book3s_xive.c
1664
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1666
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
1673
prio = xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
1722
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1736
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive.c
1738
if (!sb)
arch/powerpc/kvm/book3s_xive.c
1741
struct kvmppc_xive_irq_state *state = &sb->irq_state[j];
arch/powerpc/kvm/book3s_xive.c
1751
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
1759
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
2037
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
2041
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
2042
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2045
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
2066
struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive.c
2069
struct kvmppc_xive_irq_state *state = &sb->irq_state[irq];
arch/powerpc/kvm/book3s_xive.c
2075
state->saved_scan_prio = xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
2082
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
2086
struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive.c
2089
struct kvmppc_xive_irq_state *state = &sb->irq_state[irq];
arch/powerpc/kvm/book3s_xive.c
2099
xive_lock_for_unmask(sb, state);
arch/powerpc/kvm/book3s_xive.c
2103
xive_finish_unmask(xive, sb, state, state->saved_scan_prio);
arch/powerpc/kvm/book3s_xive.c
2106
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
2133
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive.c
2134
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2137
xive_pre_save_mask_irq(xive, sb, j);
arch/powerpc/kvm/book3s_xive.c
2153
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive.c
2154
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2157
xive_pre_save_unmask_irq(xive, sb, j);
arch/powerpc/kvm/book3s_xive.c
2167
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive.c
2168
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2171
sb->irq_state[j].in_queue = false;
arch/powerpc/kvm/book3s_xive.c
2183
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
2189
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
2190
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2193
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
2267
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
2279
sb = kzalloc_obj(*sb);
arch/powerpc/kvm/book3s_xive.c
2280
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2283
sb->id = bid;
arch/powerpc/kvm/book3s_xive.c
2286
sb->irq_state[i].number = (bid << KVMPPC_XICS_ICS_SHIFT) | i;
arch/powerpc/kvm/book3s_xive.c
2287
sb->irq_state[i].eisn = 0;
arch/powerpc/kvm/book3s_xive.c
2288
sb->irq_state[i].guest_priority = MASKED;
arch/powerpc/kvm/book3s_xive.c
2289
sb->irq_state[i].saved_priority = MASKED;
arch/powerpc/kvm/book3s_xive.c
2290
sb->irq_state[i].act_priority = MASKED;
arch/powerpc/kvm/book3s_xive.c
2293
xive->src_blocks[bid] = sb;
arch/powerpc/kvm/book3s_xive.c
2326
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
2341
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
2342
if (!sb) {
arch/powerpc/kvm/book3s_xive.c
2344
sb = kvmppc_xive_create_src_block(xive, irq);
arch/powerpc/kvm/book3s_xive.c
2345
if (!sb) {
arch/powerpc/kvm/book3s_xive.c
2350
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
2386
xive_lock_and_mask(xive, sb, state);
arch/powerpc/kvm/book3s_xive.c
2404
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
2475
xive_finish_unmask(xive, sb, state, guest_prio);
arch/powerpc/kvm/book3s_xive.c
2491
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
2498
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive.c
2499
if (!sb)
arch/powerpc/kvm/book3s_xive.c
2503
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive.c
2614
void kvmppc_xive_free_sources(struct kvmppc_xive_src_block *sb)
arch/powerpc/kvm/book3s_xive.c
2619
struct kvmppc_xive_irq_state *state = &sb->irq_state[i];
arch/powerpc/kvm/book3s_xive.c
2843
struct kvmppc_xive_src_block *sb)
arch/powerpc/kvm/book3s_xive.c
2849
struct kvmppc_xive_irq_state *state = &sb->irq_state[i];
arch/powerpc/kvm/book3s_xive.c
2938
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive.c
2940
if (sb) {
arch/powerpc/kvm/book3s_xive.c
2941
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
2942
kvmppc_xive_debug_show_sources(m, sb);
arch/powerpc/kvm/book3s_xive.c
2943
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
399
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
423
sb = kvmppc_xive_find_source(xive, irq, &src);
arch/powerpc/kvm/book3s_xive.c
424
if (!sb)
arch/powerpc/kvm/book3s_xive.c
426
state = &sb->irq_state[src];
arch/powerpc/kvm/book3s_xive.c
518
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive.c
550
sb = kvmppc_xive_find_source(xive, irq, &src);
arch/powerpc/kvm/book3s_xive.c
551
if (!sb) {
arch/powerpc/kvm/book3s_xive.c
558
state = &sb->irq_state[src];
arch/powerpc/kvm/book3s_xive.c
573
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
575
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.c
583
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive.h
294
struct kvmppc_xive_src_block *sb);
arch/powerpc/kvm/book3s_xive.h
297
void kvmppc_xive_free_sources(struct kvmppc_xive_src_block *sb);
arch/powerpc/kvm/book3s_xive_native.c
1245
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive_native.c
1247
if (sb) {
arch/powerpc/kvm/book3s_xive_native.c
1248
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
1249
kvmppc_xive_debug_show_sources(m, sb);
arch/powerpc/kvm/book3s_xive_native.c
1250
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
233
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive_native.c
249
sb = kvmppc_xive_find_source(xive, irq, &src);
arch/powerpc/kvm/book3s_xive_native.c
250
if (!sb) {
arch/powerpc/kvm/book3s_xive_native.c
255
state = &sb->irq_state[src];
arch/powerpc/kvm/book3s_xive_native.c
265
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
272
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
342
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive_native.c
354
sb = kvmppc_xive_find_source(xive, irq, &idx);
arch/powerpc/kvm/book3s_xive_native.c
355
if (!sb) {
arch/powerpc/kvm/book3s_xive_native.c
357
sb = kvmppc_xive_create_src_block(xive, irq);
arch/powerpc/kvm/book3s_xive_native.c
358
if (!sb) {
arch/powerpc/kvm/book3s_xive_native.c
363
state = &sb->irq_state[idx];
arch/powerpc/kvm/book3s_xive_native.c
370
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
411
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
417
struct kvmppc_xive_src_block *sb,
arch/powerpc/kvm/book3s_xive_native.c
426
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
459
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
466
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive_native.c
476
sb = kvmppc_xive_find_source(xive, irq, &src);
arch/powerpc/kvm/book3s_xive_native.c
477
if (!sb)
arch/powerpc/kvm/book3s_xive_native.c
480
state = &sb->irq_state[src];
arch/powerpc/kvm/book3s_xive_native.c
505
return kvmppc_xive_native_update_source_config(xive, sb, state, server,
arch/powerpc/kvm/book3s_xive_native.c
512
struct kvmppc_xive_src_block *sb;
arch/powerpc/kvm/book3s_xive_native.c
521
sb = kvmppc_xive_find_source(xive, irq, &src);
arch/powerpc/kvm/book3s_xive_native.c
522
if (!sb)
arch/powerpc/kvm/book3s_xive_native.c
525
state = &sb->irq_state[src];
arch/powerpc/kvm/book3s_xive_native.c
529
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
537
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
779
static void kvmppc_xive_reset_sources(struct kvmppc_xive_src_block *sb)
arch/powerpc/kvm/book3s_xive_native.c
784
struct kvmppc_xive_irq_state *state = &sb->irq_state[i];
arch/powerpc/kvm/book3s_xive_native.c
842
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive_native.c
844
if (sb) {
arch/powerpc/kvm/book3s_xive_native.c
845
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
846
kvmppc_xive_reset_sources(sb);
arch/powerpc/kvm/book3s_xive_native.c
847
arch_spin_unlock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
856
static void kvmppc_xive_native_sync_sources(struct kvmppc_xive_src_block *sb)
arch/powerpc/kvm/book3s_xive_native.c
861
struct kvmppc_xive_irq_state *state = &sb->irq_state[j];
arch/powerpc/kvm/book3s_xive_native.c
924
struct kvmppc_xive_src_block *sb = xive->src_blocks[i];
arch/powerpc/kvm/book3s_xive_native.c
926
if (sb) {
arch/powerpc/kvm/book3s_xive_native.c
927
arch_spin_lock(&sb->lock);
arch/powerpc/kvm/book3s_xive_native.c
928
kvmppc_xive_native_sync_sources(sb);
arch/powerpc/kvm/book3s_xive_native.c
929
arch_spin_unlock(&sb->lock);
arch/powerpc/platforms/cell/spufs/inode.c
110
spufs_new_file(struct super_block *sb, struct dentry *dentry,
arch/powerpc/platforms/cell/spufs/inode.c
121
inode = spufs_new_inode(sb, S_IFREG | mode);
arch/powerpc/platforms/cell/spufs/inode.c
42
static struct spufs_sb_info *spufs_get_sb_info(struct super_block *sb)
arch/powerpc/platforms/cell/spufs/inode.c
44
return sb->s_fs_info;
arch/powerpc/platforms/cell/spufs/inode.c
48
spufs_alloc_inode(struct super_block *sb)
arch/powerpc/platforms/cell/spufs/inode.c
665
static int spufs_create_root(struct super_block *sb, struct fs_context *fc)
arch/powerpc/platforms/cell/spufs/inode.c
673
inode = spufs_new_inode(sb, S_IFDIR | ctx->mode);
arch/powerpc/platforms/cell/spufs/inode.c
684
sb->s_root = d_make_root(inode);
arch/powerpc/platforms/cell/spufs/inode.c
685
if (!sb->s_root)
arch/powerpc/platforms/cell/spufs/inode.c
698
static int spufs_fill_super(struct super_block *sb, struct fs_context *fc)
arch/powerpc/platforms/cell/spufs/inode.c
700
sb->s_maxbytes = MAX_LFS_FILESIZE;
arch/powerpc/platforms/cell/spufs/inode.c
701
sb->s_blocksize = PAGE_SIZE;
arch/powerpc/platforms/cell/spufs/inode.c
702
sb->s_blocksize_bits = PAGE_SHIFT;
arch/powerpc/platforms/cell/spufs/inode.c
703
sb->s_magic = SPUFS_MAGIC;
arch/powerpc/platforms/cell/spufs/inode.c
704
sb->s_op = &spufs_ops;
arch/powerpc/platforms/cell/spufs/inode.c
706
return spufs_create_root(sb, fc);
arch/powerpc/platforms/cell/spufs/inode.c
77
spufs_new_inode(struct super_block *sb, umode_t mode)
arch/powerpc/platforms/cell/spufs/inode.c
81
inode = new_inode(sb);
arch/s390/hypfs/inode.c
151
struct super_block *sb = file_inode(iocb->ki_filp)->i_sb;
arch/s390/hypfs/inode.c
152
struct hypfs_sb_info *fs_info = sb->s_fs_info;
arch/s390/hypfs/inode.c
172
rc = hypfs_vm_create_files(sb->s_root);
arch/s390/hypfs/inode.c
174
rc = hypfs_diag_create_files(sb->s_root);
arch/s390/hypfs/inode.c
180
hypfs_update_update(sb);
arch/s390/hypfs/inode.c
240
static int hypfs_fill_super(struct super_block *sb, struct fs_context *fc)
arch/s390/hypfs/inode.c
242
struct hypfs_sb_info *sbi = sb->s_fs_info;
arch/s390/hypfs/inode.c
247
sb->s_blocksize = PAGE_SIZE;
arch/s390/hypfs/inode.c
248
sb->s_blocksize_bits = PAGE_SHIFT;
arch/s390/hypfs/inode.c
249
sb->s_magic = HYPFS_MAGIC;
arch/s390/hypfs/inode.c
250
sb->s_op = &hypfs_s_ops;
arch/s390/hypfs/inode.c
252
root_inode = hypfs_make_inode(sb, S_IFDIR | 0755);
arch/s390/hypfs/inode.c
257
sb->s_root = root_dentry = d_make_root(root_inode);
arch/s390/hypfs/inode.c
270
hypfs_update_update(sb);
arch/s390/hypfs/inode.c
308
static void hypfs_kill_super(struct super_block *sb)
arch/s390/hypfs/inode.c
310
struct hypfs_sb_info *sb_info = sb->s_fs_info;
arch/s390/hypfs/inode.c
313
kill_anon_super(sb);
arch/s390/hypfs/inode.c
50
static void hypfs_update_update(struct super_block *sb)
arch/s390/hypfs/inode.c
52
struct hypfs_sb_info *sb_info = sb->s_fs_info;
arch/s390/hypfs/inode.c
78
static struct inode *hypfs_make_inode(struct super_block *sb, umode_t mode)
arch/s390/hypfs/inode.c
80
struct inode *ret = new_inode(sb);
arch/s390/hypfs/inode.c
83
struct hypfs_sb_info *hypfs_info = sb->s_fs_info;
arch/s390/include/uapi/asm/kvm.h
312
__u64 sb; /* Guest addr of summary bit */
arch/s390/kvm/pci.c
601
if (args->u.reg_aen.sb != 0) {
arch/s390/kvm/pci.c
602
fib.fmt0.aisb = args->u.reg_aen.sb;
arch/x86/kernel/acpi/boot.c
885
struct acpi_table_boot *sb = (struct acpi_table_boot *)table;
arch/x86/kernel/acpi/boot.c
887
sbf_port = sb->cmos_index; /* Save CMOS port */
arch/x86/mm/pat/cpa-test.c
119
struct split_state sa, sb, sc;
arch/x86/mm/pat/cpa-test.c
215
failed += print_split(&sb);
block/bdev.c
219
static int sb_validate_large_blocksize(struct super_block *sb, int size)
block/bdev.c
223
if (!(sb->s_type->fs_flags & FS_LBS))
block/bdev.c
232
sb->s_type->name, size, PAGE_SIZE, err_str);
block/bdev.c
236
int sb_set_blocksize(struct super_block *sb, int size)
block/bdev.c
238
if (size > PAGE_SIZE && sb_validate_large_blocksize(sb, size))
block/bdev.c
240
if (set_blocksize(sb->s_bdev_file, size))
block/bdev.c
243
sb->s_blocksize = size;
block/bdev.c
244
sb->s_blocksize_bits = blksize_bits(size);
block/bdev.c
245
return sb->s_blocksize;
block/bdev.c
250
int __must_check sb_min_blocksize(struct super_block *sb, int size)
block/bdev.c
252
int minsize = bdev_logical_block_size(sb->s_bdev);
block/bdev.c
255
return sb_set_blocksize(sb, size);
block/bdev.c
378
static struct inode *bdev_alloc_inode(struct super_block *sb)
block/bdev.c
380
struct bdev_inode *ei = alloc_inode_sb(sb, bdev_cachep, GFP_KERNEL);
block/blk-mq-debugfs.c
432
sbitmap_bitmap_show(&hctx->tags->bitmap_tags.sb, m);
block/blk-mq-debugfs.c
464
sbitmap_bitmap_show(&hctx->sched_tags->bitmap_tags.sb, m);
block/blk-mq-tag.c
324
sbitmap_for_each_set(&bt->sb, bt_iter, &iter_data);
block/blk-mq-tag.c
391
sbitmap_for_each_set(&bt->sb, bt_tags_iter, &iter_data);
block/blk-mq.c
1799
static bool flush_busy_ctx(struct sbitmap *sb, unsigned int bitnr, void *data)
block/blk-mq.c
1808
sbitmap_clear_bit(sb, bitnr);
block/blk-mq.c
1832
static bool dispatch_rq_from_ctx(struct sbitmap *sb, unsigned int bitnr,
block/blk-mq.c
1845
sbitmap_clear_bit(sb, bitnr);
block/blk-mq.h
414
if (bt->sb.depth == 1)
block/blk-mq.h
434
depth = max((bt->sb.depth + users - 1) / users, 4U);
block/kyber-iosched.c
263
if (depth != kqd->domain_tokens[sched_domain].sb.depth) {
block/kyber-iosched.c
343
orig_depth = kqd->domain_tokens[sched_domain].sb.depth;
block/kyber-iosched.c
653
static bool flush_busy_kcq(struct sbitmap *sb, unsigned int bitnr, void *data)
block/kyber-iosched.c
661
sbitmap_clear_bit(sb, bitnr);
crypto/twofish_common.c
485
ctx->s[1][i] = mds[1][q0[(b) ^ sb] ^ sf]; \
crypto/twofish_common.c
493
ctx->s[1][i] = mds[1][q0[q1[(b) ^ sb] ^ sf] ^ sj]; \
crypto/twofish_common.c
501
ctx->s[1][i] = mds[1][q0[q1[q1[(a) ^ sb] ^ sf] ^ sj] ^ sn]; \
crypto/twofish_common.c
579
u8 sa = 0, sb = 0, sc = 0, sd = 0, se = 0, sf = 0, sg = 0, sh = 0;
crypto/twofish_common.c
593
CALC_S (sa, sb, sc, sd, 0, 0x00, 0x2D, 0x01, 0x2D); /* 01 A4 02 A4 */
crypto/twofish_common.c
594
CALC_S (sa, sb, sc, sd, 1, 0x2D, 0xA4, 0x44, 0x8A); /* A4 56 A1 55 */
crypto/twofish_common.c
595
CALC_S (sa, sb, sc, sd, 2, 0x8A, 0xD5, 0xBF, 0xD1); /* 55 82 FC 87 */
crypto/twofish_common.c
596
CALC_S (sa, sb, sc, sd, 3, 0xD1, 0x7F, 0x3D, 0x99); /* 87 F3 C1 5A */
crypto/twofish_common.c
597
CALC_S (sa, sb, sc, sd, 4, 0x99, 0x46, 0x66, 0x96); /* 5A 1E 47 58 */
crypto/twofish_common.c
598
CALC_S (sa, sb, sc, sd, 5, 0x96, 0x3C, 0x5B, 0xED); /* 58 C6 AE DB */
crypto/twofish_common.c
599
CALC_S (sa, sb, sc, sd, 6, 0xED, 0x37, 0x4F, 0xE0); /* DB 68 3D 9E */
crypto/twofish_common.c
600
CALC_S (sa, sb, sc, sd, 7, 0xE0, 0xD0, 0x8C, 0x17); /* 9E E5 19 03 */
drivers/accel/habanalabs/common/hldio.c
100
if (inode->i_blocks << sb->s_blocksize_bits < i_size_read(inode)) {
drivers/accel/habanalabs/common/hldio.c
71
struct super_block *sb;
drivers/accel/habanalabs/common/hldio.c
96
sb = inode->i_sb;
drivers/accel/habanalabs/common/hldio.c
97
bd = sb->s_bdev;
drivers/android/binder/rust_binderfs.c
123
struct super_block *sb = ref_inode->i_sb;
drivers/android/binder/rust_binderfs.c
124
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binder/rust_binderfs.c
158
inode = new_inode(sb);
drivers/android/binder/rust_binderfs.c
180
root = sb->s_root;
drivers/android/binder/rust_binderfs.c
376
static int binderfs_binder_ctl_create(struct super_block *sb)
drivers/android/binder/rust_binderfs.c
382
struct dentry *root = sb->s_root;
drivers/android/binder/rust_binderfs.c
383
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binder/rust_binderfs.c
395
inode = new_inode(sb);
drivers/android/binder/rust_binderfs.c
446
static struct inode *binderfs_make_inode(struct super_block *sb, int mode)
drivers/android/binder/rust_binderfs.c
450
ret = new_inode(sb);
drivers/android/binder/rust_binderfs.c
452
ret->i_ino = iunique(sb, BINDERFS_MAX_MINOR + INODE_OFFSET);
drivers/android/binder/rust_binderfs.c
540
static int init_binder_features(struct super_block *sb)
drivers/android/binder/rust_binderfs.c
544
dir = binderfs_create_dir(sb->s_root, "features");
drivers/android/binder/rust_binderfs.c
569
static int init_binder_logs(struct super_block *sb)
drivers/android/binder/rust_binderfs.c
575
binder_logs_root_dir = binderfs_create_dir(sb->s_root,
drivers/android/binder/rust_binderfs.c
608
info = sb->s_fs_info;
drivers/android/binder/rust_binderfs.c
615
static int binderfs_fill_super(struct super_block *sb, struct fs_context *fc)
drivers/android/binder/rust_binderfs.c
625
sb->s_blocksize = PAGE_SIZE;
drivers/android/binder/rust_binderfs.c
626
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/android/binder/rust_binderfs.c
639
sb->s_iflags &= ~SB_I_NODEV;
drivers/android/binder/rust_binderfs.c
640
sb->s_iflags |= SB_I_NOEXEC;
drivers/android/binder/rust_binderfs.c
641
sb->s_magic = RUST_BINDERFS_SUPER_MAGIC;
drivers/android/binder/rust_binderfs.c
642
sb->s_op = &binderfs_super_ops;
drivers/android/binder/rust_binderfs.c
643
sb->s_time_gran = 1;
drivers/android/binder/rust_binderfs.c
645
sb->s_fs_info = kzalloc_obj(struct binderfs_info);
drivers/android/binder/rust_binderfs.c
646
if (!sb->s_fs_info)
drivers/android/binder/rust_binderfs.c
648
info = sb->s_fs_info;
drivers/android/binder/rust_binderfs.c
652
info->root_gid = make_kgid(sb->s_user_ns, 0);
drivers/android/binder/rust_binderfs.c
655
info->root_uid = make_kuid(sb->s_user_ns, 0);
drivers/android/binder/rust_binderfs.c
661
inode = new_inode(sb);
drivers/android/binder/rust_binderfs.c
672
sb->s_root = d_make_root(inode);
drivers/android/binder/rust_binderfs.c
673
if (!sb->s_root)
drivers/android/binder/rust_binderfs.c
676
ret = binderfs_binder_ctl_create(sb);
drivers/android/binder/rust_binderfs.c
691
ret = init_binder_features(sb);
drivers/android/binder/rust_binderfs.c
696
return init_binder_logs(sb);
drivers/android/binder/rust_binderfs.c
737
static void binderfs_kill_super(struct super_block *sb)
drivers/android/binder/rust_binderfs.c
739
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binder/rust_binderfs.c
745
kill_anon_super(sb);
drivers/android/binder/rust_binderfs.c
91
static inline struct binderfs_info *BINDERFS_SB(const struct super_block *sb)
drivers/android/binder/rust_binderfs.c
93
return sb->s_fs_info;
drivers/android/binderfs.c
123
struct super_block *sb = ref_inode->i_sb;
drivers/android/binderfs.c
124
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binderfs.c
152
inode = new_inode(sb);
drivers/android/binderfs.c
185
root = sb->s_root;
drivers/android/binderfs.c
385
static int binderfs_binder_ctl_create(struct super_block *sb)
drivers/android/binderfs.c
391
struct dentry *root = sb->s_root;
drivers/android/binderfs.c
392
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binderfs.c
404
inode = new_inode(sb);
drivers/android/binderfs.c
456
static struct inode *binderfs_make_inode(struct super_block *sb, int mode)
drivers/android/binderfs.c
460
ret = new_inode(sb);
drivers/android/binderfs.c
462
ret->i_ino = iunique(sb, BINDERFS_MAX_MINOR + INODE_OFFSET);
drivers/android/binderfs.c
475
struct super_block *sb;
drivers/android/binderfs.c
483
sb = parent_inode->i_sb;
drivers/android/binderfs.c
484
new_inode = binderfs_make_inode(sb, S_IFREG | 0444);
drivers/android/binderfs.c
503
struct super_block *sb;
drivers/android/binderfs.c
511
sb = parent_inode->i_sb;
drivers/android/binderfs.c
512
new_inode = binderfs_make_inode(sb, S_IFDIR | 0755);
drivers/android/binderfs.c
539
static int init_binder_features(struct super_block *sb)
drivers/android/binderfs.c
543
dir = binderfs_create_dir(sb->s_root, "features");
drivers/android/binderfs.c
574
static int init_binder_logs(struct super_block *sb)
drivers/android/binderfs.c
581
binder_logs_root_dir = binderfs_create_dir(sb->s_root,
drivers/android/binderfs.c
604
info = sb->s_fs_info;
drivers/android/binderfs.c
611
static int binderfs_fill_super(struct super_block *sb, struct fs_context *fc)
drivers/android/binderfs.c
621
sb->s_blocksize = PAGE_SIZE;
drivers/android/binderfs.c
622
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/android/binderfs.c
635
sb->s_iflags &= ~SB_I_NODEV;
drivers/android/binderfs.c
636
sb->s_iflags |= SB_I_NOEXEC;
drivers/android/binderfs.c
637
sb->s_magic = BINDERFS_SUPER_MAGIC;
drivers/android/binderfs.c
638
sb->s_op = &binderfs_super_ops;
drivers/android/binderfs.c
639
sb->s_time_gran = 1;
drivers/android/binderfs.c
641
sb->s_fs_info = kzalloc_obj(struct binderfs_info);
drivers/android/binderfs.c
642
if (!sb->s_fs_info)
drivers/android/binderfs.c
644
info = sb->s_fs_info;
drivers/android/binderfs.c
648
info->root_gid = make_kgid(sb->s_user_ns, 0);
drivers/android/binderfs.c
651
info->root_uid = make_kuid(sb->s_user_ns, 0);
drivers/android/binderfs.c
657
inode = new_inode(sb);
drivers/android/binderfs.c
668
sb->s_root = d_make_root(inode);
drivers/android/binderfs.c
669
if (!sb->s_root)
drivers/android/binderfs.c
672
ret = binderfs_binder_ctl_create(sb);
drivers/android/binderfs.c
687
ret = init_binder_features(sb);
drivers/android/binderfs.c
692
return init_binder_logs(sb);
drivers/android/binderfs.c
733
static void binderfs_kill_super(struct super_block *sb)
drivers/android/binderfs.c
735
struct binderfs_info *info = sb->s_fs_info;
drivers/android/binderfs.c
741
kill_anon_super(sb);
drivers/android/binderfs.c
83
static inline struct binderfs_info *BINDERFS_SB(const struct super_block *sb)
drivers/android/binderfs.c
85
return sb->s_fs_info;
drivers/android/tests/binder_alloc_kunit.c
118
static void stringify_free_seq(struct kunit *test, int *seq, struct seq_buf *sb)
drivers/android/tests/binder_alloc_kunit.c
123
seq_buf_printf(sb, "[%d]", seq[i]);
drivers/android/tests/binder_alloc_kunit.c
125
KUNIT_EXPECT_FALSE(test, seq_buf_has_overflowed(sb));
drivers/android/tests/binder_alloc_kunit.c
129
struct seq_buf *sb)
drivers/android/tests/binder_alloc_kunit.c
134
seq_buf_printf(sb, "[ %d:%s ]", i,
drivers/android/tests/binder_alloc_kunit.c
137
KUNIT_EXPECT_FALSE(test, seq_buf_has_overflowed(sb));
drivers/ata/libata-scsi.c
252
unsigned char *sb = cmd->sense_buffer;
drivers/ata/libata-scsi.c
260
if ((sb[0] & 0x7f) >= 0x72) {
drivers/ata/libata-scsi.c
265
len = sb[7];
drivers/ata/libata-scsi.c
266
desc = (char *)scsi_sense_desc_find(sb, len + 8, 9);
drivers/ata/libata-scsi.c
270
sb[7] = len + 14;
drivers/ata/libata-scsi.c
271
desc = sb + 8 + len;
drivers/ata/libata-scsi.c
300
sb[0] |= 0x80;
drivers/ata/libata-scsi.c
301
sb[3] = tf->error;
drivers/ata/libata-scsi.c
302
sb[4] = tf->status;
drivers/ata/libata-scsi.c
303
sb[5] = tf->device;
drivers/ata/libata-scsi.c
304
sb[6] = tf->nsect;
drivers/ata/libata-scsi.c
306
sb[8] |= 0x80;
drivers/ata/libata-scsi.c
308
sb[8] |= 0x40;
drivers/ata/libata-scsi.c
310
sb[8] |= 0x20;
drivers/ata/libata-scsi.c
312
sb[9] = tf->lbal;
drivers/ata/libata-scsi.c
313
sb[10] = tf->lbam;
drivers/ata/libata-scsi.c
314
sb[11] = tf->lbah;
drivers/atm/idt77252.c
1070
struct sk_buff *sb;
drivers/atm/idt77252.c
1076
if ((sb = dev_alloc_skb(64)) == NULL) {
drivers/atm/idt77252.c
1082
if (!atm_charge(vcc, sb->truesize)) {
drivers/atm/idt77252.c
1086
dev_kfree_skb(sb);
drivers/atm/idt77252.c
1094
*((u32 *) sb->data) = aal0;
drivers/atm/idt77252.c
1095
skb_put(sb, sizeof(u32));
drivers/atm/idt77252.c
1096
skb_put_data(sb, cell, ATM_CELL_PAYLOAD);
drivers/atm/idt77252.c
1098
ATM_SKB(sb)->vcc = vcc;
drivers/atm/idt77252.c
1099
__net_timestamp(sb);
drivers/atm/idt77252.c
1100
vcc->push(vcc, sb);
drivers/atm/idt77252.c
1148
struct sk_buff *sb;
drivers/atm/idt77252.c
1163
skb_queue_walk(&rpp->queue, sb)
drivers/atm/idt77252.c
1164
skb_put_data(skb, sb->data, sb->len);
drivers/atm/idt77252.c
1246
struct sk_buff *sb;
drivers/atm/idt77252.c
1311
if ((sb = dev_alloc_skb(64)) == NULL) {
drivers/atm/idt77252.c
1318
if (!atm_charge(vcc, sb->truesize)) {
drivers/atm/idt77252.c
1321
dev_kfree_skb(sb);
drivers/atm/idt77252.c
1325
*((u32 *) sb->data) = header;
drivers/atm/idt77252.c
1326
skb_put(sb, sizeof(u32));
drivers/atm/idt77252.c
1327
skb_put_data(sb, &(queue->data[16]), ATM_CELL_PAYLOAD);
drivers/atm/idt77252.c
1329
ATM_SKB(sb)->vcc = vcc;
drivers/atm/idt77252.c
1330
__net_timestamp(sb);
drivers/atm/idt77252.c
1331
vcc->push(vcc, sb);
drivers/atm/nicstar.c
1168
struct sk_buff *sb;
drivers/atm/nicstar.c
1174
sb = dev_alloc_skb(NS_SMSKBSIZE);
drivers/atm/nicstar.c
1175
if (sb == NULL) {
drivers/atm/nicstar.c
1181
NS_PRV_BUFTYPE(sb) = BUF_SM;
drivers/atm/nicstar.c
1182
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
1183
skb_reserve(sb, NS_AAL0_HEADER);
drivers/atm/nicstar.c
1184
push_rxbufs(card, sb);
drivers/atm/nicstar.c
144
static void dequeue_sm_buf(ns_dev * card, struct sk_buff *sb);
drivers/atm/nicstar.c
2029
struct sk_buff *sb;
drivers/atm/nicstar.c
2035
sb = dev_alloc_skb(NS_SMSKBSIZE);
drivers/atm/nicstar.c
2036
if (!sb) {
drivers/atm/nicstar.c
2043
if (!atm_charge(vcc, sb->truesize)) {
drivers/atm/nicstar.c
2048
dev_kfree_skb_any(sb);
drivers/atm/nicstar.c
2052
*((u32 *) sb->data) = le32_to_cpu(rsqe->word_1) << 4 |
drivers/atm/nicstar.c
2055
*((u32 *) sb->data) |= 0x00000002;
drivers/atm/nicstar.c
2056
skb_put(sb, NS_AAL0_HEADER);
drivers/atm/nicstar.c
2057
memcpy(skb_tail_pointer(sb), cell, ATM_CELL_PAYLOAD);
drivers/atm/nicstar.c
2058
skb_put(sb, ATM_CELL_PAYLOAD);
drivers/atm/nicstar.c
2059
ATM_SKB(sb)->vcc = vcc;
drivers/atm/nicstar.c
2060
__net_timestamp(sb);
drivers/atm/nicstar.c
2061
vcc->push(vcc, sb);
drivers/atm/nicstar.c
207
struct sk_buff *sb;
drivers/atm/nicstar.c
2183
struct sk_buff *sb;
drivers/atm/nicstar.c
2185
sb = (struct sk_buff *)(iov - 1)->iov_base;
drivers/atm/nicstar.c
2189
if (!atm_charge(vcc, sb->truesize)) {
drivers/atm/nicstar.c
2190
push_rxbufs(card, sb);
drivers/atm/nicstar.c
2193
skb_put(sb, len);
drivers/atm/nicstar.c
2194
dequeue_sm_buf(card, sb);
drivers/atm/nicstar.c
2195
ATM_SKB(sb)->vcc = vcc;
drivers/atm/nicstar.c
2196
__net_timestamp(sb);
drivers/atm/nicstar.c
2197
vcc->push(vcc, sb);
drivers/atm/nicstar.c
2211
skb_copy_from_linear_data(sb, skb->data,
drivers/atm/nicstar.c
2220
push_rxbufs(card, sb);
drivers/atm/nicstar.c
2226
struct sk_buff *hb, *sb, *lb;
drivers/atm/nicstar.c
2292
sb = (struct sk_buff *)iov->iov_base;
drivers/atm/nicstar.c
2293
skb_copy_from_linear_data(sb, hb->data,
drivers/atm/nicstar.c
2299
push_rxbufs(card, sb);
drivers/atm/nicstar.c
2358
static void dequeue_sm_buf(ns_dev * card, struct sk_buff *sb)
drivers/atm/nicstar.c
2360
skb_unlink(sb, &card->sbpool.queue);
drivers/atm/nicstar.c
244
while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL)
drivers/atm/nicstar.c
245
dev_kfree_skb_any(sb);
drivers/atm/nicstar.c
2584
struct sk_buff *sb;
drivers/atm/nicstar.c
2586
sb = __dev_alloc_skb(NS_SMSKBSIZE, GFP_KERNEL);
drivers/atm/nicstar.c
2587
if (sb == NULL)
drivers/atm/nicstar.c
2589
NS_PRV_BUFTYPE(sb) = BUF_SM;
drivers/atm/nicstar.c
2590
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
2591
skb_reserve(sb, NS_AAL0_HEADER);
drivers/atm/nicstar.c
2592
push_rxbufs(card, sb);
drivers/atm/nicstar.c
714
struct sk_buff *sb;
drivers/atm/nicstar.c
715
sb = __dev_alloc_skb(NS_SMSKBSIZE, GFP_KERNEL);
drivers/atm/nicstar.c
716
if (sb == NULL) {
drivers/atm/nicstar.c
724
NS_PRV_BUFTYPE(sb) = BUF_SM;
drivers/atm/nicstar.c
725
skb_queue_tail(&card->sbpool.queue, sb);
drivers/atm/nicstar.c
726
skb_reserve(sb, NS_AAL0_HEADER);
drivers/atm/nicstar.c
727
push_rxbufs(card, sb);
drivers/atm/nicstar.c
828
struct sk_buff *sb;
drivers/atm/nicstar.c
829
while ((sb = skb_dequeue(&card->sbpool.queue)) != NULL)
drivers/atm/nicstar.c
830
dev_kfree_skb_any(sb);
drivers/base/devtmpfs.c
78
struct super_block *sb = mnt->mnt_sb;
drivers/base/devtmpfs.c
80
atomic_inc(&sb->s_active);
drivers/base/devtmpfs.c
81
down_write(&sb->s_umount);
drivers/base/devtmpfs.c
82
fc->root = dget(sb->s_root);
drivers/block/zloop.c
550
struct super_block *sb = file_inode(zlo->data_dir)->i_sb;
drivers/block/zloop.c
553
down_read(&sb->s_umount);
drivers/block/zloop.c
554
ret = sync_filesystem(sb);
drivers/block/zloop.c
555
up_read(&sb->s_umount);
drivers/bluetooth/hci_bcm.c
962
struct acpi_resource_uart_serialbus *sb;
drivers/bluetooth/hci_bcm.c
982
sb = &ares->data.uart_serial_bus;
drivers/bluetooth/hci_bcm.c
983
if (sb->type == ACPI_RESOURCE_SERIAL_TYPE_UART) {
drivers/bluetooth/hci_bcm.c
984
dev->init_speed = sb->default_baud_rate;
drivers/clk/spear/spear3xx_clock.c
385
static inline void spear320_clk_init(void __iomem *sb, struct clk *rc) { }
drivers/crypto/ccp/ccp-dev-v3.c
27
start = (u32)bitmap_find_next_zero_area(ccp->sb,
drivers/crypto/ccp/ccp-dev-v3.c
275
cr[2] = op->src.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v3.c
284
cr[4] = op->dst.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v3.c
32
bitmap_set(ccp->sb, start, count);
drivers/crypto/ccp/ccp-dev-v3.c
60
bitmap_clear(ccp->sb, start - KSB_START, count);
drivers/crypto/ccp/ccp-dev-v5.c
530
u32 key_addr = op->src.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev-v5.c
542
u32 key_addr = op->dst.u.sb * CCP_SB_BYTES;
drivers/crypto/ccp/ccp-dev.h
418
DECLARE_BITMAP(sb, KSB_COUNT);
drivers/crypto/ccp/ccp-dev.h
489
u32 sb;
drivers/crypto/ccp/ccp-ops.c
431
struct ccp_dm_workarea *wa, u32 jobid, u32 sb,
drivers/crypto/ccp/ccp-ops.c
445
op.src.u.sb = sb;
drivers/crypto/ccp/ccp-ops.c
454
op.dst.u.sb = sb;
drivers/crypto/ccp/ccp-ops.c
463
struct ccp_dm_workarea *wa, u32 jobid, u32 sb,
drivers/crypto/ccp/ccp-ops.c
466
return ccp_copy_to_from_sb(cmd_q, wa, jobid, sb, byte_swap, false);
drivers/crypto/ccp/ccp-ops.c
470
struct ccp_dm_workarea *wa, u32 jobid, u32 sb,
drivers/crypto/ccp/ccp-ops.c
473
return ccp_copy_to_from_sb(cmd_q, wa, jobid, sb, byte_swap, true);
drivers/dax/super.c
352
static struct inode *dax_alloc_inode(struct super_block *sb)
drivers/dax/super.c
357
dax_dev = alloc_inode_sb(sb, dax_cache, GFP_KERNEL);
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1858
static void amdgpu_display_set_pixel(struct drm_scanout_buffer *sb,
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1869
offset = x * 4 + y * sb->pitch[0];
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1880
struct drm_scanout_buffer *sb)
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1899
sb->width = fb->width;
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1900
sb->height = fb->height;
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1902
sb->format = drm_format_info(fb->format->format);
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1903
if (!sb->format)
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1906
sb->pitch[0] = fb->pitches[0];
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1918
sb->set_pixel = amdgpu_display_set_pixel;
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1928
iosys_map_set_vaddr_iomem(&sb->map[0], abo->kmap.virtual);
drivers/gpu/drm/amd/amdgpu/amdgpu_display.c
1930
iosys_map_set_vaddr(&sb->map[0], abo->kmap.virtual);
drivers/gpu/drm/amd/amdgpu/amdgpu_display.h
56
struct drm_scanout_buffer *sb);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
379
const struct dc_surface_update *sb = (struct dc_surface_update *)b;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
382
return sb->surface->layer_index - sa->surface->layer_index;
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8268
out->sb[0] = buf[3]; /* checksum */
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8278
out->sb[0] = 0x01; /* version */
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8279
out->sb[1] = 0x1A; /* length */
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8287
memcpy(&out->sb[i], &buf[4], 26);
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8290
print_hex_dump(KERN_DEBUG, "HDR SB:", DUMP_PREFIX_NONE, 16, 1, out->sb,
drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c
8291
sizeof(out->sb), false);
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4482
ptr = &gamut_packet->sb[3];
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4484
for (i = 0; i <= gamut_packet->sb[1]; i++)
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4487
gamut_packet->sb[2] = (uint8_t) (0x100 - chk_sum);
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4728
check_sum = &hdmi_info.packet_raw_data.sb[0];
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4733
*check_sum += hdmi_info.packet_raw_data.sb[byte_index];
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4743
for (byte_index = 0; byte_index < sizeof(hdmi_info.packet_raw_data.sb); byte_index++)
drivers/gpu/drm/amd/display/dc/core/dc_resource.c
4744
info_packet->sb[byte_index] = hdmi_info.packet_raw_data.sb[byte_index];
drivers/gpu/drm/amd/display/dc/dc_types.h
562
uint8_t sb[32];
drivers/gpu/drm/amd/display/dc/dc_types.h
571
uint8_t sb[128];
drivers/gpu/drm/amd/display/dc/dce/dce_stream_encoder.c
118
(const uint32_t *) &info_packet->sb[0];
drivers/gpu/drm/amd/display/dc/dce/dce_stream_encoder.c
735
(const uint32_t *) &info_frame->avi.sb[0];
drivers/gpu/drm/amd/display/dc/dcn30/dcn30_vpg.c
100
(const uint32_t *) &info_packet->sb[0];
drivers/gpu/drm/amd/display/dc/dio/dcn10/dcn10_stream_encoder.c
108
(const uint32_t *) &info_packet->sb[0];
drivers/gpu/drm/amd/display/dc/dio/dcn20/dcn20_stream_encoder.c
223
const uint32_t *content = (const uint32_t *) &info_packet->sb[0];
drivers/gpu/drm/amd/display/dc/dio/dcn20/dcn20_stream_encoder.c
314
memcpy(&pps_sdp.sb[0], dsc_packed_pps, sizeof(pps_sdp.sb));
drivers/gpu/drm/amd/display/dc/dio/dcn30/dcn30_dio_stream_encoder.c
348
memcpy(pps_sdp.sb, &dsc_packed_pps[i * 32], 32);
drivers/gpu/drm/amd/display/dc/hpo/dcn31/dcn31_hpo_dp_stream_encoder.c
572
memcpy(pps_sdp.sb, &dsc_packed_pps[i * 32], 32);
drivers/gpu/drm/amd/display/dc/link/accessories/link_dp_cts.c
905
pipe_ctx->stream->vsc_infopacket.sb[17] |= (1 << 7); // sb17 bit 7 Dynamic Range: 0 = VESA range, 1 = CTA range
drivers/gpu/drm/amd/display/dc/link/accessories/link_dp_cts.c
907
pipe_ctx->stream->vsc_infopacket.sb[17] &= ~(1 << 7);
drivers/gpu/drm/amd/display/dc/link/accessories/link_dp_cts.c
910
pipe_ctx->stream->vsc_infopacket.sb[16] &= 0xf0;
drivers/gpu/drm/amd/display/dc/link/accessories/link_dp_cts.c
912
pipe_ctx->stream->vsc_infopacket.sb[16] |= 1;
drivers/gpu/drm/amd/display/include/set_mode_types.h
53
uint8_t sb[28]; /* sb0~sb27 */
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
568
infopacket->sb[1] = 0x1A;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
571
infopacket->sb[2] = 0x00;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
574
infopacket->sb[3] = 0x00;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
584
infopacket->sb[6] |= 0x01;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
589
infopacket->sb[6] |= 0x02;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
595
infopacket->sb[6] |= 0x04;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
599
infopacket->sb[6] |= 0x04;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
606
infopacket->sb[7] = (unsigned char)((vrr->min_refresh_in_uhz + 500000) / 1000000);
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
608
infopacket->sb[7] = (unsigned char)((vrr->max_refresh_in_uhz + 500000) / 1000000);
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
614
infopacket->sb[8] = (unsigned char)((vrr->max_refresh_in_uhz + 500000) / 1000000);
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
627
infopacket->sb[1] = 0x1A;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
630
infopacket->sb[2] = 0x00;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
633
infopacket->sb[3] = 0x00;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
643
infopacket->sb[6] |= 0x01;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
648
infopacket->sb[6] |= 0x02;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
654
infopacket->sb[6] |= 0x04;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
658
infopacket->sb[6] |= 0x04;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
671
infopacket->sb[7] = min_programmed & 0xFF;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
674
infopacket->sb[8] = max_refresh & 0xFF;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
677
infopacket->sb[11] = (min_programmed >> 8) & 0x03;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
680
infopacket->sb[12] = (max_refresh >> 8) & 0x03;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
683
infopacket->sb[16] = (vrr->state == VRR_STATE_ACTIVE_FIXED) ? 1 : 0;
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
693
infopacket->sb[9] |= 0x20; // PB9 = [Bit 5 = PQ EOTF Active]
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
695
infopacket->sb[6] |= 0x08; // PB6 = [Bit 3 = Native Color Active]
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
697
infopacket->sb[9] |= 0x04; // PB9 = [Bit 2 = Gamma 2.2 EOTF Active]
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
864
checksum += infopacket->sb[idx];
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
867
infopacket->sb[0] = (unsigned char)(0x100 - checksum);
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
933
infopacket->sb[idx] = infopacket->sb[idx-1];
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
935
infopacket->sb[1] = size; // Length
drivers/gpu/drm/amd/display/modules/freesync/freesync.c
936
infopacket->sb[0] = (infopacket->hb3 >> 2) & 0x3F;//Version
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
249
info_packet->sb[16] = (pixelEncoding << 4) | colorimetryFormat;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
257
info_packet->sb[17] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
260
info_packet->sb[17] = 1;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
263
info_packet->sb[17] = 2;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
266
info_packet->sb[17] = 3;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
270
info_packet->sb[17] = 4;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
273
info_packet->sb[17] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
281
info_packet->sb[17] |= 0x80; /* DB17 bit 7 set to 1 for CEA timing. */
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
291
info_packet->sb[18] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
347
info_packet->sb[i] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
370
info_packet->sb[i] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
392
info_packet->sb[i] = 0;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
440
info_packet->sb[0] = 0x02; // Stacked Frame, Left view is on top and right view on bottom.
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
444
info_packet->sb[0] = 0x01; // Frame/Field Sequential, L + R view indication based on MISC1 bit 2:1
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
448
info_packet->sb[0] = 0x04; // Side-by-side
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
451
info_packet->sb[0] = 0x00; // No Stereo Video, Shall be cleared to 0x0.
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
529
info_packet->sb[1] = 0x03;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
530
info_packet->sb[2] = 0x0C;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
531
info_packet->sb[3] = 0x00;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
534
info_packet->sb[4] = (2 << 5);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
537
info_packet->sb[4] = (1 << 5);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
542
info_packet->sb[5] = (0x0 << 4);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
547
info_packet->sb[5] = (0x8 << 4);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
553
info_packet->sb[5] = (0x6 << 4);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
561
info_packet->sb[5] = stream->timing.hdmi_vic;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
572
checksum += info_packet->sb[i];
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
574
info_packet->sb[0] = (uint8_t) (0x100 - checksum);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
631
info_packet->sb[0] = param->supportMode; //1: AVT; 0: FAVT
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
632
info_packet->sb[1] = (stream->timing.v_total & 0x00FF);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
633
info_packet->sb[2] = (stream->timing.v_total & 0xFF00) >> 8;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
635
info_packet->sb[4] = (param->increase.support << 6 | param->decrease.support << 7);
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
636
info_packet->sb[5] = param->increase.frame_duration_hex;
drivers/gpu/drm/amd/display/modules/info_packet/info_packet.c
637
info_packet->sb[6] = param->decrease.frame_duration_hex;
drivers/gpu/drm/ast/ast_mode.c
612
struct drm_scanout_buffer *sb)
drivers/gpu/drm/ast/ast_mode.c
617
sb->format = plane->state->fb->format;
drivers/gpu/drm/ast/ast_mode.c
618
sb->width = plane->state->fb->width;
drivers/gpu/drm/ast/ast_mode.c
619
sb->height = plane->state->fb->height;
drivers/gpu/drm/ast/ast_mode.c
620
sb->pitch[0] = plane->state->fb->pitches[0];
drivers/gpu/drm/ast/ast_mode.c
621
iosys_map_set_vaddr_iomem(&sb->map[0], ast_plane_vaddr(ast_plane));
drivers/gpu/drm/drm_blend.c
440
const struct drm_plane_state *sb = *(struct drm_plane_state **)b;
drivers/gpu/drm/drm_blend.c
442
if (sa->zpos != sb->zpos)
drivers/gpu/drm/drm_blend.c
443
return sa->zpos - sb->zpos;
drivers/gpu/drm/drm_blend.c
445
return sa->plane->base.id - sb->plane->base.id;
drivers/gpu/drm/drm_fb_dma_helper.c
167
struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_fb_dma_helper.c
190
iosys_map_set_vaddr(&sb->map[0], dma_obj->vaddr);
drivers/gpu/drm/drm_fb_dma_helper.c
191
sb->format = fb->format;
drivers/gpu/drm/drm_fb_dma_helper.c
192
sb->height = fb->height;
drivers/gpu/drm/drm_fb_dma_helper.c
193
sb->width = fb->width;
drivers/gpu/drm/drm_fb_dma_helper.c
194
sb->pitch[0] = fb->pitches[0];
drivers/gpu/drm/drm_panic.c
141
static void drm_panic_blit_pixel(struct drm_scanout_buffer *sb, struct drm_rect *clip,
drivers/gpu/drm/drm_panic.c
150
sb->set_pixel(sb, clip->x1 + x, clip->y1 + y, fg_color);
drivers/gpu/drm/drm_panic.c
284
static void drm_panic_blit(struct drm_scanout_buffer *sb, struct drm_rect *clip,
drivers/gpu/drm/drm_panic.c
291
if (sb->set_pixel)
drivers/gpu/drm/drm_panic.c
292
return drm_panic_blit_pixel(sb, clip, sbuf8, spitch, scale, fg_color);
drivers/gpu/drm/drm_panic.c
294
if (sb->pages)
drivers/gpu/drm/drm_panic.c
295
return drm_panic_blit_page(sb->pages, sb->pitch[0], sb->format->cpp[0],
drivers/gpu/drm/drm_panic.c
298
map = sb->map[0];
drivers/gpu/drm/drm_panic.c
299
iosys_map_incr(&map, clip->y1 * sb->pitch[0] + clip->x1 * sb->format->cpp[0]);
drivers/gpu/drm/drm_panic.c
301
switch (sb->format->cpp[0]) {
drivers/gpu/drm/drm_panic.c
303
drm_draw_blit16(&map, sb->pitch[0], sbuf8, spitch,
drivers/gpu/drm/drm_panic.c
307
drm_draw_blit24(&map, sb->pitch[0], sbuf8, spitch,
drivers/gpu/drm/drm_panic.c
311
drm_draw_blit32(&map, sb->pitch[0], sbuf8, spitch,
drivers/gpu/drm/drm_panic.c
315
WARN_ONCE(1, "Can't blit with pixel width %d\n", sb->format->cpp[0]);
drivers/gpu/drm/drm_panic.c
319
static void drm_panic_fill_pixel(struct drm_scanout_buffer *sb,
drivers/gpu/drm/drm_panic.c
327
sb->set_pixel(sb, clip->x1 + x, clip->y1 + y, color);
drivers/gpu/drm/drm_panic.c
375
static void drm_panic_fill(struct drm_scanout_buffer *sb, struct drm_rect *clip,
drivers/gpu/drm/drm_panic.c
380
if (sb->set_pixel)
drivers/gpu/drm/drm_panic.c
381
return drm_panic_fill_pixel(sb, clip, color);
drivers/gpu/drm/drm_panic.c
383
if (sb->pages)
drivers/gpu/drm/drm_panic.c
384
return drm_panic_fill_page(sb->pages, sb->pitch[0], sb->format->cpp[0],
drivers/gpu/drm/drm_panic.c
387
map = sb->map[0];
drivers/gpu/drm/drm_panic.c
388
iosys_map_incr(&map, clip->y1 * sb->pitch[0] + clip->x1 * sb->format->cpp[0]);
drivers/gpu/drm/drm_panic.c
390
switch (sb->format->cpp[0]) {
drivers/gpu/drm/drm_panic.c
392
drm_draw_fill16(&map, sb->pitch[0], drm_rect_height(clip),
drivers/gpu/drm/drm_panic.c
396
drm_draw_fill24(&map, sb->pitch[0], drm_rect_height(clip),
drivers/gpu/drm/drm_panic.c
400
drm_draw_fill32(&map, sb->pitch[0], drm_rect_height(clip),
drivers/gpu/drm/drm_panic.c
404
WARN_ONCE(1, "Can't fill with pixel width %d\n", sb->format->cpp[0]);
drivers/gpu/drm/drm_panic.c
421
static void draw_txt_rectangle(struct drm_scanout_buffer *sb,
drivers/gpu/drm/drm_panic.c
448
drm_panic_blit(sb, &rec, src, font_pitch, 1, color);
drivers/gpu/drm/drm_panic.c
465
static void drm_panic_logo_draw(struct drm_scanout_buffer *sb, struct drm_rect *rect,
drivers/gpu/drm/drm_panic.c
468
if (rect->x2 > sb->width || rect->y2 > sb->height)
drivers/gpu/drm/drm_panic.c
472
drm_panic_blit(sb, rect, logo_mono->data,
drivers/gpu/drm/drm_panic.c
475
draw_txt_rectangle(sb, font, logo_ascii, logo_ascii_lines, false, rect,
drivers/gpu/drm/drm_panic.c
479
static void draw_panic_screen_user(struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_panic.c
482
sb->format->format);
drivers/gpu/drm/drm_panic.c
484
sb->format->format);
drivers/gpu/drm/drm_panic.c
485
const struct font_desc *font = get_default_font(sb->width, sb->height, NULL, NULL);
drivers/gpu/drm/drm_panic.c
492
r_screen = DRM_RECT_INIT(0, 0, sb->width, sb->height);
drivers/gpu/drm/drm_panic.c
495
msg_width = min(get_max_line_len(panic_msg, panic_msg_lines) * font->width, sb->width);
drivers/gpu/drm/drm_panic.c
496
msg_height = min(panic_msg_lines * font->height, sb->height);
drivers/gpu/drm/drm_panic.c
500
drm_rect_translate(&r_msg, (sb->width - r_msg.x2) / 2, (sb->height - r_msg.y2) / 2);
drivers/gpu/drm/drm_panic.c
503
drm_panic_fill(sb, &r_screen, bg_color);
drivers/gpu/drm/drm_panic.c
506
drm_panic_logo_draw(sb, &r_logo, font, fg_color);
drivers/gpu/drm/drm_panic.c
508
draw_txt_rectangle(sb, font, panic_msg, panic_msg_lines, true, &r_msg, fg_color);
drivers/gpu/drm/drm_panic.c
515
static int draw_line_with_wrap(struct drm_scanout_buffer *sb, const struct font_desc *font,
drivers/gpu/drm/drm_panic.c
518
int chars_per_row = sb->width / font->width;
drivers/gpu/drm/drm_panic.c
519
struct drm_rect r_txt = DRM_RECT_INIT(0, yoffset, sb->width, font->height);
drivers/gpu/drm/drm_panic.c
525
draw_txt_rectangle(sb, font, &line_wrap, 1, false, &r_txt, fg_color);
drivers/gpu/drm/drm_panic.c
532
draw_txt_rectangle(sb, font, &line_wrap, 1, false, &r_txt, fg_color);
drivers/gpu/drm/drm_panic.c
538
draw_txt_rectangle(sb, font, line, 1, false, &r_txt, fg_color);
drivers/gpu/drm/drm_panic.c
548
static void draw_panic_screen_kmsg(struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_panic.c
551
sb->format->format);
drivers/gpu/drm/drm_panic.c
553
sb->format->format);
drivers/gpu/drm/drm_panic.c
554
const struct font_desc *font = get_default_font(sb->width, sb->height, NULL, NULL);
drivers/gpu/drm/drm_panic.c
555
struct drm_rect r_screen = DRM_RECT_INIT(0, 0, sb->width, sb->height);
drivers/gpu/drm/drm_panic.c
562
if (!font || font->width > sb->width)
drivers/gpu/drm/drm_panic.c
565
yoffset = sb->height - font->height - (sb->height % font->height) / 2;
drivers/gpu/drm/drm_panic.c
568
drm_panic_fill(sb, &r_screen, bg_color);
drivers/gpu/drm/drm_panic.c
585
yoffset = draw_line_with_wrap(sb, font, &line, yoffset, fg_color);
drivers/gpu/drm/drm_panic.c
736
static int _draw_panic_screen_qr_code(struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_panic.c
739
sb->format->format);
drivers/gpu/drm/drm_panic.c
741
sb->format->format);
drivers/gpu/drm/drm_panic.c
742
const struct font_desc *font = get_default_font(sb->width, sb->height, NULL, NULL);
drivers/gpu/drm/drm_panic.c
752
r_screen = DRM_RECT_INIT(0, 0, sb->width, sb->height);
drivers/gpu/drm/drm_panic.c
756
msg_width = min(get_max_line_len(panic_msg, panic_msg_lines) * font->width, sb->width);
drivers/gpu/drm/drm_panic.c
757
msg_height = min(panic_msg_lines * font->height, sb->height);
drivers/gpu/drm/drm_panic.c
760
max_qr_size = min(3 * sb->width / 4, 3 * sb->height / 4);
drivers/gpu/drm/drm_panic.c
775
v_margin = sb->height - drm_rect_height(&r_qr_canvas) - drm_rect_height(&r_msg);
drivers/gpu/drm/drm_panic.c
780
drm_rect_translate(&r_qr_canvas, (sb->width - r_qr_canvas.x2) / 2, 2 * v_margin);
drivers/gpu/drm/drm_panic.c
785
drm_rect_translate(&r_msg, (sb->width - r_msg.x2) / 2,
drivers/gpu/drm/drm_panic.c
789
drm_panic_fill(sb, &r_screen, bg_color);
drivers/gpu/drm/drm_panic.c
792
drm_panic_logo_draw(sb, &r_logo, font, fg_color);
drivers/gpu/drm/drm_panic.c
794
draw_txt_rectangle(sb, font, panic_msg, panic_msg_lines, true, &r_msg, fg_color);
drivers/gpu/drm/drm_panic.c
798
drm_panic_fill(sb, &r_qr_canvas, fg_color);
drivers/gpu/drm/drm_panic.c
799
drm_panic_fill(sb, &r_qr, bg_color);
drivers/gpu/drm/drm_panic.c
800
drm_panic_blit(sb, &r_qr, qr_image, qr_pitch, scale, fg_color);
drivers/gpu/drm/drm_panic.c
804
static void draw_panic_screen_qr_code(struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_panic.c
806
if (_draw_panic_screen_qr_code(sb))
drivers/gpu/drm/drm_panic.c
807
draw_panic_screen_user(sb);
drivers/gpu/drm/drm_panic.c
878
static void draw_panic_dispatch(struct drm_scanout_buffer *sb)
drivers/gpu/drm/drm_panic.c
882
draw_panic_screen_kmsg(sb);
drivers/gpu/drm/drm_panic.c
887
draw_panic_screen_qr_code(sb);
drivers/gpu/drm/drm_panic.c
893
draw_panic_screen_user(sb);
drivers/gpu/drm/drm_panic.c
923
struct drm_scanout_buffer sb = { };
drivers/gpu/drm/drm_panic.c
930
ret = plane->helper_private->get_scanout_buffer(plane, &sb);
drivers/gpu/drm/drm_panic.c
932
if (ret || !drm_panic_is_format_supported(sb.format))
drivers/gpu/drm/drm_panic.c
936
if (!sb.set_pixel && !sb.pages && iosys_map_is_null(&sb.map[0]))
drivers/gpu/drm/drm_panic.c
941
draw_panic_dispatch(&sb);
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
194
struct drm_scanout_buffer *sb)
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
200
sb->format = plane->state->fb->format;
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
201
sb->width = plane->state->fb->width;
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
202
sb->height = plane->state->fb->height;
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
203
sb->pitch[0] = plane->state->fb->pitches[0];
drivers/gpu/drm/hyperv/hyperv_drm_modeset.c
204
sb->map[0] = map;
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
445
struct acpi_resource_i2c_serialbus *sb;
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
450
if (!i2c_acpi_get_i2c_resource(ares, &sb))
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
453
if (lookup->target_addr != sb->slave_address)
drivers/gpu/drm/i915/display/intel_dsi_vbt.c
457
sb->resource_source.string_ptr,
drivers/gpu/drm/i915/display/intel_parent.c
68
int intel_parent_panic_setup(struct intel_display *display, struct intel_panic *panic, struct drm_scanout_buffer *sb)
drivers/gpu/drm/i915/display/intel_parent.c
70
return display->parent->panic->setup(panic, sb);
drivers/gpu/drm/i915/display/intel_parent.h
32
int intel_parent_panic_setup(struct intel_display *display, struct intel_panic *panic, struct drm_scanout_buffer *sb);
drivers/gpu/drm/i915/display/intel_plane.c
1402
struct drm_scanout_buffer *sb)
drivers/gpu/drm/i915/display/intel_plane.c
1420
intel_fbdev_get_map(display->fbdev.fbdev, &sb->map[0]);
drivers/gpu/drm/i915/display/intel_plane.c
1431
sb->private = fb;
drivers/gpu/drm/i915/display/intel_plane.c
1432
ret = intel_parent_panic_setup(display, fb->panic, sb);
drivers/gpu/drm/i915/display/intel_plane.c
1436
sb->width = fb->base.width;
drivers/gpu/drm/i915/display/intel_plane.c
1437
sb->height = fb->base.height;
drivers/gpu/drm/i915/display/intel_plane.c
1441
sb->format = drm_format_info(fb->base.format->format);
drivers/gpu/drm/i915/display/intel_plane.c
1442
sb->pitch[0] = fb->base.pitches[0];
drivers/gpu/drm/i915/gem/i915_gem_object.h
697
int i915_gem_object_panic_setup(struct intel_panic *panic, struct drm_scanout_buffer *sb,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
395
static void i915_gem_object_panic_map_set_pixel(struct drm_scanout_buffer *sb, unsigned int x,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
398
struct intel_framebuffer *fb = (struct intel_framebuffer *)sb->private;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
399
unsigned int offset = fb->panic_tiling(sb->width, x, y);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
401
iosys_map_wr(&sb->map[0], offset, u32, color);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
409
static void i915_gem_object_panic_page_set_pixel(struct drm_scanout_buffer *sb, unsigned int x,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
414
struct intel_framebuffer *fb = (struct intel_framebuffer *)sb->private;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
418
offset = fb->panic_tiling(sb->width, x, y);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
420
offset = y * sb->pitch[0] + x * sb->format->cpp[0];
drivers/gpu/drm/i915/gem/i915_gem_pages.c
450
int i915_gem_object_panic_setup(struct intel_panic *panic, struct drm_scanout_buffer *sb,
drivers/gpu/drm/i915/gem/i915_gem_pages.c
460
iosys_map_set_vaddr_iomem(&sb->map[0], (void __iomem *)ptr);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
462
iosys_map_set_vaddr(&sb->map[0], ptr);
drivers/gpu/drm/i915/gem/i915_gem_pages.c
465
sb->set_pixel = i915_gem_object_panic_map_set_pixel;
drivers/gpu/drm/i915/gem/i915_gem_pages.c
473
sb->set_pixel = i915_gem_object_panic_page_set_pixel;
drivers/gpu/drm/i915/i915_panic.c
18
static int intel_panic_setup(struct intel_panic *panic, struct drm_scanout_buffer *sb)
drivers/gpu/drm/i915/i915_panic.c
20
struct intel_framebuffer *fb = sb->private;
drivers/gpu/drm/i915/i915_panic.c
23
return i915_gem_object_panic_setup(panic, sb, obj, fb->panic_tiling);
drivers/gpu/drm/i915/intel_wakeref.c
207
char *buf, *sb, *se;
drivers/gpu/drm/i915/intel_wakeref.c
218
for (sb = buf; *sb; sb = se + 1) {
drivers/gpu/drm/i915/intel_wakeref.c
219
se = strchrnul(sb, '\n');
drivers/gpu/drm/i915/intel_wakeref.c
220
drm_printf(p, "%.*s", (int)(se - sb + 1), sb);
drivers/gpu/drm/mgag200/mgag200_drv.h
375
struct drm_scanout_buffer *sb);
drivers/gpu/drm/mgag200/mgag200_mode.c
561
struct drm_scanout_buffer *sb)
drivers/gpu/drm/mgag200/mgag200_mode.c
567
sb->format = plane->state->fb->format;
drivers/gpu/drm/mgag200/mgag200_mode.c
568
sb->width = plane->state->fb->width;
drivers/gpu/drm/mgag200/mgag200_mode.c
569
sb->height = plane->state->fb->height;
drivers/gpu/drm/mgag200/mgag200_mode.c
570
sb->pitch[0] = plane->state->fb->pitches[0];
drivers/gpu/drm/mgag200/mgag200_mode.c
571
sb->map[0] = map;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
616
nv50_set_pixel_swizzle(struct drm_scanout_buffer *sb, unsigned int x,
drivers/gpu/drm/nouveau/dispnv50/wndw.c
621
blk_off = nv50_get_block_off(x, y, sb->pitch[0]);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
632
iosys_map_wr(&sb->map[0], off, u32, color);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
636
nv50_set_pixel(struct drm_scanout_buffer *sb, unsigned int x, unsigned int y,
drivers/gpu/drm/nouveau/dispnv50/wndw.c
641
blk_off = nv50_get_block_off(x, y, sb->width);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
647
iosys_map_wr(&sb->map[0], off, u32, color);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
651
nv50_wndw_get_scanout_buffer(struct drm_plane *plane, struct drm_scanout_buffer *sb)
drivers/gpu/drm/nouveau/dispnv50/wndw.c
677
iosys_map_set_vaddr_iomem(&sb->map[0], (void __iomem *)nvbo->kmap.virtual);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
679
iosys_map_set_vaddr(&sb->map[0], nvbo->kmap.virtual);
drivers/gpu/drm/nouveau/dispnv50/wndw.c
681
sb->height = fb->height;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
682
sb->width = fb->width;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
683
sb->pitch[0] = fb->pitches[0];
drivers/gpu/drm/nouveau/dispnv50/wndw.c
684
sb->format = fb->format;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
697
sb->set_pixel = nv50_set_pixel_swizzle;
drivers/gpu/drm/nouveau/dispnv50/wndw.c
699
sb->set_pixel = nv50_set_pixel;
drivers/gpu/drm/nouveau/include/nvkm/subdev/gsp.h
82
struct nvkm_falcon_fw sb;
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/fwsec.c
325
ret = nvkm_gsp_fwsec_boot(gsp, &gsp->fws.falcon.sb);
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/fwsec.c
342
return nvkm_gsp_fwsec_init(gsp, &gsp->fws.falcon.sb, "fwsec-sb",
drivers/gpu/drm/nouveau/nvkm/subdev/gsp/tu102.c
42
nvkm_falcon_fw_dtor(&gsp->fws.falcon.sb);
drivers/gpu/drm/omapdrm/omap_drv.c
123
const struct drm_plane_state *sb = *(struct drm_plane_state **)b;
drivers/gpu/drm/omapdrm/omap_drv.c
125
if (sa->normalized_zpos != sb->normalized_zpos)
drivers/gpu/drm/omapdrm/omap_drv.c
126
return sa->normalized_zpos - sb->normalized_zpos;
drivers/gpu/drm/omapdrm/omap_drv.c
128
return sa->plane->base.id - sb->plane->base.id;
drivers/gpu/drm/qxl/qxl_kms.c
111
int r, sb;
drivers/gpu/drm/qxl/qxl_kms.c
134
sb = 4;
drivers/gpu/drm/qxl/qxl_kms.c
135
qdev->surfaceram_base = pci_resource_start(pdev, sb);
drivers/gpu/drm/qxl/qxl_kms.c
136
qdev->surfaceram_size = pci_resource_len(pdev, sb);
drivers/gpu/drm/qxl/qxl_kms.c
143
sb = 1;
drivers/gpu/drm/qxl/qxl_kms.c
144
qdev->surfaceram_base = pci_resource_start(pdev, sb);
drivers/gpu/drm/qxl/qxl_kms.c
145
qdev->surfaceram_size = pci_resource_len(pdev, sb);
drivers/gpu/drm/qxl/qxl_kms.c
162
(unsigned long long)pci_resource_end(pdev, sb),
drivers/gpu/drm/qxl/qxl_kms.c
165
(sb == 4) ? "64bit" : "32bit");
drivers/gpu/drm/sysfb/drm_sysfb_helper.h
120
struct drm_scanout_buffer *sb);
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
409
struct drm_scanout_buffer *sb)
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
413
sb->width = sysfb->fb_mode.hdisplay;
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
414
sb->height = sysfb->fb_mode.vdisplay;
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
415
sb->format = sysfb->fb_format;
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
416
sb->pitch[0] = sysfb->fb_pitch;
drivers/gpu/drm/sysfb/drm_sysfb_modeset.c
417
sb->map[0] = sysfb->fb_addr;
drivers/gpu/drm/tests/drm_panic_test.c
100
iosys_map_set_vaddr(&sb->map[0], fb);
drivers/gpu/drm/tests/drm_panic_test.c
101
sb->width = params->width;
drivers/gpu/drm/tests/drm_panic_test.c
102
sb->height = params->height;
drivers/gpu/drm/tests/drm_panic_test.c
103
sb->pitch[0] = params->width * sb->format->cpp[0];
drivers/gpu/drm/tests/drm_panic_test.c
105
params->draw_screen(sb);
drivers/gpu/drm/tests/drm_panic_test.c
120
struct drm_scanout_buffer *sb = test->priv;
drivers/gpu/drm/tests/drm_panic_test.c
126
sb->format = drm_format_info(params->format);
drivers/gpu/drm/tests/drm_panic_test.c
127
fb_size = params->width * params->height * sb->format->cpp[0];
drivers/gpu/drm/tests/drm_panic_test.c
144
sb->pages = pages;
drivers/gpu/drm/tests/drm_panic_test.c
145
sb->width = params->width;
drivers/gpu/drm/tests/drm_panic_test.c
146
sb->height = params->height;
drivers/gpu/drm/tests/drm_panic_test.c
147
sb->pitch[0] = params->width * sb->format->cpp[0];
drivers/gpu/drm/tests/drm_panic_test.c
149
params->draw_screen(sb);
drivers/gpu/drm/tests/drm_panic_test.c
167
static void drm_test_panic_set_pixel(struct drm_scanout_buffer *sb,
drivers/gpu/drm/tests/drm_panic_test.c
172
struct kunit *test = (struct kunit *)sb->private;
drivers/gpu/drm/tests/drm_panic_test.c
174
KUNIT_ASSERT_TRUE(test, x < sb->width && y < sb->height);
drivers/gpu/drm/tests/drm_panic_test.c
183
struct drm_scanout_buffer *sb = test->priv;
drivers/gpu/drm/tests/drm_panic_test.c
186
sb->format = drm_format_info(params->format);
drivers/gpu/drm/tests/drm_panic_test.c
187
sb->set_pixel = drm_test_panic_set_pixel;
drivers/gpu/drm/tests/drm_panic_test.c
188
sb->width = params->width;
drivers/gpu/drm/tests/drm_panic_test.c
189
sb->height = params->height;
drivers/gpu/drm/tests/drm_panic_test.c
190
sb->private = test;
drivers/gpu/drm/tests/drm_panic_test.c
192
params->draw_screen(sb);
drivers/gpu/drm/tests/drm_panic_test.c
33
void (*draw_screen)(struct drm_scanout_buffer *sb);
drivers/gpu/drm/tests/drm_panic_test.c
86
struct drm_scanout_buffer *sb = test->priv;
drivers/gpu/drm/tests/drm_panic_test.c
92
sb->format = drm_format_info(params->format);
drivers/gpu/drm/tests/drm_panic_test.c
93
fb_size = params->width * params->height * sb->format->cpp[0];
drivers/gpu/drm/tiny/bochs.c
477
struct drm_scanout_buffer *sb)
drivers/gpu/drm/tiny/bochs.c
483
sb->format = plane->state->fb->format;
drivers/gpu/drm/tiny/bochs.c
484
sb->width = plane->state->fb->width;
drivers/gpu/drm/tiny/bochs.c
485
sb->height = plane->state->fb->height;
drivers/gpu/drm/tiny/bochs.c
486
sb->pitch[0] = plane->state->fb->pitches[0];
drivers/gpu/drm/tiny/bochs.c
487
sb->map[0] = map;
drivers/gpu/drm/virtio/virtgpu_plane.c
503
struct drm_scanout_buffer *sb)
drivers/gpu/drm/virtio/virtgpu_plane.c
516
iosys_map_set_vaddr(&sb->map[0], bo->base.vaddr);
drivers/gpu/drm/virtio/virtgpu_plane.c
523
sb->pages = shmem->pages;
drivers/gpu/drm/virtio/virtgpu_plane.c
526
sb->format = plane->state->fb->format;
drivers/gpu/drm/virtio/virtgpu_plane.c
527
sb->height = plane->state->fb->height;
drivers/gpu/drm/virtio/virtgpu_plane.c
528
sb->width = plane->state->fb->width;
drivers/gpu/drm/virtio/virtgpu_plane.c
529
sb->pitch[0] = plane->state->fb->pitches[0];
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
976
struct vmw_dx_shader_bindings *sb = &cbs->per_shader[0];
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
980
for (i = 0; i < SVGA3D_NUM_SHADERTYPE_DX10; ++i, ++sb) {
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
981
if (!test_bit(VMW_BINDING_PS_SR_BIT, &sb->dirty))
drivers/gpu/drm/vmwgfx/vmwgfx_binding.c
988
__clear_bit(VMW_BINDING_PS_SR_BIT, &sb->dirty);
drivers/gpu/drm/xe/display/xe_panic.c
36
static void xe_panic_page_set_pixel(struct drm_scanout_buffer *sb, unsigned int x,
drivers/gpu/drm/xe/display/xe_panic.c
39
struct intel_framebuffer *fb = (struct intel_framebuffer *)sb->private;
drivers/gpu/drm/xe/display/xe_panic.c
46
offset = fb->panic_tiling(sb->width, x, y);
drivers/gpu/drm/xe/display/xe_panic.c
48
offset = y * sb->pitch[0] + x * sb->format->cpp[0];
drivers/gpu/drm/xe/display/xe_panic.c
87
static int xe_panic_setup(struct intel_panic *panic, struct drm_scanout_buffer *sb)
drivers/gpu/drm/xe/display/xe_panic.c
89
struct intel_framebuffer *fb = (struct intel_framebuffer *)sb->private;
drivers/gpu/drm/xe/display/xe_panic.c
96
sb->set_pixel = xe_panic_page_set_pixel;
drivers/hwmon/hwmon.c
1292
struct pci_dev *sb;
drivers/hwmon/hwmon.c
1297
sb = pci_get_device(PCI_VENDOR_ID_ATI, 0x436c, NULL);
drivers/hwmon/hwmon.c
1298
if (sb) {
drivers/hwmon/hwmon.c
1299
if (sb->subsystem_vendor == 0x1462 && /* MSI */
drivers/hwmon/hwmon.c
1300
sb->subsystem_device == 0x0031) { /* MS-7031 */
drivers/hwmon/hwmon.c
1301
pci_read_config_byte(sb, 0x48, &enable);
drivers/hwmon/hwmon.c
1302
pci_read_config_word(sb, 0x64, &base);
drivers/hwmon/hwmon.c
1305
dev_info(&sb->dev,
drivers/hwmon/hwmon.c
1307
pci_write_config_word(sb, 0x64, 0x295);
drivers/hwmon/hwmon.c
1308
pci_write_config_byte(sb, 0x48,
drivers/hwmon/hwmon.c
1312
pci_dev_put(sb);
drivers/i2c/algos/i2c-algo-bit.c
155
int sb;
drivers/i2c/algos/i2c-algo-bit.c
161
sb = (c >> i) & 1;
drivers/i2c/algos/i2c-algo-bit.c
162
setsda(adap, sb);
drivers/i2c/i2c-core-acpi.c
108
struct acpi_resource_i2c_serialbus *sb;
drivers/i2c/i2c-core-acpi.c
111
if (info->addr || !i2c_acpi_get_i2c_resource(ares, &sb))
drivers/i2c/i2c-core-acpi.c
118
sb->resource_source.string_ptr,
drivers/i2c/i2c-core-acpi.c
123
info->addr = sb->slave_address;
drivers/i2c/i2c-core-acpi.c
124
lookup->speed = sb->connection_speed;
drivers/i2c/i2c-core-acpi.c
125
if (sb->access_mode == ACPI_I2C_10BIT_MODE)
drivers/i2c/i2c-core-acpi.c
58
struct acpi_resource_i2c_serialbus *sb;
drivers/i2c/i2c-core-acpi.c
63
sb = &ares->data.i2c_serial_bus;
drivers/i2c/i2c-core-acpi.c
64
if (sb->type != ACPI_RESOURCE_SERIAL_TYPE_I2C)
drivers/i2c/i2c-core-acpi.c
67
*i2c = sb;
drivers/i2c/i2c-core-acpi.c
673
struct acpi_resource_i2c_serialbus *sb;
drivers/i2c/i2c-core-acpi.c
692
if (!value64 || !i2c_acpi_get_i2c_resource(ares, &sb)) {
drivers/i2c/i2c-core-acpi.c
698
client->addr = sb->slave_address;
drivers/i2c/i2c-core-acpi.c
700
if (sb->access_mode == ACPI_I2C_10BIT_MODE)
drivers/i2c/i2c-core-acpi.c
74
struct acpi_resource_i2c_serialbus *sb;
drivers/i2c/i2c-core-acpi.c
77
if (i2c_acpi_get_i2c_resource(ares, &sb))
drivers/iio/imu/inv_mpu6050/inv_mpu_acpi.c
88
struct acpi_resource_i2c_serialbus *sb;
drivers/iio/imu/inv_mpu6050/inv_mpu_acpi.c
91
if (i2c_acpi_get_i2c_resource(ares, &sb)) {
drivers/iio/imu/inv_mpu6050/inv_mpu_acpi.c
93
*addr |= (sb->slave_address << 16);
drivers/iio/imu/inv_mpu6050/inv_mpu_acpi.c
95
*addr = sb->slave_address;
drivers/infiniband/core/cma.c
462
struct sockaddr *sb = cma_dst_addr(id_priv);
drivers/infiniband/core/cma.c
467
if (sa->sa_family != sb->sa_family)
drivers/infiniband/core/cma.c
468
return sa->sa_family - sb->sa_family;
drivers/infiniband/core/cma.c
473
&((struct sockaddr_in *)sb)->sin_addr,
drivers/infiniband/core/cma.c
480
&((struct sockaddr_in6 *)sb)->sin6_addr);
drivers/infiniband/hw/bng_re/bng_fw.c
346
if (__get_cmdq_base_resp_size(msg->req, msg->req_sz) && msg->sb) {
drivers/infiniband/hw/bng_re/bng_fw.c
347
struct bng_re_rcfw_sbuf *sbuf = msg->sb;
drivers/infiniband/hw/bng_re/bng_fw.h
111
void *sb;
drivers/infiniband/hw/bng_re/bng_fw.h
134
void *sb;
drivers/infiniband/hw/bng_re/bng_fw.h
148
void *req, void *resp, void *sb,
drivers/infiniband/hw/bng_re/bng_fw.h
153
msg->sb = sb;
drivers/infiniband/hw/bng_re/bng_sp.c
100
attr->max_inline_data = le32_to_cpu(sb->max_inline_data);
drivers/infiniband/hw/bng_re/bng_sp.c
107
attr->max_sgid = le32_to_cpu(sb->max_gid);
drivers/infiniband/hw/bng_re/bng_sp.c
109
attr->dev_cap_flags = le16_to_cpu(sb->dev_cap_flags);
drivers/infiniband/hw/bng_re/bng_sp.c
110
attr->dev_cap_flags2 = le16_to_cpu(sb->dev_cap_ext_flags_2);
drivers/infiniband/hw/bng_re/bng_sp.c
113
attr->max_srq += le16_to_cpu(sb->max_srq_ext);
drivers/infiniband/hw/bng_re/bng_sp.c
117
temp = le32_to_cpu(sb->tqm_alloc_reqs[i]);
drivers/infiniband/hw/bng_re/bng_sp.c
125
attr->max_dpi = le32_to_cpu(sb->max_dpi);
drivers/infiniband/hw/bng_re/bng_sp.c
129
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bng_re/bng_sp.c
46
struct creq_query_func_resp_sb *sb;
drivers/infiniband/hw/bng_re/bng_sp.c
57
sbuf.size = ALIGN(sizeof(*sb), BNG_FW_CMDQE_UNITS);
drivers/infiniband/hw/bng_re/bng_sp.c
58
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bng_re/bng_sp.c
60
if (!sbuf.sb)
drivers/infiniband/hw/bng_re/bng_sp.c
62
sb = sbuf.sb;
drivers/infiniband/hw/bng_re/bng_sp.c
70
attr->max_qp = le32_to_cpu(sb->max_qp);
drivers/infiniband/hw/bng_re/bng_sp.c
74
sb->max_qp_rd_atom > BNG_RE_MAX_OUT_RD_ATOM ?
drivers/infiniband/hw/bng_re/bng_sp.c
75
BNG_RE_MAX_OUT_RD_ATOM : sb->max_qp_rd_atom;
drivers/infiniband/hw/bng_re/bng_sp.c
77
sb->max_qp_init_rd_atom > BNG_RE_MAX_OUT_RD_ATOM ?
drivers/infiniband/hw/bng_re/bng_sp.c
78
BNG_RE_MAX_OUT_RD_ATOM : sb->max_qp_init_rd_atom;
drivers/infiniband/hw/bng_re/bng_sp.c
79
attr->max_qp_wqes = le16_to_cpu(sb->max_qp_wr) - 1;
drivers/infiniband/hw/bng_re/bng_sp.c
84
attr->max_qp_sges = min_t(u32, sb->max_sge_var_wqe, BNG_VAR_MAX_SGE);
drivers/infiniband/hw/bng_re/bng_sp.c
85
attr->max_cq = le32_to_cpu(sb->max_cq);
drivers/infiniband/hw/bng_re/bng_sp.c
86
attr->max_cq_wqes = le32_to_cpu(sb->max_cqe);
drivers/infiniband/hw/bng_re/bng_sp.c
88
attr->max_mr = le32_to_cpu(sb->max_mr);
drivers/infiniband/hw/bng_re/bng_sp.c
89
attr->max_mw = le32_to_cpu(sb->max_mw);
drivers/infiniband/hw/bng_re/bng_sp.c
91
attr->max_mr_size = le64_to_cpu(sb->max_mr_size);
drivers/infiniband/hw/bng_re/bng_sp.c
93
attr->max_raw_ethy_qp = le32_to_cpu(sb->max_raw_eth_qp);
drivers/infiniband/hw/bng_re/bng_sp.c
94
attr->max_ah = le32_to_cpu(sb->max_ah);
drivers/infiniband/hw/bng_re/bng_sp.c
96
attr->max_srq = le16_to_cpu(sb->max_srq);
drivers/infiniband/hw/bng_re/bng_sp.c
97
attr->max_srq_wqes = le32_to_cpu(sb->max_srq_wr) - 1;
drivers/infiniband/hw/bng_re/bng_sp.c
98
attr->max_srq_sges = sb->max_srq_sge;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1451
struct creq_query_qp_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1456
sbuf.size = ALIGN(sizeof(*sb), BNXT_QPLIB_CMDQE_UNITS);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1457
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1459
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1461
sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1475
qp->state = sb->en_sqd_async_notify_state &
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1477
qp->en_sqd_async_notify = sb->en_sqd_async_notify_state &
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1479
qp->access = sb->access;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1480
qp->pkey_index = le16_to_cpu(sb->pkey);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1481
qp->qkey = le32_to_cpu(sb->qkey);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1482
qp->udp_sport = le16_to_cpu(sb->udp_src_port);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1484
temp32[0] = le32_to_cpu(sb->dgid[0]);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1485
temp32[1] = le32_to_cpu(sb->dgid[1]);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1486
temp32[2] = le32_to_cpu(sb->dgid[2]);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1487
temp32[3] = le32_to_cpu(sb->dgid[3]);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1490
qp->ah.flow_label = le32_to_cpu(sb->flow_label);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1494
if (res->sgid_tbl.hw_id[i] == le16_to_cpu(sb->sgid_index)) {
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1502
qp->ah.hop_limit = sb->hop_limit;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1503
qp->ah.traffic_class = sb->traffic_class;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1504
memcpy(qp->ah.dmac, sb->dest_mac, 6);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1505
qp->ah.vlan_id = (le16_to_cpu(sb->path_mtu_dest_vlan_id) &
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1508
qp->path_mtu = (le16_to_cpu(sb->path_mtu_dest_vlan_id) &
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1511
qp->timeout = sb->timeout;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1512
qp->retry_cnt = sb->retry_cnt;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1513
qp->rnr_retry = sb->rnr_retry;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1514
qp->min_rnr_timer = sb->min_rnr_timer;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1515
qp->rq.psn = le32_to_cpu(sb->rq_psn);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1516
qp->max_rd_atomic = ORRQ_SLOTS_TO_ORD_LIMIT(sb->max_rd_atomic);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1517
qp->sq.psn = le32_to_cpu(sb->sq_psn);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1519
IRRQ_SLOTS_TO_IRD_LIMIT(sb->max_dest_rd_atomic);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1522
qp->sq.max_sge = le16_to_cpu(sb->sq_sge);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1523
qp->rq.max_sge = le16_to_cpu(sb->rq_sge);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1524
qp->max_inline_data = le32_to_cpu(sb->max_inline_data);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1525
qp->dest_qpn = le32_to_cpu(sb->dest_qp_id);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1526
memcpy(qp->smac, sb->src_mac, 6);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1527
qp->vlan_id = le16_to_cpu(sb->vlan_pcp_vlan_dei_vlan_id);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1528
qp->port_id = le16_to_cpu(sb->port_id);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
1531
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
724
struct creq_query_srq_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
733
sbuf.size = ALIGN(sizeof(*sb), BNXT_QPLIB_CMDQE_UNITS);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
734
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_fp.c
736
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_fp.c
740
sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_fp.c
745
srq->threshold = le16_to_cpu(sb->srq_limit);
drivers/infiniband/hw/bnxt_re/qplib_fp.c
747
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
330
if (__get_cmdq_base_resp_size(msg->req, msg->req_sz) && msg->sb) {
drivers/infiniband/hw/bnxt_re/qplib_rcfw.c
331
struct bnxt_qplib_rcfw_sbuf *sbuf = msg->sb;
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
166
void *sb;
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
245
void *sb;
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
252
void *req, void *resp, void *sb,
drivers/infiniband/hw/bnxt_re/qplib_rcfw.h
257
msg->sb = sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1003
struct creq_query_roce_cc_gen1_resp_sb_tlv *sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1005
cc_ext->inact_th_hi = le16_to_cpu(sb->inactivity_th_hi);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1006
cc_ext->min_delta_cnp = le16_to_cpu(sb->min_time_between_cnps);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1007
cc_ext->init_cp = le16_to_cpu(sb->init_cp);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1008
cc_ext->tr_update_mode = sb->tr_update_mode;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1009
cc_ext->tr_update_cyls = sb->tr_update_cycles;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1010
cc_ext->fr_rtt = sb->fr_num_rtts;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1011
cc_ext->ai_rate_incr = sb->ai_rate_increase;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1012
cc_ext->rr_rtt_th = le16_to_cpu(sb->reduction_relax_rtts_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1013
cc_ext->ar_cr_th = le16_to_cpu(sb->additional_relax_cr_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1014
cc_ext->cr_min_th = le16_to_cpu(sb->cr_min_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1015
cc_ext->bw_avg_weight = sb->bw_avg_weight;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1016
cc_ext->cr_factor = sb->actual_cr_factor;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1017
cc_ext->cr_th_max_cp = le16_to_cpu(sb->max_cp_cr_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1018
cc_ext->cp_bias_en = sb->cp_bias_en;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1019
cc_ext->cp_bias = sb->cp_bias;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1020
cc_ext->cnp_ecn = sb->cnp_ecn;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1021
cc_ext->rtt_jitter_en = sb->rtt_jitter_en;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1022
cc_ext->bytes_per_usec = le16_to_cpu(sb->link_bytes_per_usec);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1023
cc_ext->cc_cr_reset_th = le16_to_cpu(sb->reset_cc_cr_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1024
cc_ext->cr_width = sb->cr_width;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1025
cc_ext->min_quota = sb->quota_period_min;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1026
cc_ext->max_quota = sb->quota_period_max;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1027
cc_ext->abs_max_quota = sb->quota_period_abs_max;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1028
cc_ext->tr_lb = le16_to_cpu(sb->tr_lower_bound);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1029
cc_ext->cr_prob_fac = sb->cr_prob_factor;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1030
cc_ext->tr_prob_fac = sb->tr_prob_factor;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1031
cc_ext->fair_cr_th = le16_to_cpu(sb->fairness_cr_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1032
cc_ext->red_div = sb->red_div;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1033
cc_ext->cnp_ratio_th = sb->cnp_ratio_th;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1034
cc_ext->ai_ext_rtt = le16_to_cpu(sb->exp_ai_rtts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1035
cc_ext->exp_crcp_ratio = sb->exp_ai_cr_cp_ratio;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1036
cc_ext->low_rate_en = sb->use_rate_table;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1037
cc_ext->cpcr_update_th = le16_to_cpu(sb->cp_exp_update_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1038
cc_ext->ai_rtt_th1 = le16_to_cpu(sb->high_exp_ai_rtts_th1);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1039
cc_ext->ai_rtt_th2 = le16_to_cpu(sb->high_exp_ai_rtts_th2);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1040
cc_ext->cf_rtt_th = le16_to_cpu(sb->actual_cr_cong_free_rtts_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1041
cc_ext->sc_cr_th1 = le16_to_cpu(sb->severe_cong_cr_th1);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1042
cc_ext->sc_cr_th2 = le16_to_cpu(sb->severe_cong_cr_th2);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1043
cc_ext->l64B_per_rtt = le32_to_cpu(sb->link64B_per_rtt);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1044
cc_ext->cc_ack_bytes = sb->cc_ack_bytes;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1045
cc_ext->reduce_cf_rtt_th = le16_to_cpu(sb->reduce_init_cong_free_rtts_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1054
struct creq_query_roce_cc_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1067
resp_size = sizeof(*sb);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1070
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1072
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1082
ext_sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1083
sb = bnxt_qplib_is_chip_gen_p5_p7(res->cctx) ? &ext_sb->base_sb :
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1086
cc_param->enable = sb->enable_cc & CREQ_QUERY_ROCE_CC_RESP_SB_ENABLE_CC;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1087
cc_param->tos_ecn = (sb->tos_dscp_tos_ecn &
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1090
cc_param->tos_dscp = (sb->tos_dscp_tos_ecn &
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1093
cc_param->alt_tos_dscp = sb->alt_tos_dscp;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1094
cc_param->alt_vlan_pcp = sb->alt_vlan_pcp;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1096
cc_param->g = sb->g;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1097
cc_param->nph_per_state = sb->num_phases_per_state;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1098
cc_param->init_cr = le16_to_cpu(sb->init_cr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1099
cc_param->init_tr = le16_to_cpu(sb->init_tr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
110
sbuf.size = ALIGN(sizeof(*sb), BNXT_QPLIB_CMDQE_UNITS);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1100
cc_param->cc_mode = sb->cc_mode;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1101
cc_param->inact_th = le16_to_cpu(sb->inactivity_th);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1102
cc_param->rtt = le16_to_cpu(sb->rtt);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1103
cc_param->tcp_cp = le16_to_cpu(sb->tcp_cp);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1104
cc_param->time_pph = sb->time_per_phase;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1105
cc_param->pkts_pph = sb->pkts_per_phase;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
111
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_sp.c
1111
dma_free_coherent(&rcfw->pdev->dev, sbuf.size, sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
113
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
115
sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
124
attr->max_qp = le32_to_cpu(sb->max_qp);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
128
sb->max_qp_rd_atom > BNXT_QPLIB_MAX_OUT_RD_ATOM ?
drivers/infiniband/hw/bnxt_re/qplib_sp.c
129
BNXT_QPLIB_MAX_OUT_RD_ATOM : sb->max_qp_rd_atom;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
131
sb->max_qp_init_rd_atom > BNXT_QPLIB_MAX_OUT_RD_ATOM ?
drivers/infiniband/hw/bnxt_re/qplib_sp.c
132
BNXT_QPLIB_MAX_OUT_RD_ATOM : sb->max_qp_init_rd_atom;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
133
attr->max_qp_wqes = le16_to_cpu(sb->max_qp_wr) - 1;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
147
min_t(u32, sb->max_sge_var_wqe, BNXT_VAR_MAX_SGE) : 6;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
148
attr->max_cq = le32_to_cpu(sb->max_cq);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
149
attr->max_cq_wqes = le32_to_cpu(sb->max_cqe);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
153
attr->max_mr = le32_to_cpu(sb->max_mr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
154
attr->max_mw = le32_to_cpu(sb->max_mw);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
156
attr->max_mr_size = le64_to_cpu(sb->max_mr_size);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
158
attr->max_raw_ethy_qp = le32_to_cpu(sb->max_raw_eth_qp);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
159
attr->max_ah = le32_to_cpu(sb->max_ah);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
161
attr->max_srq = le16_to_cpu(sb->max_srq);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
162
attr->max_srq_wqes = le32_to_cpu(sb->max_srq_wr) - 1;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
163
attr->max_srq_sges = sb->max_srq_sge;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
167
attr->l2_db_size = (sb->l2_db_space_size + 1) *
drivers/infiniband/hw/bnxt_re/qplib_sp.c
175
attr->max_sgid = le32_to_cpu(sb->max_gid);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
177
attr->dev_cap_flags = le16_to_cpu(sb->dev_cap_flags);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
178
attr->dev_cap_flags2 = le16_to_cpu(sb->dev_cap_ext_flags_2);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
181
attr->max_srq += le16_to_cpu(sb->max_srq_ext);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
184
temp = le32_to_cpu(sb->tqm_alloc_reqs[i]);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
193
attr->max_dpi = le32_to_cpu(sb->max_dpi);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
198
attr->rate_limit_min = le16_to_cpu(sb->rate_limit_min);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
199
attr->rate_limit_max = le32_to_cpu(sb->rate_limit_max);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
203
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
715
struct creq_query_roce_stats_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
725
sbuf.size = ALIGN(sizeof(*sb), BNXT_QPLIB_CMDQE_UNITS);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
726
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_sp.c
728
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
730
sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
739
stats->to_retransmits = le64_to_cpu(sb->to_retransmits);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
740
stats->seq_err_naks_rcvd = le64_to_cpu(sb->seq_err_naks_rcvd);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
741
stats->max_retry_exceeded = le64_to_cpu(sb->max_retry_exceeded);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
742
stats->rnr_naks_rcvd = le64_to_cpu(sb->rnr_naks_rcvd);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
743
stats->missing_resp = le64_to_cpu(sb->missing_resp);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
744
stats->unrecoverable_err = le64_to_cpu(sb->unrecoverable_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
745
stats->bad_resp_err = le64_to_cpu(sb->bad_resp_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
746
stats->local_qp_op_err = le64_to_cpu(sb->local_qp_op_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
747
stats->local_protection_err = le64_to_cpu(sb->local_protection_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
748
stats->mem_mgmt_op_err = le64_to_cpu(sb->mem_mgmt_op_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
749
stats->remote_invalid_req_err = le64_to_cpu(sb->remote_invalid_req_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
750
stats->remote_access_err = le64_to_cpu(sb->remote_access_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
751
stats->remote_op_err = le64_to_cpu(sb->remote_op_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
752
stats->dup_req = le64_to_cpu(sb->dup_req);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
753
stats->res_exceed_max = le64_to_cpu(sb->res_exceed_max);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
754
stats->res_length_mismatch = le64_to_cpu(sb->res_length_mismatch);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
755
stats->res_exceeds_wqe = le64_to_cpu(sb->res_exceeds_wqe);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
756
stats->res_opcode_err = le64_to_cpu(sb->res_opcode_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
757
stats->res_rx_invalid_rkey = le64_to_cpu(sb->res_rx_invalid_rkey);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
758
stats->res_rx_domain_err = le64_to_cpu(sb->res_rx_domain_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
759
stats->res_rx_no_perm = le64_to_cpu(sb->res_rx_no_perm);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
760
stats->res_rx_range_err = le64_to_cpu(sb->res_rx_range_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
761
stats->res_tx_invalid_rkey = le64_to_cpu(sb->res_tx_invalid_rkey);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
762
stats->res_tx_domain_err = le64_to_cpu(sb->res_tx_domain_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
763
stats->res_tx_no_perm = le64_to_cpu(sb->res_tx_no_perm);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
764
stats->res_tx_range_err = le64_to_cpu(sb->res_tx_range_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
765
stats->res_irrq_oflow = le64_to_cpu(sb->res_irrq_oflow);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
766
stats->res_unsup_opcode = le64_to_cpu(sb->res_unsup_opcode);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
767
stats->res_unaligned_atomic = le64_to_cpu(sb->res_unaligned_atomic);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
768
stats->res_rem_inv_err = le64_to_cpu(sb->res_rem_inv_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
769
stats->res_mem_error = le64_to_cpu(sb->res_mem_error);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
770
stats->res_srq_err = le64_to_cpu(sb->res_srq_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
771
stats->res_cmp_err = le64_to_cpu(sb->res_cmp_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
772
stats->res_invalid_dup_rkey = le64_to_cpu(sb->res_invalid_dup_rkey);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
773
stats->res_wqe_format_err = le64_to_cpu(sb->res_wqe_format_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
774
stats->res_cq_load_err = le64_to_cpu(sb->res_cq_load_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
775
stats->res_srq_load_err = le64_to_cpu(sb->res_srq_load_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
776
stats->res_tx_pci_err = le64_to_cpu(sb->res_tx_pci_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
777
stats->res_rx_pci_err = le64_to_cpu(sb->res_rx_pci_err);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
779
rcfw->oos_prev = le64_to_cpu(sb->res_oos_drop_count);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
783
(le64_to_cpu(sb->res_oos_drop_count) -
drivers/infiniband/hw/bnxt_re/qplib_sp.c
785
rcfw->oos_prev = le64_to_cpu(sb->res_oos_drop_count);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
790
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
798
struct creq_query_roce_stats_ext_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
804
sbuf.size = ALIGN(sizeof(*sb), BNXT_QPLIB_CMDQE_UNITS);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
805
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_sp.c
807
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
810
sb = sbuf.sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
831
estat->tx_atomic_req = le64_to_cpu(sb->tx_atomic_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
832
estat->tx_read_req = le64_to_cpu(sb->tx_read_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
833
estat->tx_read_res = le64_to_cpu(sb->tx_read_res_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
834
estat->tx_write_req = le64_to_cpu(sb->tx_write_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
835
estat->tx_send_req = le64_to_cpu(sb->tx_send_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
836
estat->tx_roce_pkts = le64_to_cpu(sb->tx_roce_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
837
estat->tx_roce_bytes = le64_to_cpu(sb->tx_roce_bytes);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
838
estat->rx_atomic_req = le64_to_cpu(sb->rx_atomic_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
839
estat->rx_read_req = le64_to_cpu(sb->rx_read_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
840
estat->rx_read_res = le64_to_cpu(sb->rx_read_res_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
841
estat->rx_write_req = le64_to_cpu(sb->rx_write_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
842
estat->rx_send_req = le64_to_cpu(sb->rx_send_req_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
843
estat->rx_roce_pkts = le64_to_cpu(sb->rx_roce_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
844
estat->rx_roce_bytes = le64_to_cpu(sb->rx_roce_bytes);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
845
estat->rx_roce_good_pkts = le64_to_cpu(sb->rx_roce_good_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
846
estat->rx_roce_good_bytes = le64_to_cpu(sb->rx_roce_good_bytes);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
847
estat->rx_out_of_buffer = le64_to_cpu(sb->rx_out_of_buffer_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
848
estat->rx_out_of_sequence = le64_to_cpu(sb->rx_out_of_sequence_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
849
estat->tx_cnp = le64_to_cpu(sb->tx_cnp_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
850
estat->rx_cnp = le64_to_cpu(sb->rx_cnp_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
851
estat->rx_ecn_marked = le64_to_cpu(sb->rx_ecn_marked_pkts);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
855
sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
97
struct creq_query_func_resp_sb *sb;
drivers/infiniband/hw/bnxt_re/qplib_sp.c
977
sbuf.sb = dma_alloc_coherent(&rcfw->pdev->dev, sbuf.size,
drivers/infiniband/hw/bnxt_re/qplib_sp.c
979
if (!sbuf.sb)
drivers/infiniband/hw/bnxt_re/qplib_sp.c
996
memcpy(resp_va, sbuf.sb, resp_size);
drivers/infiniband/hw/bnxt_re/qplib_sp.c
998
dma_free_coherent(&rcfw->pdev->dev, sbuf.size, sbuf.sb, sbuf.dma_addr);
drivers/infiniband/hw/qedr/main.c
385
cnq->sb = &dev->sb_array[i];
drivers/infiniband/hw/qedr/main.c
439
qed_sb_ack(cnq->sb, IGU_INT_DISABLE, 0);
drivers/infiniband/hw/qedr/main.c
441
qed_sb_update_sb_idx(cnq->sb);
drivers/infiniband/hw/qedr/main.c
492
qed_sb_ack(cnq->sb, IGU_INT_ENABLE, 1);
drivers/infiniband/hw/qedr/qedr.h
72
struct qed_sb_info *sb;
drivers/md/bcache/alloc.c
216
if (ca->fifo_last_bucket < ca->sb.first_bucket ||
drivers/md/bcache/alloc.c
217
ca->fifo_last_bucket >= ca->sb.nbuckets)
drivers/md/bcache/alloc.c
218
ca->fifo_last_bucket = ca->sb.first_bucket;
drivers/md/bcache/alloc.c
225
if (++checked >= ca->sb.nbuckets) {
drivers/md/bcache/alloc.c
243
n %= (size_t) (ca->sb.nbuckets - ca->sb.first_bucket);
drivers/md/bcache/alloc.c
244
n += ca->sb.first_bucket;
drivers/md/bcache/alloc.c
251
if (++checked >= ca->sb.nbuckets / 2) {
drivers/md/bcache/alloc.c
263
switch (CACHE_REPLACEMENT(&ca->sb)) {
drivers/md/bcache/alloc.c
348
if (CACHE_SYNC(&ca->sb)) {
drivers/md/bcache/alloc.c
437
SET_GC_SECTORS_USED(b, ca->sb.bucket_size);
drivers/md/bcache/alloc.c
497
ca->sb.nr_this_dev);
drivers/md/bcache/alloc.c
569
ret->sectors_free = c->cache->sb.bucket_size;
drivers/md/bcache/alloc.c
663
if (b->sectors_free < c->cache->sb.block_size)
drivers/md/bcache/alloc.c
87
unsigned long next = c->nbuckets * c->cache->sb.bucket_size / 1024;
drivers/md/bcache/bcache.h
1043
struct cache_set *bch_cache_set_alloc(struct cache_sb *sb);
drivers/md/bcache/bcache.h
308
struct cache_sb sb;
drivers/md/bcache/bcache.h
422
struct cache_sb sb;
drivers/md/bcache/bcache.h
775
#define bucket_bytes(ca) ((ca)->sb.bucket_size << 9)
drivers/md/bcache/bcache.h
776
#define block_bytes(ca) ((ca)->sb.block_size << 9)
drivers/md/bcache/bcache.h
778
static inline unsigned int meta_bucket_pages(struct cache_sb *sb)
drivers/md/bcache/bcache.h
786
n = sb->bucket_size / PAGE_SECTORS;
drivers/md/bcache/bcache.h
793
static inline unsigned int meta_bucket_bytes(struct cache_sb *sb)
drivers/md/bcache/bcache.h
795
return meta_bucket_pages(sb) << PAGE_SHIFT;
drivers/md/bcache/bcache.h
799
((meta_bucket_bytes(&(ca)->sb) - sizeof(struct prio_set)) / \
drivers/md/bcache/bcache.h
803
DIV_ROUND_UP((size_t) (ca)->sb.nbuckets, prios_per_bucket(ca))
drivers/md/bcache/bcache.h
817
return s & (c->cache->sb.bucket_size - 1);
drivers/md/bcache/bcache.h
899
for (b = (ca)->buckets + (ca)->sb.first_bucket; \
drivers/md/bcache/bcache.h
900
b < (ca)->buckets + (ca)->sb.nbuckets; b++)
drivers/md/bcache/bcache_ondisk.h
278
static inline _Bool SB_IS_BDEV(const struct cache_sb *sb)
drivers/md/bcache/bcache_ondisk.h
280
return sb->version == BCACHE_SB_VERSION_BDEV
drivers/md/bcache/bcache_ondisk.h
281
|| sb->version == BCACHE_SB_VERSION_BDEV_WITH_OFFSET
drivers/md/bcache/bcache_ondisk.h
282
|| sb->version == BCACHE_SB_VERSION_BDEV_WITH_FEATURES;
drivers/md/bcache/bcache_ondisk.h
314
static inline __u64 jset_magic(struct cache_sb *sb)
drivers/md/bcache/bcache_ondisk.h
316
return sb->set_magic ^ JSET_MAGIC;
drivers/md/bcache/bcache_ondisk.h
319
static inline __u64 pset_magic(struct cache_sb *sb)
drivers/md/bcache/bcache_ondisk.h
321
return sb->set_magic ^ PSET_MAGIC;
drivers/md/bcache/bcache_ondisk.h
324
static inline __u64 bset_magic(struct cache_sb *sb)
drivers/md/bcache/bcache_ondisk.h
326
return sb->set_magic ^ BSET_MAGIC;
drivers/md/bcache/btree.c
1147
bch_bset_init_next(&b->keys, b->keys.set->data, bset_magic(&b->c->cache->sb));
drivers/md/bcache/btree.c
122
bset_magic(&b->c->cache->sb));
drivers/md/bcache/btree.c
160
iter->size = b->c->cache->sb.bucket_size / b->c->cache->sb.block_size;
drivers/md/bcache/btree.c
1801
for (k = ca->sb.d; k < ca->sb.d + ca->sb.keys; k++)
drivers/md/bcache/btree.c
183
if (i->magic != bset_magic(&b->c->cache->sb))
drivers/md/bcache/btree.c
224
bset_magic(&b->c->cache->sb));
drivers/md/bcache/btree.c
428
atomic_long_add(set_blocks(i, block_bytes(b->c->cache)) * b->c->cache->sb.block_size,
drivers/md/bcache/btree.c
762
free_pages((unsigned long) c->verify_ondisk, ilog2(meta_bucket_pages(&c->cache->sb)));
drivers/md/bcache/btree.c
810
ilog2(meta_bucket_pages(&c->cache->sb)));
drivers/md/bcache/btree.h
197
atomic_set(&c->sectors_to_gc, c->cache->sb.bucket_size * c->nbuckets / 16);
drivers/md/bcache/extents.c
139
if (n >= b->c->cache->sb.first_bucket && n < b->c->cache->sb.nbuckets)
drivers/md/bcache/extents.c
57
if (KEY_SIZE(k) + r > c->cache->sb.bucket_size ||
drivers/md/bcache/extents.c
58
bucket < ca->sb.first_bucket ||
drivers/md/bcache/extents.c
59
bucket >= ca->sb.nbuckets)
drivers/md/bcache/extents.c
78
if (KEY_SIZE(k) + r > c->cache->sb.bucket_size)
drivers/md/bcache/extents.c
80
if (bucket < ca->sb.first_bucket)
drivers/md/bcache/extents.c
82
if (bucket >= ca->sb.nbuckets)
drivers/md/bcache/features.c
33
if (BCH_HAS_ ## type ## _FEATURE(&c->cache->sb, f->mask)) { \
drivers/md/bcache/features.c
47
if (BCH_HAS_ ## type ## _FEATURE(&c->cache->sb, f->mask)) \
drivers/md/bcache/features.h
101
return ((sb->feature_ro_compat & ~BCH_FEATURE_RO_COMPAT_SUPP) != 0);
drivers/md/bcache/features.h
104
static inline bool bch_has_unknown_incompat_features(struct cache_sb *sb)
drivers/md/bcache/features.h
106
return ((sb->feature_incompat & ~BCH_FEATURE_INCOMPAT_SUPP) != 0);
drivers/md/bcache/features.h
27
#define BCH_HAS_COMPAT_FEATURE(sb, mask) \
drivers/md/bcache/features.h
28
((sb)->feature_compat & (mask))
drivers/md/bcache/features.h
29
#define BCH_HAS_RO_COMPAT_FEATURE(sb, mask) \
drivers/md/bcache/features.h
30
((sb)->feature_ro_compat & (mask))
drivers/md/bcache/features.h
31
#define BCH_HAS_INCOMPAT_FEATURE(sb, mask) \
drivers/md/bcache/features.h
32
((sb)->feature_incompat & (mask))
drivers/md/bcache/features.h
35
static inline int bch_has_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
37
if (sb->version < BCACHE_SB_VERSION_CDEV_WITH_FEATURES) \
drivers/md/bcache/features.h
39
return (((sb)->feature_compat & \
drivers/md/bcache/features.h
42
static inline void bch_set_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
44
(sb)->feature_compat |= \
drivers/md/bcache/features.h
47
static inline void bch_clear_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
49
(sb)->feature_compat &= \
drivers/md/bcache/features.h
54
static inline int bch_has_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
56
if (sb->version < BCACHE_SB_VERSION_CDEV_WITH_FEATURES) \
drivers/md/bcache/features.h
58
return (((sb)->feature_ro_compat & \
drivers/md/bcache/features.h
61
static inline void bch_set_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
63
(sb)->feature_ro_compat |= \
drivers/md/bcache/features.h
66
static inline void bch_clear_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
68
(sb)->feature_ro_compat &= \
drivers/md/bcache/features.h
73
static inline int bch_has_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
75
if (sb->version < BCACHE_SB_VERSION_CDEV_WITH_FEATURES) \
drivers/md/bcache/features.h
77
return (((sb)->feature_incompat & \
drivers/md/bcache/features.h
80
static inline void bch_set_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
82
(sb)->feature_incompat |= \
drivers/md/bcache/features.h
85
static inline void bch_clear_feature_##name(struct cache_sb *sb) \
drivers/md/bcache/features.h
87
(sb)->feature_incompat &= \
drivers/md/bcache/features.h
94
static inline bool bch_has_unknown_compat_features(struct cache_sb *sb)
drivers/md/bcache/features.h
96
return ((sb->feature_compat & ~BCH_FEATURE_COMPAT_SUPP) != 0);
drivers/md/bcache/features.h
99
static inline bool bch_has_unknown_ro_compat_features(struct cache_sb *sb)
drivers/md/bcache/io.c
29
bio_init_inline(bio, NULL, meta_bucket_pages(&c->cache->sb), 0);
drivers/md/bcache/journal.c
161
offset += blocks * ca->sb.block_size;
drivers/md/bcache/journal.c
162
len -= blocks * ca->sb.block_size;
drivers/md/bcache/journal.c
189
pr_debug("%u journal buckets\n", ca->sb.njournal_buckets);
drivers/md/bcache/journal.c
195
for (i = 0; i < ca->sb.njournal_buckets; i++) {
drivers/md/bcache/journal.c
201
l = (i * 2654435769U) % ca->sb.njournal_buckets;
drivers/md/bcache/journal.c
216
for_each_clear_bit(l, bitmap, ca->sb.njournal_buckets)
drivers/md/bcache/journal.c
221
if (l == ca->sb.njournal_buckets)
drivers/md/bcache/journal.c
228
r = find_next_bit(bitmap, ca->sb.njournal_buckets, l + 1);
drivers/md/bcache/journal.c
250
m, ca->sb.njournal_buckets);
drivers/md/bcache/journal.c
255
l = ca->sb.njournal_buckets - 1;
drivers/md/bcache/journal.c
269
for (i = 0; i < ca->sb.njournal_buckets; i++)
drivers/md/bcache/journal.c
278
ja->last_idx = (i + 1) % ca->sb.njournal_buckets;
drivers/md/bcache/journal.c
46
sector_t bucket = bucket_to_sector(ca->set, ca->sb.d[bucket_index]);
drivers/md/bcache/journal.c
52
while (offset < ca->sb.bucket_size) {
drivers/md/bcache/journal.c
53
reread: left = ca->sb.bucket_size - offset;
drivers/md/bcache/journal.c
564
n = ca->sb.njournal_buckets + ja->last_idx - ja->cur_idx;
drivers/md/bcache/journal.c
594
ca->sb.njournal_buckets;
drivers/md/bcache/journal.c
602
ja->cur_idx = (ja->cur_idx + 1) % ca->sb.njournal_buckets;
drivers/md/bcache/journal.c
604
bucket_to_sector(c, ca->sb.d[ja->cur_idx]),
drivers/md/bcache/journal.c
605
ca->sb.nr_this_dev);
drivers/md/bcache/journal.c
610
c->journal.blocks_free = ca->sb.bucket_size >> c->block_bits;
drivers/md/bcache/journal.c
679
ca->sb.block_size;
drivers/md/bcache/journal.c
705
w->data->prio_bucket[ca->sb.nr_this_dev] = ca->prio_buckets[0];
drivers/md/bcache/journal.c
706
w->data->magic = jset_magic(&ca->sb);
drivers/md/bcache/journal.c
78
if (j->magic != jset_magic(&ca->sb)) {
drivers/md/bcache/journal.c
790
block_bytes(ca)) * ca->sb.block_size;
drivers/md/bcache/journal.c
793
c->journal.blocks_free * ca->sb.block_size,
drivers/md/bcache/journal.c
858
if (!CACHE_SYNC(&c->cache->sb))
drivers/md/bcache/movinggc.c
209
reserve_sectors = ca->sb.bucket_size *
drivers/md/bcache/movinggc.c
217
GC_SECTORS_USED(b) == ca->sb.bucket_size ||
drivers/md/bcache/request.c
1203
bio->bi_iter.bi_sector += dc->sb.data_offset;
drivers/md/bcache/request.c
32
return BDEV_CACHE_MODE(&dc->sb);
drivers/md/bcache/request.c
383
if (BDEV_STATE(&dc->sb) == BDEV_STATE_CLEAN &&
drivers/md/bcache/request.c
411
if (bio->bi_iter.bi_sector & (c->cache->sb.block_size - 1) ||
drivers/md/bcache/request.c
412
bio_sectors(bio) & (c->cache->sb.block_size - 1)) {
drivers/md/bcache/super.c
101
sb->nr_in_set = le16_to_cpu(s->nr_in_set);
drivers/md/bcache/super.c
102
sb->nr_this_dev = le16_to_cpu(s->nr_this_dev);
drivers/md/bcache/super.c
105
if (sb->keys > SB_JOURNAL_BUCKETS)
drivers/md/bcache/super.c
1054
char *buf = kmemdup_nul(dc->sb.label, SB_LABEL_SIZE, GFP_KERNEL);
drivers/md/bcache/super.c
1057
kasprintf(GFP_KERNEL, "CACHED_UUID=%pU", dc->sb.uuid),
drivers/md/bcache/super.c
1075
BDEV_STATE(&dc->sb) != BDEV_STATE_NONE) {
drivers/md/bcache/super.c
1080
SET_BDEV_STATE(&dc->sb, BDEV_STATE_STALE);
drivers/md/bcache/super.c
109
if (sb->nbuckets > LONG_MAX)
drivers/md/bcache/super.c
113
if (sb->nbuckets < 1 << 7)
drivers/md/bcache/super.c
117
if (!is_power_of_2(sb->block_size))
drivers/md/bcache/super.c
1205
(!set_uuid && memcmp(dc->sb.set_uuid, c->set_uuid, 16)))
drivers/md/bcache/super.c
121
if (sb->block_size > PAGE_SECTORS)
drivers/md/bcache/super.c
1218
if (dc->sb.block_size < c->cache->sb.block_size) {
drivers/md/bcache/super.c
1227
if (!memcmp(dc->sb.uuid, exist_dc->sb.uuid, 16)) {
drivers/md/bcache/super.c
1235
u = uuid_find(c, dc->sb.uuid);
drivers/md/bcache/super.c
1238
(BDEV_STATE(&dc->sb) == BDEV_STATE_STALE ||
drivers/md/bcache/super.c
1239
BDEV_STATE(&dc->sb) == BDEV_STATE_NONE)) {
drivers/md/bcache/super.c
1246
if (BDEV_STATE(&dc->sb) == BDEV_STATE_DIRTY) {
drivers/md/bcache/super.c
125
if (!is_power_of_2(sb->bucket_size))
drivers/md/bcache/super.c
1268
memcpy(u->uuid, dc->sb.uuid, 16);
drivers/md/bcache/super.c
1269
memcpy(u->label, dc->sb.label, SB_LABEL_SIZE);
drivers/md/bcache/super.c
1273
memcpy(dc->sb.set_uuid, c->set_uuid, 16);
drivers/md/bcache/super.c
1274
SET_BDEV_STATE(&dc->sb, BDEV_STATE_CLEAN);
drivers/md/bcache/super.c
129
if (sb->bucket_size < PAGE_SECTORS)
drivers/md/bcache/super.c
1303
if (BDEV_STATE(&dc->sb) == BDEV_STATE_DIRTY) {
drivers/md/bcache/super.c
1328
if (bch_has_feature_obso_large_bucket(&(c->cache->sb))) {
drivers/md/bcache/super.c
134
sb->bucket_size * sb->nbuckets)
drivers/md/bcache/super.c
138
if (bch_is_zero(sb->set_uuid, 16))
drivers/md/bcache/super.c
142
if (!sb->nr_in_set ||
drivers/md/bcache/super.c
143
sb->nr_in_set <= sb->nr_this_dev ||
drivers/md/bcache/super.c
1431
bdev_nr_sectors(dc->bdev) - dc->sb.data_offset,
drivers/md/bcache/super.c
144
sb->nr_in_set > MAX_CACHES_PER_SET)
drivers/md/bcache/super.c
1449
static int register_bdev(struct cache_sb *sb, struct cache_sb_disk *sb_disk,
drivers/md/bcache/super.c
1457
memcpy(&dc->sb, sb, sizeof(struct cache_sb));
drivers/md/bcache/super.c
1462
if (cached_dev_init(dc, sb->block_size << 9))
drivers/md/bcache/super.c
1478
if (BDEV_STATE(&dc->sb) == BDEV_STATE_NONE ||
drivers/md/bcache/super.c
1479
BDEV_STATE(&dc->sb) == BDEV_STATE_STALE) {
drivers/md/bcache/super.c
148
for (i = 0; i < sb->keys; i++)
drivers/md/bcache/super.c
149
if (sb->d[i] != sb->first_bucket + i)
drivers/md/bcache/super.c
153
if (sb->first_bucket + sb->keys > sb->nbuckets)
drivers/md/bcache/super.c
1556
if (bch_has_feature_obso_large_bucket(&c->cache->sb)) {
drivers/md/bcache/super.c
157
if (sb->first_bucket * sb->bucket_size < 16)
drivers/md/bcache/super.c
166
static const char *read_super(struct cache_sb *sb, struct block_device *bdev,
drivers/md/bcache/super.c
1686
free_pages((unsigned long) c->uuids, ilog2(meta_bucket_pages(&c->cache->sb)));
drivers/md/bcache/super.c
180
sb->offset = le64_to_cpu(s->offset);
drivers/md/bcache/super.c
181
sb->version = le64_to_cpu(s->version);
drivers/md/bcache/super.c
183
memcpy(sb->magic, s->magic, 16);
drivers/md/bcache/super.c
184
memcpy(sb->uuid, s->uuid, 16);
drivers/md/bcache/super.c
185
memcpy(sb->set_uuid, s->set_uuid, 16);
drivers/md/bcache/super.c
1859
#define alloc_meta_bucket_pages(gfp, sb) \
drivers/md/bcache/super.c
186
memcpy(sb->label, s->label, SB_LABEL_SIZE);
drivers/md/bcache/super.c
1860
((void *) __get_free_pages(__GFP_ZERO|__GFP_COMP|gfp, ilog2(meta_bucket_pages(sb))))
drivers/md/bcache/super.c
1862
struct cache_set *bch_cache_set_alloc(struct cache_sb *sb)
drivers/md/bcache/super.c
1865
struct cache *ca = container_of(sb, struct cache, sb);
drivers/md/bcache/super.c
188
sb->flags = le64_to_cpu(s->flags);
drivers/md/bcache/super.c
1887
memcpy(c->set_uuid, sb->set_uuid, 16);
drivers/md/bcache/super.c
189
sb->seq = le64_to_cpu(s->seq);
drivers/md/bcache/super.c
1891
c->bucket_bits = ilog2(sb->bucket_size);
drivers/md/bcache/super.c
1892
c->block_bits = ilog2(sb->block_size);
drivers/md/bcache/super.c
1893
c->nr_uuids = meta_bucket_bytes(sb) / sizeof(struct uuid_entry);
drivers/md/bcache/super.c
1896
c->btree_pages = meta_bucket_pages(sb);
drivers/md/bcache/super.c
190
sb->last_mount = le32_to_cpu(s->last_mount);
drivers/md/bcache/super.c
191
sb->keys = le16_to_cpu(s->keys);
drivers/md/bcache/super.c
1923
((meta_bucket_pages(sb) * PAGE_SECTORS) / sb->block_size) *
drivers/md/bcache/super.c
1935
sizeof(struct bio_vec) * meta_bucket_pages(sb)))
drivers/md/bcache/super.c
194
sb->d[i] = le64_to_cpu(s->d[i]);
drivers/md/bcache/super.c
1945
c->uuids = alloc_meta_bucket_pages(GFP_KERNEL, sb);
drivers/md/bcache/super.c
197
sb->version, sb->flags, sb->seq, sb->keys);
drivers/md/bcache/super.c
1989
c->nbuckets = ca->sb.nbuckets;
drivers/md/bcache/super.c
1992
if (CACHE_SYNC(&c->cache->sb)) {
drivers/md/bcache/super.c
200
if (sb->offset != SB_SECTOR)
drivers/md/bcache/super.c
2009
if (prio_read(ca, j->prio_bucket[ca->sb.nr_this_dev]))
drivers/md/bcache/super.c
204
if (memcmp(sb->magic, bcache_magic, 16))
drivers/md/bcache/super.c
2077
ca->sb.keys = clamp_t(int, ca->sb.nbuckets >> 7,
drivers/md/bcache/super.c
2080
for (j = 0; j < ca->sb.keys; j++)
drivers/md/bcache/super.c
2081
ca->sb.d[j] = ca->sb.first_bucket + j;
drivers/md/bcache/super.c
2115
SET_CACHE_SYNC(&c->cache->sb, true);
drivers/md/bcache/super.c
212
if (bch_is_zero(sb->uuid, 16))
drivers/md/bcache/super.c
2126
c->cache->sb.last_mount = (u32)ktime_get_real_seconds();
drivers/md/bcache/super.c
2129
if (bch_has_feature_obso_large_bucket(&c->cache->sb))
drivers/md/bcache/super.c
215
sb->block_size = le16_to_cpu(s->block_size);
drivers/md/bcache/super.c
2161
if (!memcmp(c->set_uuid, ca->sb.set_uuid, 16)) {
drivers/md/bcache/super.c
2168
c = bch_cache_set_alloc(&ca->sb);
drivers/md/bcache/super.c
218
if (sb->block_size << 9 < bdev_logical_block_size(bdev))
drivers/md/bcache/super.c
2184
sprintf(buf, "cache%i", ca->sb.nr_this_dev);
drivers/md/bcache/super.c
221
switch (sb->version) {
drivers/md/bcache/super.c
2216
free_pages((unsigned long) ca->disk_buckets, ilog2(meta_bucket_pages(&ca->sb)));
drivers/md/bcache/super.c
223
sb->data_offset = BDEV_DATA_START_DEFAULT;
drivers/md/bcache/super.c
227
sb->data_offset = le64_to_cpu(s->data_offset);
drivers/md/bcache/super.c
2289
btree_buckets = clamp_t(size_t, ca->sb.nbuckets >> 7,
drivers/md/bcache/super.c
2291
free = roundup_pow_of_two(ca->sb.nbuckets) >> 10;
drivers/md/bcache/super.c
230
if (sb->data_offset < BDEV_DATA_START_DEFAULT)
drivers/md/bcache/super.c
2331
ca->sb.nbuckets));
drivers/md/bcache/super.c
2345
ca->disk_buckets = alloc_meta_bucket_pages(GFP_KERNEL, &ca->sb);
drivers/md/bcache/super.c
236
err = read_super_common(sb, bdev, s);
drivers/md/bcache/super.c
2381
static int register_cache(struct cache_sb *sb, struct cache_sb_disk *sb_disk,
drivers/md/bcache/super.c
2388
memcpy(&ca->sb, sb, sizeof(struct cache_sb));
drivers/md/bcache/super.c
245
sb->feature_compat = le64_to_cpu(s->feature_compat);
drivers/md/bcache/super.c
246
sb->feature_incompat = le64_to_cpu(s->feature_incompat);
drivers/md/bcache/super.c
247
sb->feature_ro_compat = le64_to_cpu(s->feature_ro_compat);
drivers/md/bcache/super.c
2484
struct cache_sb *sb;
drivers/md/bcache/super.c
2497
if (register_bdev(args->sb, args->sb_disk, args->bdev_file,
drivers/md/bcache/super.c
2505
kfree(args->sb);
drivers/md/bcache/super.c
251
if (bch_has_unknown_compat_features(sb))
drivers/md/bcache/super.c
2518
if (register_cache(args->sb, args->sb_disk, args->bdev_file,
drivers/md/bcache/super.c
2525
kfree(args->sb);
drivers/md/bcache/super.c
2533
if (SB_IS_BDEV(args->sb))
drivers/md/bcache/super.c
2542
static void *alloc_holder_object(struct cache_sb *sb)
drivers/md/bcache/super.c
2544
if (SB_IS_BDEV(sb))
drivers/md/bcache/super.c
255
if (bch_has_unknown_ro_compat_features(sb))
drivers/md/bcache/super.c
2554
struct cache_sb *sb;
drivers/md/bcache/super.c
2583
sb = kmalloc_obj(struct cache_sb);
drivers/md/bcache/super.c
2584
if (!sb)
drivers/md/bcache/super.c
259
if (bch_has_unknown_incompat_features(sb))
drivers/md/bcache/super.c
2593
err = read_super(sb, file_bdev(bdev_file), &sb_disk);
drivers/md/bcache/super.c
2597
holder = alloc_holder_object(sb);
drivers/md/bcache/super.c
262
err = read_super_common(sb, bdev, s);
drivers/md/bcache/super.c
2644
args->sb = sb;
drivers/md/bcache/super.c
2653
if (SB_IS_BDEV(sb)) {
drivers/md/bcache/super.c
2655
ret = register_bdev(sb, sb_disk, bdev_file, holder);
drivers/md/bcache/super.c
2662
ret = register_cache(sb, sb_disk, bdev_file, holder);
drivers/md/bcache/super.c
2667
kfree(sb);
drivers/md/bcache/super.c
2681
kfree(sb);
drivers/md/bcache/super.c
271
sb->last_mount = (u32)ktime_get_real_seconds();
drivers/md/bcache/super.c
2720
char *pdev_set_uuid = pdev->dc->sb.set_uuid;
drivers/md/bcache/super.c
289
static void __write_super(struct cache_sb *sb, struct cache_sb_disk *out,
drivers/md/bcache/super.c
298
out->offset = cpu_to_le64(sb->offset);
drivers/md/bcache/super.c
300
memcpy(out->uuid, sb->uuid, 16);
drivers/md/bcache/super.c
301
memcpy(out->set_uuid, sb->set_uuid, 16);
drivers/md/bcache/super.c
302
memcpy(out->label, sb->label, SB_LABEL_SIZE);
drivers/md/bcache/super.c
304
out->flags = cpu_to_le64(sb->flags);
drivers/md/bcache/super.c
305
out->seq = cpu_to_le64(sb->seq);
drivers/md/bcache/super.c
307
out->last_mount = cpu_to_le32(sb->last_mount);
drivers/md/bcache/super.c
308
out->first_bucket = cpu_to_le16(sb->first_bucket);
drivers/md/bcache/super.c
309
out->keys = cpu_to_le16(sb->keys);
drivers/md/bcache/super.c
311
for (i = 0; i < sb->keys; i++)
drivers/md/bcache/super.c
312
out->d[i] = cpu_to_le64(sb->d[i]);
drivers/md/bcache/super.c
314
if (sb->version >= BCACHE_SB_VERSION_CDEV_WITH_FEATURES) {
drivers/md/bcache/super.c
315
out->feature_compat = cpu_to_le64(sb->feature_compat);
drivers/md/bcache/super.c
316
out->feature_incompat = cpu_to_le64(sb->feature_incompat);
drivers/md/bcache/super.c
317
out->feature_ro_compat = cpu_to_le64(sb->feature_ro_compat);
drivers/md/bcache/super.c
320
out->version = cpu_to_le64(sb->version);
drivers/md/bcache/super.c
324
sb->version, sb->flags, sb->seq);
drivers/md/bcache/super.c
350
__write_super(&dc->sb, dc->sb_disk, bio);
drivers/md/bcache/super.c
382
ca->sb.seq++;
drivers/md/bcache/super.c
384
if (ca->sb.version < version)
drivers/md/bcache/super.c
385
ca->sb.version = version;
drivers/md/bcache/super.c
392
__write_super(&ca->sb, ca->sb_disk, bio);
drivers/md/bcache/super.c
511
size = meta_bucket_pages(&ca->sb) * PAGE_SECTORS;
drivers/md/bcache/super.c
517
atomic_long_add(ca->sb.bucket_size, &ca->meta_sectors_written);
drivers/md/bcache/super.c
597
bio->bi_iter.bi_sector = bucket * ca->sb.bucket_size;
drivers/md/bcache/super.c
599
bio->bi_iter.bi_size = meta_bucket_bytes(&ca->sb);
drivers/md/bcache/super.c
639
atomic_long_add(ca->sb.bucket_size * prio_buckets(ca),
drivers/md/bcache/super.c
64
static unsigned int get_bucket_size(struct cache_sb *sb, struct cache_sb_disk *s)
drivers/md/bcache/super.c
649
b < ca->buckets + ca->sb.nbuckets && d < end;
drivers/md/bcache/super.c
656
p->magic = pset_magic(&ca->sb);
drivers/md/bcache/super.c
657
p->csum = bch_crc64(&p->magic, meta_bucket_bytes(&ca->sb) - 8);
drivers/md/bcache/super.c
68
if (sb->version >= BCACHE_SB_VERSION_CDEV_WITH_FEATURES) {
drivers/md/bcache/super.c
69
if (bch_has_feature_large_bucket(sb)) {
drivers/md/bcache/super.c
700
b < ca->buckets + ca->sb.nbuckets;
drivers/md/bcache/super.c
710
bch_crc64(&p->magic, meta_bucket_bytes(&ca->sb) - 8)) {
drivers/md/bcache/super.c
715
if (p->magic != pset_magic(&ca->sb)) {
drivers/md/bcache/super.c
82
} else if (bch_has_feature_obso_large_bucket(sb)) {
drivers/md/bcache/super.c
91
static const char *read_super_common(struct cache_sb *sb, struct block_device *bdev,
drivers/md/bcache/super.c
97
sb->first_bucket= le16_to_cpu(s->first_bucket);
drivers/md/bcache/super.c
98
sb->nbuckets = le64_to_cpu(s->nbuckets);
drivers/md/bcache/super.c
99
sb->bucket_size = get_bucket_size(sb, s);
drivers/md/bcache/sysfs.c
1037
sysfs_print(nbuckets, ca->sb.nbuckets);
drivers/md/bcache/sysfs.c
1051
CACHE_REPLACEMENT(&ca->sb));
drivers/md/bcache/sysfs.c
1055
size_t n = ca->sb.nbuckets, i;
drivers/md/bcache/sysfs.c
1063
ca->sb.nbuckets));
drivers/md/bcache/sysfs.c
1079
for (i = ca->sb.first_bucket; i < n; i++)
drivers/md/bcache/sysfs.c
1115
unused * 100 / (size_t) ca->sb.nbuckets,
drivers/md/bcache/sysfs.c
1116
available * 100 / (size_t) ca->sb.nbuckets,
drivers/md/bcache/sysfs.c
1117
dirty * 100 / (size_t) ca->sb.nbuckets,
drivers/md/bcache/sysfs.c
1118
meta * 100 / (size_t) ca->sb.nbuckets, sum,
drivers/md/bcache/sysfs.c
1119
n * ca->sb.bucket_size / (ARRAY_SIZE(q) + 1));
drivers/md/bcache/sysfs.c
1148
if ((unsigned int) v != CACHE_REPLACEMENT(&ca->sb)) {
drivers/md/bcache/sysfs.c
1150
SET_CACHE_REPLACEMENT(&ca->sb, v);
drivers/md/bcache/sysfs.c
182
BDEV_CACHE_MODE(&dc->sb));
drivers/md/bcache/sysfs.c
263
sysfs_print(state, states[BDEV_STATE(&dc->sb)]);
drivers/md/bcache/sysfs.c
266
memcpy(buf, dc->sb.label, SB_LABEL_SIZE);
drivers/md/bcache/sysfs.c
280
snprintf(buf, 36+1, "%pU", dc->sb.uuid);
drivers/md/bcache/sysfs.c
381
if ((unsigned int) v != BDEV_CACHE_MODE(&dc->sb)) {
drivers/md/bcache/sysfs.c
382
SET_BDEV_CACHE_MODE(&dc->sb, v);
drivers/md/bcache/sysfs.c
407
memcpy(dc->sb.label, buf, size);
drivers/md/bcache/sysfs.c
409
dc->sb.label[size] = '\0';
drivers/md/bcache/sysfs.c
410
if (size && dc->sb.label[size - 1] == '\n')
drivers/md/bcache/sysfs.c
411
dc->sb.label[size - 1] = '\0';
drivers/md/bcache/sysfs.c
422
add_uevent_var(env, "CACHED_UUID=%pU", dc->sb.uuid);
drivers/md/bcache/sysfs.c
728
sysfs_print(synchronous, CACHE_SYNC(&c->cache->sb));
drivers/md/bcache/sysfs.c
829
if (sync != CACHE_SYNC(&c->cache->sb)) {
drivers/md/bcache/sysfs.c
830
SET_CACHE_SYNC(&c->cache->sb, sync);
drivers/md/bcache/writeback.c
108
div_s64((dirty_buckets * c->cache->sb.bucket_size), dirty);
drivers/md/bcache/writeback.c
38
uint64_t cache_sectors = c->nbuckets * c->cache->sb.bucket_size -
drivers/md/bcache/writeback.c
778
SET_BDEV_STATE(&dc->sb, BDEV_STATE_CLEAN);
drivers/md/bcache/writeback.c
790
memset(&dc->sb.set_uuid, 0, 16);
drivers/md/bcache/writeback.c
791
SET_BDEV_STATE(&dc->sb, BDEV_STATE_NONE);
drivers/md/bcache/writeback.h
138
if (BDEV_STATE(&dc->sb) != BDEV_STATE_DIRTY) {
drivers/md/bcache/writeback.h
139
SET_BDEV_STATE(&dc->sb, BDEV_STATE_DIRTY);
drivers/md/dm-cache-target.c
1017
sector_t sb = bio->bi_iter.bi_sector;
drivers/md/dm-cache-target.c
1020
*b = to_dblock(dm_sector_div_up(sb, cache->discard_block_size));
drivers/md/dm-cache-target.c
1022
if (se - sb < cache->discard_block_size)
drivers/md/dm-clone-metadata.c
169
struct superblock_disk *sb;
drivers/md/dm-clone-metadata.c
172
sb = dm_block_data(b);
drivers/md/dm-clone-metadata.c
173
sb->blocknr = cpu_to_le64(dm_block_location(b));
drivers/md/dm-clone-metadata.c
175
csum = dm_bm_checksum(&sb->flags, sb_block_size - sizeof(__le32),
drivers/md/dm-clone-metadata.c
177
sb->csum = cpu_to_le32(csum);
drivers/md/dm-clone-metadata.c
183
struct superblock_disk *sb;
drivers/md/dm-clone-metadata.c
186
sb = dm_block_data(b);
drivers/md/dm-clone-metadata.c
188
if (dm_block_location(b) != le64_to_cpu(sb->blocknr)) {
drivers/md/dm-clone-metadata.c
190
le64_to_cpu(sb->blocknr),
drivers/md/dm-clone-metadata.c
195
if (le64_to_cpu(sb->magic) != SUPERBLOCK_MAGIC) {
drivers/md/dm-clone-metadata.c
197
le64_to_cpu(sb->magic),
drivers/md/dm-clone-metadata.c
202
csum = dm_bm_checksum(&sb->flags, sb_block_size - sizeof(__le32),
drivers/md/dm-clone-metadata.c
204
if (sb->csum != cpu_to_le32(csum)) {
drivers/md/dm-clone-metadata.c
206
csum, le32_to_cpu(sb->csum));
drivers/md/dm-clone-metadata.c
211
metadata_version = le32_to_cpu(sb->version);
drivers/md/dm-clone-metadata.c
299
struct superblock_disk *sb)
drivers/md/dm-clone-metadata.c
301
sb->flags = cpu_to_le32(0UL);
drivers/md/dm-clone-metadata.c
304
memset(sb->uuid, 0, sizeof(sb->uuid));
drivers/md/dm-clone-metadata.c
306
sb->magic = cpu_to_le64(SUPERBLOCK_MAGIC);
drivers/md/dm-clone-metadata.c
307
sb->version = cpu_to_le32(DM_CLONE_MAX_METADATA_VERSION);
drivers/md/dm-clone-metadata.c
310
memcpy(&sb->metadata_space_map_root, &cmd->metadata_space_map_root,
drivers/md/dm-clone-metadata.c
313
sb->region_size = cpu_to_le64(cmd->region_size);
drivers/md/dm-clone-metadata.c
314
sb->target_size = cpu_to_le64(cmd->target_size);
drivers/md/dm-clone-metadata.c
315
sb->bitset_root = cpu_to_le64(cmd->bitset_root);
drivers/md/dm-clone-metadata.c
322
struct superblock_disk *sb;
drivers/md/dm-clone-metadata.c
331
sb = dm_block_data(sblock);
drivers/md/dm-clone-metadata.c
334
if (cmd->region_size != le64_to_cpu(sb->region_size) ||
drivers/md/dm-clone-metadata.c
335
cmd->target_size != le64_to_cpu(sb->target_size)) {
drivers/md/dm-clone-metadata.c
342
sb->metadata_space_map_root,
drivers/md/dm-clone-metadata.c
343
sizeof(sb->metadata_space_map_root),
drivers/md/dm-clone-metadata.c
352
cmd->bitset_root = le64_to_cpu(sb->bitset_root);
drivers/md/dm-clone-metadata.c
364
struct superblock_disk *sb;
drivers/md/dm-clone-metadata.c
406
sb = dm_block_data(sblock);
drivers/md/dm-clone-metadata.c
407
__prepare_superblock(cmd, sb);
drivers/md/dm-clone-metadata.c
684
struct superblock_disk *sb;
drivers/md/dm-clone-metadata.c
715
sb = dm_block_data(sblock);
drivers/md/dm-clone-metadata.c
716
__prepare_superblock(cmd, sb);
drivers/md/dm-integrity.c
1663
if (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) {
drivers/md/dm-integrity.c
1664
r = crypto_shash_update(req, (__u8 *)&ic->sb->salt, SALT_SIZE);
drivers/md/dm-integrity.c
1722
if (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) {
drivers/md/dm-integrity.c
1724
sg_set_buf(s, (const __u8 *)&ic->sb->salt, SALT_SIZE);
drivers/md/dm-integrity.c
180
struct superblock *sb;
drivers/md/dm-integrity.c
1900
checksums = kmalloc((PAGE_SIZE >> SECTOR_SHIFT >> ic->sb->log2_sectors_per_block) * ic->tag_size + extra_space,
drivers/md/dm-integrity.c
1921
unsigned int this_step_blocks = bi_size >> (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
drivers/md/dm-integrity.c
1932
bi_size -= this_step_blocks << (SECTOR_SHIFT + ic->sb->log2_sectors_per_block);
drivers/md/dm-integrity.c
2106
unsigned int wanted_tag_size = bio_sectors(bio) >> ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
2315
(sector_t)ic->free_sectors << ic->sb->log2_sectors_per_block);
drivers/md/dm-integrity.c
2322
range_sectors = dio->range.n_sectors >> ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
2426
recalc_sector = le64_to_cpu(ic->sb->recalc_sector);
drivers/md/dm-integrity.c
2480
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING) &&
drivers/md/dm-integrity.c
2529
dio->payload_len = ic->tuple_size * (bio_sectors(bio) >> ic->sb->log2_sectors_per_block);
drivers/md/dm-integrity.c
2537
unsigned sectors = ((x_size - extra_size) / ic->tuple_size) << ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
2552
if (!(ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING)))
drivers/md/dm-integrity.c
2561
recalc_sector = le64_to_cpu(smp_load_acquire(&ic->sb->recalc_sector));
drivers/md/dm-integrity.c
2566
recalc_sector = le64_to_cpu(ic->sb->recalc_sector);
drivers/md/dm-integrity.c
2756
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING) &&
drivers/md/dm-integrity.c
2975
if (area2 != area || offset2 != offset + ((k - j) << ic->sb->log2_sectors_per_block))
drivers/md/dm-integrity.c
2984
io->range.n_sectors = (k - j) << ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
3035
integrity_sector_checksum(ic, &ic->journal_ahash_req, sec + ((l - j) << ic->sb->log2_sectors_per_block),
drivers/md/dm-integrity.c
3051
copy_from_journal(ic, i, j << ic->sb->log2_sectors_per_block,
drivers/md/dm-integrity.c
3052
(k - j) << ic->sb->log2_sectors_per_block,
drivers/md/dm-integrity.c
3138
if (recalc_sectors >= 1U << ic->sb->log2_sectors_per_block)
drivers/md/dm-integrity.c
3143
recalc_tags_size = (recalc_sectors >> ic->sb->log2_sectors_per_block) * ic->tag_size;
drivers/md/dm-integrity.c
3153
DEBUG_print("start recalculation... (position %llx)\n", le64_to_cpu(ic->sb->recalc_sector));
drivers/md/dm-integrity.c
3162
range.logical_sector = le64_to_cpu(ic->sb->recalc_sector);
drivers/md/dm-integrity.c
3175
range.n_sectors = min(range.n_sectors, ((sector_t)1U << ic->sb->log2_interleave_sectors) - (unsigned int)offset);
drivers/md/dm-integrity.c
3249
(ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit)) <<
drivers/md/dm-integrity.c
3250
(ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
3252
(ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit)) <<
drivers/md/dm-integrity.c
3253
(ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
3262
ic->sb->recalc_sector = cpu_to_le64(range.logical_sector + range.n_sectors);
drivers/md/dm-integrity.c
3302
if (recalc_sectors >= 1U << ic->sb->log2_sectors_per_block)
drivers/md/dm-integrity.c
3308
recalc_tags_size = (recalc_sectors >> ic->sb->log2_sectors_per_block) * ic->tuple_size;
drivers/md/dm-integrity.c
3324
range.logical_sector = le64_to_cpu(ic->sb->recalc_sector);
drivers/md/dm-integrity.c
3393
smp_store_release(&ic->sb->recalc_sector, cpu_to_le64(range.logical_sector + range.n_sectors));
drivers/md/dm-integrity.c
3395
ic->sb->recalc_sector = cpu_to_le64(range.logical_sector + range.n_sectors);
drivers/md/dm-integrity.c
3486
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING)) {
drivers/md/dm-integrity.c
3487
limit = le64_to_cpu(ic->sb->recalc_sector)
drivers/md/dm-integrity.c
3488
>> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit)
drivers/md/dm-integrity.c
3489
<< (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
3775
ic->sb->flags &= ~cpu_to_le32(SB_FLAG_DIRTY_BITMAP);
drivers/md/dm-integrity.c
3790
__u64 old_provided_data_sectors = le64_to_cpu(ic->sb->provided_data_sectors);
drivers/md/dm-integrity.c
3798
flags = ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
3802
if ((ic->sb->flags & flags) != flags) {
drivers/md/dm-integrity.c
3803
ic->sb->flags |= flags;
drivers/md/dm-integrity.c
3812
ic->sb->flags & cpu_to_le32(SB_FLAG_DIRTY_BITMAP) &&
drivers/md/dm-integrity.c
3813
ic->sb->log2_blocks_per_bitmap_bit == ic->log2_blocks_per_bitmap_bit) {
drivers/md/dm-integrity.c
3822
ic->sb->provided_data_sectors = cpu_to_le64(ic->provided_data_sectors);
drivers/md/dm-integrity.c
3828
if (ic->sb->flags & cpu_to_le32(SB_FLAG_DIRTY_BITMAP)) {
drivers/md/dm-integrity.c
3833
if (ic->sb->log2_blocks_per_bitmap_bit == ic->log2_blocks_per_bitmap_bit &&
drivers/md/dm-integrity.c
3839
ic->sb->flags |= cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
3840
ic->sb->recalc_sector = cpu_to_le64(0);
drivers/md/dm-integrity.c
3844
ic->sb->log2_blocks_per_bitmap_bit, ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
3845
ic->sb->log2_blocks_per_bitmap_bit = ic->log2_blocks_per_bitmap_bit;
drivers/md/dm-integrity.c
3851
ic->sb->flags |= cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
3852
ic->sb->recalc_sector = cpu_to_le64(0);
drivers/md/dm-integrity.c
3855
if (!(ic->sb->log2_blocks_per_bitmap_bit == ic->log2_blocks_per_bitmap_bit &&
drivers/md/dm-integrity.c
3858
ic->sb->flags |= cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
3859
ic->sb->recalc_sector = cpu_to_le64(0);
drivers/md/dm-integrity.c
3863
ic->sb->flags &= ~cpu_to_le32(SB_FLAG_DIRTY_BITMAP);
drivers/md/dm-integrity.c
3871
ic->sb->flags |= cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
3872
ic->sb->recalc_sector = cpu_to_le64(0);
drivers/md/dm-integrity.c
3875
ic->sb->flags |= cpu_to_le32(SB_FLAG_DIRTY_BITMAP);
drivers/md/dm-integrity.c
3876
ic->sb->log2_blocks_per_bitmap_bit = ic->log2_blocks_per_bitmap_bit;
drivers/md/dm-integrity.c
3884
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING) &&
drivers/md/dm-integrity.c
3885
le64_to_cpu(ic->sb->recalc_sector) < ic->provided_data_sectors) {
drivers/md/dm-integrity.c
3886
block_bitmap_op(ic, ic->journal, le64_to_cpu(ic->sb->recalc_sector),
drivers/md/dm-integrity.c
3887
ic->provided_data_sectors - le64_to_cpu(ic->sb->recalc_sector), BITMAP_OP_SET);
drivers/md/dm-integrity.c
3888
block_bitmap_op(ic, ic->recalc_bitmap, le64_to_cpu(ic->sb->recalc_sector),
drivers/md/dm-integrity.c
3889
ic->provided_data_sectors - le64_to_cpu(ic->sb->recalc_sector), BITMAP_OP_SET);
drivers/md/dm-integrity.c
3890
block_bitmap_op(ic, ic->may_write_bitmap, le64_to_cpu(ic->sb->recalc_sector),
drivers/md/dm-integrity.c
3891
ic->provided_data_sectors - le64_to_cpu(ic->sb->recalc_sector), BITMAP_OP_SET);
drivers/md/dm-integrity.c
3898
DEBUG_print("testing recalc: %x\n", ic->sb->flags);
drivers/md/dm-integrity.c
3899
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING)) {
drivers/md/dm-integrity.c
3900
__u64 recalc_pos = le64_to_cpu(ic->sb->recalc_sector);
drivers/md/dm-integrity.c
3906
ic->sb->recalc_sector = cpu_to_le64(ic->provided_data_sectors);
drivers/md/dm-integrity.c
3934
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING))
drivers/md/dm-integrity.c
3935
DMEMIT(" %llu", le64_to_cpu(ic->sb->recalc_sector));
drivers/md/dm-integrity.c
3944
arg_count += !!(ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING));
drivers/md/dm-integrity.c
3956
arg_count += (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_PADDING)) != 0;
drivers/md/dm-integrity.c
3957
arg_count += (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) != 0;
drivers/md/dm-integrity.c
3965
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING))
drivers/md/dm-integrity.c
397
if (!(ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) ?
drivers/md/dm-integrity.c
3972
DMEMIT(" interleave_sectors:%u", 1U << ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
3987
if ((ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_PADDING)) != 0)
drivers/md/dm-integrity.c
3989
if ((ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) != 0)
drivers/md/dm-integrity.c
4017
DMEMIT(",recalculate=%c", (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING)) ?
drivers/md/dm-integrity.c
4021
((ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_PADDING)) != 0) ? 'y' : 'n');
drivers/md/dm-integrity.c
4023
((ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) != 0) ? 'y' : 'n');
drivers/md/dm-integrity.c
4027
DMEMIT(",interleave_sectors=%u", 1U << ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
4064
ic->sb->log2_sectors_per_block + SECTOR_SHIFT;
drivers/md/dm-integrity.c
4074
ic->journal_sections = le32_to_cpu(ic->sb->journal_sections);
drivers/md/dm-integrity.c
4078
if (ic->sb->flags & cpu_to_le32(SB_FLAG_HAVE_JOURNAL_MAC))
drivers/md/dm-integrity.c
4082
ic->journal_section_sectors = (ic->journal_section_entries << ic->sb->log2_sectors_per_block) + JOURNAL_BLOCK_SECTORS;
drivers/md/dm-integrity.c
4104
ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_PADDING) ?
drivers/md/dm-integrity.c
4108
ic->metadata_run = round_up((__u64)ic->tag_size << (ic->sb->log2_interleave_sectors - ic->sb->log2_sectors_per_block),
drivers/md/dm-integrity.c
4120
__u64 meta_size = (ic->provided_data_sectors >> ic->sb->log2_sectors_per_block) * ic->tag_size;
drivers/md/dm-integrity.c
4160
memset(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT);
drivers/md/dm-integrity.c
4161
memcpy(ic->sb->magic, SB_MAGIC, 8);
drivers/md/dm-integrity.c
4163
ic->sb->flags |= cpu_to_le32(SB_FLAG_INLINE);
drivers/md/dm-integrity.c
4164
ic->sb->integrity_tag_size = cpu_to_le16(ic->tag_size);
drivers/md/dm-integrity.c
4165
ic->sb->log2_sectors_per_block = __ffs(ic->sectors_per_block);
drivers/md/dm-integrity.c
4167
ic->sb->flags |= cpu_to_le32(SB_FLAG_HAVE_JOURNAL_MAC);
drivers/md/dm-integrity.c
4177
ic->sb->flags |= cpu_to_le32(SB_FLAG_FIXED_HMAC);
drivers/md/dm-integrity.c
4178
get_random_bytes(ic->sb->salt, SALT_SIZE);
drivers/md/dm-integrity.c
418
__u8 log2_interleave_sectors = ic->sb->log2_interleave_sectors;
drivers/md/dm-integrity.c
4183
ic->sb->flags |= cpu_to_le32(SB_FLAG_FIXED_PADDING);
drivers/md/dm-integrity.c
4184
ic->sb->journal_sections = cpu_to_le32(journal_sections);
drivers/md/dm-integrity.c
4187
ic->sb->log2_interleave_sectors = __fls(interleave_sectors);
drivers/md/dm-integrity.c
4188
ic->sb->log2_interleave_sectors = max_t(__u8, MIN_LOG2_INTERLEAVE_SECTORS, ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
4189
ic->sb->log2_interleave_sectors = min_t(__u8, MAX_LOG2_INTERLEAVE_SECTORS, ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
4195
ic->sb->log2_interleave_sectors = 0;
drivers/md/dm-integrity.c
4202
ic->sb->journal_sections = cpu_to_le32(0);
drivers/md/dm-integrity.c
4204
__u32 prev_journal_sections = le32_to_cpu(ic->sb->journal_sections);
drivers/md/dm-integrity.c
4209
ic->sb->journal_sections = cpu_to_le32(test_journal_sections);
drivers/md/dm-integrity.c
4211
ic->sb->journal_sections = cpu_to_le32(prev_journal_sections);
drivers/md/dm-integrity.c
4214
if (!le32_to_cpu(ic->sb->journal_sections)) {
drivers/md/dm-integrity.c
4223
ic->sb->provided_data_sectors = cpu_to_le64(ic->provided_data_sectors);
drivers/md/dm-integrity.c
430
(n) >>= (ic)->sb->log2_sectors_per_block; \
drivers/md/dm-integrity.c
439
ms = area << ic->sb->log2_interleave_sectors;
drivers/md/dm-integrity.c
466
result = area << ic->sb->log2_interleave_sectors;
drivers/md/dm-integrity.c
486
if (ic->sb->flags & cpu_to_le32(SB_FLAG_INLINE))
drivers/md/dm-integrity.c
487
ic->sb->version = SB_VERSION_6;
drivers/md/dm-integrity.c
488
else if (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC))
drivers/md/dm-integrity.c
489
ic->sb->version = SB_VERSION_5;
drivers/md/dm-integrity.c
490
else if (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_PADDING))
drivers/md/dm-integrity.c
491
ic->sb->version = SB_VERSION_4;
drivers/md/dm-integrity.c
492
else if (ic->mode == 'B' || ic->sb->flags & cpu_to_le32(SB_FLAG_DIRTY_BITMAP))
drivers/md/dm-integrity.c
493
ic->sb->version = SB_VERSION_3;
drivers/md/dm-integrity.c
494
else if (ic->meta_dev || ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING))
drivers/md/dm-integrity.c
495
ic->sb->version = SB_VERSION_2;
drivers/md/dm-integrity.c
497
ic->sb->version = SB_VERSION_1;
drivers/md/dm-integrity.c
505
__u8 *sb = (__u8 *)ic->sb;
drivers/md/dm-integrity.c
5053
ic->sb = alloc_pages_exact(SB_SECTORS << SECTOR_SHIFT, GFP_KERNEL);
drivers/md/dm-integrity.c
5054
if (!ic->sb) {
drivers/md/dm-integrity.c
506
__u8 *mac = sb + (1 << SECTOR_SHIFT) - mac_size;
drivers/md/dm-integrity.c
5066
if (memcmp(ic->sb->magic, SB_MAGIC, 8)) {
drivers/md/dm-integrity.c
5068
if (memchr_inv(ic->sb, 0, SB_SECTORS << SECTOR_SHIFT)) {
drivers/md/dm-integrity.c
5084
if (!ic->sb->version || ic->sb->version > SB_VERSION_6) {
drivers/md/dm-integrity.c
5089
if (!!(ic->sb->flags & cpu_to_le32(SB_FLAG_INLINE)) != (ic->mode == 'I')) {
drivers/md/dm-integrity.c
5094
if (le16_to_cpu(ic->sb->integrity_tag_size) != ic->tag_size) {
drivers/md/dm-integrity.c
5099
if (ic->sb->log2_sectors_per_block != __ffs(ic->sectors_per_block)) {
drivers/md/dm-integrity.c
5105
if (!le32_to_cpu(ic->sb->journal_sections)) {
drivers/md/dm-integrity.c
5111
if (le32_to_cpu(ic->sb->journal_sections)) {
drivers/md/dm-integrity.c
5119
if (ic->sb->log2_interleave_sectors < MIN_LOG2_INTERLEAVE_SECTORS ||
drivers/md/dm-integrity.c
5120
ic->sb->log2_interleave_sectors > MAX_LOG2_INTERLEAVE_SECTORS) {
drivers/md/dm-integrity.c
5126
if (ic->sb->log2_interleave_sectors) {
drivers/md/dm-integrity.c
5132
if (!!(ic->sb->flags & cpu_to_le32(SB_FLAG_HAVE_JOURNAL_MAC)) != !!ic->journal_mac_alg.alg_string) {
drivers/md/dm-integrity.c
5160
if (log2_sectors_per_bitmap_bit < ic->sb->log2_sectors_per_block)
drivers/md/dm-integrity.c
5161
log2_sectors_per_bitmap_bit = ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
517
r = crypto_shash_digest(desc, sb, mac - sb, mac);
drivers/md/dm-integrity.c
5170
log2_blocks_per_bitmap_bit = log2_sectors_per_bitmap_bit - ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
5173
ic->sb->log2_blocks_per_bitmap_bit = log2_blocks_per_bitmap_bit;
drivers/md/dm-integrity.c
5175
n_bitmap_bits = ((ic->provided_data_sectors >> ic->sb->log2_sectors_per_block)
drivers/md/dm-integrity.c
5194
DEBUG_print(" integrity_tag_size %u\n", le16_to_cpu(ic->sb->integrity_tag_size));
drivers/md/dm-integrity.c
5199
DEBUG_print(" journal_sections %u\n", (unsigned int)le32_to_cpu(ic->sb->journal_sections));
drivers/md/dm-integrity.c
5201
DEBUG_print(" log2_interleave_sectors %d\n", ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
5210
if (ic->recalculate_flag && !(ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING))) {
drivers/md/dm-integrity.c
5211
ic->sb->flags |= cpu_to_le32(SB_FLAG_RECALCULATING);
drivers/md/dm-integrity.c
5212
ic->sb->recalc_sector = cpu_to_le64(0);
drivers/md/dm-integrity.c
5225
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING)) {
drivers/md/dm-integrity.c
5232
if (ic->sb->flags & cpu_to_le32(SB_FLAG_RECALCULATING) &&
drivers/md/dm-integrity.c
5233
le64_to_cpu(ic->sb->recalc_sector) < ic->provided_data_sectors &&
drivers/md/dm-integrity.c
525
r = crypto_shash_digest(desc, sb, mac - sb, actual_mac);
drivers/md/dm-integrity.c
5315
r = dm_set_target_max_io_len(ti, 1U << ic->sb->log2_interleave_sectors);
drivers/md/dm-integrity.c
5410
if (ic->sb)
drivers/md/dm-integrity.c
5411
free_pages_exact(ic->sb, SB_SECTORS << SECTOR_SHIFT);
drivers/md/dm-integrity.c
549
io_req.mem.ptr.addr = ic->sb;
drivers/md/dm-integrity.c
558
if (ic->journal_mac && ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) {
drivers/md/dm-integrity.c
570
if (ic->mode != 'R' && ic->journal_mac && ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) {
drivers/md/dm-integrity.c
591
if (unlikely(((sector | n_sectors) & ((1 << ic->sb->log2_sectors_per_block) - 1)) != 0)) {
drivers/md/dm-integrity.c
595
ic->sb->log2_sectors_per_block,
drivers/md/dm-integrity.c
604
bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
606
(ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
706
unsigned int bit = sector >> (ic->sb->log2_sectors_per_block + ic->log2_blocks_per_bitmap_bit);
drivers/md/dm-integrity.c
778
n <<= ic->sb->log2_sectors_per_block;
drivers/md/dm-integrity.c
801
if (ic->sb->flags & cpu_to_le32(SB_FLAG_FIXED_HMAC)) {
drivers/md/dm-integrity.c
804
r = crypto_shash_update(desc, (__u8 *)&ic->sb->salt, SALT_SIZE);
drivers/md/dm-pcache/cache_dev.c
130
static int sb_read(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb)
drivers/md/dm-pcache/cache_dev.c
134
if (copy_mc_to_kernel(sb, sb_addr, sizeof(struct pcache_sb)))
drivers/md/dm-pcache/cache_dev.c
140
static void sb_write(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb)
drivers/md/dm-pcache/cache_dev.c
144
memcpy_flushcache(sb_addr, sb, sizeof(struct pcache_sb));
drivers/md/dm-pcache/cache_dev.c
148
static int sb_init(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb)
drivers/md/dm-pcache/cache_dev.c
156
magic = le64_to_cpu(sb->magic);
drivers/md/dm-pcache/cache_dev.c
172
sb->flags = cpu_to_le32(flags);
drivers/md/dm-pcache/cache_dev.c
173
sb->magic = cpu_to_le64(PCACHE_MAGIC);
drivers/md/dm-pcache/cache_dev.c
174
sb->seg_num = cpu_to_le32(nr_segs);
drivers/md/dm-pcache/cache_dev.c
175
sb->crc = cpu_to_le32(crc32c(PCACHE_CRC_SEED, (void *)(sb) + 4, sizeof(struct pcache_sb) - 4));
drivers/md/dm-pcache/cache_dev.c
184
static int sb_validate(struct pcache_cache_dev *cache_dev, struct pcache_sb *sb)
drivers/md/dm-pcache/cache_dev.c
190
if (le64_to_cpu(sb->magic) != PCACHE_MAGIC) {
drivers/md/dm-pcache/cache_dev.c
192
le64_to_cpu(sb->magic));
drivers/md/dm-pcache/cache_dev.c
196
crc = crc32c(PCACHE_CRC_SEED, (void *)(sb) + 4, sizeof(struct pcache_sb) - 4);
drivers/md/dm-pcache/cache_dev.c
197
if (crc != le32_to_cpu(sb->crc)) {
drivers/md/dm-pcache/cache_dev.c
198
pcache_dev_err(pcache, "corrupted sb: %u, expected: %u\n", crc, le32_to_cpu(sb->crc));
drivers/md/dm-pcache/cache_dev.c
202
flags = le32_to_cpu(sb->flags);
drivers/md/dm-pcache/cache_dev.c
243
struct pcache_sb sb;
drivers/md/dm-pcache/cache_dev.c
256
ret = sb_read(cache_dev, &sb);
drivers/md/dm-pcache/cache_dev.c
260
if (le64_to_cpu(sb.magic) == 0) {
drivers/md/dm-pcache/cache_dev.c
262
ret = sb_init(cache_dev, &sb);
drivers/md/dm-pcache/cache_dev.c
267
ret = sb_validate(cache_dev, &sb);
drivers/md/dm-pcache/cache_dev.c
271
cache_dev->sb_flags = le32_to_cpu(sb.flags);
drivers/md/dm-pcache/cache_dev.c
272
ret = cache_dev_init(cache_dev, le32_to_cpu(sb.seg_num));
drivers/md/dm-pcache/cache_dev.c
277
sb_write(cache_dev, &sb);
drivers/md/dm-raid.c
2083
static void sb_retrieve_failed_devices(struct dm_raid_superblock *sb, uint64_t *failed_devices)
drivers/md/dm-raid.c
2085
failed_devices[0] = le64_to_cpu(sb->failed_devices);
drivers/md/dm-raid.c
2086
memset(failed_devices + 1, 0, sizeof(sb->extended_failed_devices));
drivers/md/dm-raid.c
2088
if (le32_to_cpu(sb->compat_features) & FEATURE_FLAG_SUPPORTS_V190) {
drivers/md/dm-raid.c
2089
int i = ARRAY_SIZE(sb->extended_failed_devices);
drivers/md/dm-raid.c
2092
failed_devices[i+1] = le64_to_cpu(sb->extended_failed_devices[i]);
drivers/md/dm-raid.c
2096
static void sb_update_failed_devices(struct dm_raid_superblock *sb, uint64_t *failed_devices)
drivers/md/dm-raid.c
2098
int i = ARRAY_SIZE(sb->extended_failed_devices);
drivers/md/dm-raid.c
2100
sb->failed_devices = cpu_to_le64(failed_devices[0]);
drivers/md/dm-raid.c
2102
sb->extended_failed_devices[i] = cpu_to_le64(failed_devices[i+1]);
drivers/md/dm-raid.c
2115
struct dm_raid_superblock *sb;
drivers/md/dm-raid.c
2124
sb = page_address(rdev->sb_page);
drivers/md/dm-raid.c
2126
sb_retrieve_failed_devices(sb, failed_devices);
drivers/md/dm-raid.c
2135
sb_update_failed_devices(sb, failed_devices);
drivers/md/dm-raid.c
2137
sb->magic = cpu_to_le32(DM_RAID_MAGIC);
drivers/md/dm-raid.c
2138
sb->compat_features = cpu_to_le32(FEATURE_FLAG_SUPPORTS_V190);
drivers/md/dm-raid.c
2140
sb->num_devices = cpu_to_le32(mddev->raid_disks);
drivers/md/dm-raid.c
2141
sb->array_position = cpu_to_le32(rdev->raid_disk);
drivers/md/dm-raid.c
2143
sb->events = cpu_to_le64(mddev->events);
drivers/md/dm-raid.c
2145
sb->disk_recovery_offset = cpu_to_le64(rdev->recovery_offset);
drivers/md/dm-raid.c
2146
sb->array_resync_offset = cpu_to_le64(mddev->resync_offset);
drivers/md/dm-raid.c
2148
sb->level = cpu_to_le32(mddev->level);
drivers/md/dm-raid.c
2149
sb->layout = cpu_to_le32(mddev->layout);
drivers/md/dm-raid.c
2150
sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors);
drivers/md/dm-raid.c
2157
sb->new_level = cpu_to_le32(mddev->new_level);
drivers/md/dm-raid.c
2158
sb->new_layout = cpu_to_le32(mddev->new_layout);
drivers/md/dm-raid.c
2159
sb->new_stripe_sectors = cpu_to_le32(mddev->new_chunk_sectors);
drivers/md/dm-raid.c
2161
sb->delta_disks = cpu_to_le32(mddev->delta_disks);
drivers/md/dm-raid.c
2164
sb->reshape_position = cpu_to_le64(mddev->reshape_position);
drivers/md/dm-raid.c
2165
if (le64_to_cpu(sb->reshape_position) != MaxSector) {
drivers/md/dm-raid.c
2167
sb->flags |= cpu_to_le32(SB_FLAG_RESHAPE_ACTIVE);
drivers/md/dm-raid.c
2170
sb->flags |= cpu_to_le32(SB_FLAG_RESHAPE_BACKWARDS);
drivers/md/dm-raid.c
2173
sb->flags &= ~(cpu_to_le32(SB_FLAG_RESHAPE_ACTIVE|SB_FLAG_RESHAPE_BACKWARDS));
drivers/md/dm-raid.c
2176
sb->array_sectors = cpu_to_le64(mddev->array_sectors);
drivers/md/dm-raid.c
2177
sb->data_offset = cpu_to_le64(rdev->data_offset);
drivers/md/dm-raid.c
2178
sb->new_data_offset = cpu_to_le64(rdev->new_data_offset);
drivers/md/dm-raid.c
2179
sb->sectors = cpu_to_le64(rdev->sectors);
drivers/md/dm-raid.c
2180
sb->incompat_features = cpu_to_le32(0);
drivers/md/dm-raid.c
2183
memset(sb + 1, 0, rdev->sb_size - sizeof(*sb));
drivers/md/dm-raid.c
2197
struct dm_raid_superblock *sb;
drivers/md/dm-raid.c
2205
sb = page_address(rdev->sb_page);
drivers/md/dm-raid.c
2212
if ((sb->magic != cpu_to_le32(DM_RAID_MAGIC)) ||
drivers/md/dm-raid.c
2217
sb->compat_features = cpu_to_le32(FEATURE_FLAG_SUPPORTS_V190);
drivers/md/dm-raid.c
2229
events_sb = le64_to_cpu(sb->events);
drivers/md/dm-raid.c
2243
struct dm_raid_superblock *sb;
drivers/md/dm-raid.c
2248
sb = page_address(rdev->sb_page);
drivers/md/dm-raid.c
2249
events_sb = le64_to_cpu(sb->events);
drivers/md/dm-raid.c
2258
mddev->raid_disks = le32_to_cpu(sb->num_devices);
drivers/md/dm-raid.c
2259
mddev->level = le32_to_cpu(sb->level);
drivers/md/dm-raid.c
2260
mddev->layout = le32_to_cpu(sb->layout);
drivers/md/dm-raid.c
2261
mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors);
drivers/md/dm-raid.c
2267
if (le32_to_cpu(sb->compat_features) & FEATURE_FLAG_SUPPORTS_V190) {
drivers/md/dm-raid.c
2269
mddev->new_level = le32_to_cpu(sb->new_level);
drivers/md/dm-raid.c
2270
mddev->new_layout = le32_to_cpu(sb->new_layout);
drivers/md/dm-raid.c
2271
mddev->new_chunk_sectors = le32_to_cpu(sb->new_stripe_sectors);
drivers/md/dm-raid.c
2272
mddev->delta_disks = le32_to_cpu(sb->delta_disks);
drivers/md/dm-raid.c
2273
mddev->array_sectors = le64_to_cpu(sb->array_sectors);
drivers/md/dm-raid.c
2276
if (le32_to_cpu(sb->flags) & SB_FLAG_RESHAPE_ACTIVE) {
drivers/md/dm-raid.c
2283
(!mddev->delta_disks && (le32_to_cpu(sb->flags) & SB_FLAG_RESHAPE_BACKWARDS)))
drivers/md/dm-raid.c
2288
mddev->reshape_position = le64_to_cpu(sb->reshape_position);
drivers/md/dm-raid.c
2316
le32_to_cpu(sb->layout), mddev->new_layout);
drivers/md/dm-raid.c
2339
mddev->resync_offset = le64_to_cpu(sb->array_resync_offset);
drivers/md/dm-raid.c
2409
sb_retrieve_failed_devices(sb, failed_devices);
drivers/md/dm-raid.c
2462
struct dm_raid_superblock *sb;
drivers/md/dm-raid.c
2467
sb = page_address(rdev->sb_page);
drivers/md/dm-raid.c
2476
if (le32_to_cpu(sb->compat_features) &&
drivers/md/dm-raid.c
2477
le32_to_cpu(sb->compat_features) != FEATURE_FLAG_SUPPORTS_V190) {
drivers/md/dm-raid.c
2482
if (sb->incompat_features) {
drivers/md/dm-raid.c
2497
if (le32_to_cpu(sb->compat_features) & FEATURE_FLAG_SUPPORTS_V190)
drivers/md/dm-raid.c
2498
rdev->sectors = le64_to_cpu(sb->sectors);
drivers/md/dm-raid.c
2500
rdev->recovery_offset = le64_to_cpu(sb->disk_recovery_offset);
drivers/md/dm-raid.c
2521
rdev->data_offset = le64_to_cpu(sb->data_offset);
drivers/md/dm-raid.c
2522
rdev->new_data_offset = le64_to_cpu(sb->new_data_offset);
drivers/md/dm-raid.c
3875
struct dm_raid_superblock *sb;
drivers/md/dm-raid.c
3941
sb = page_address(r->sb_page);
drivers/md/dm-raid.c
3942
sb_retrieve_failed_devices(sb, failed_devices);
drivers/md/dm-raid.c
3947
sb_update_failed_devices(sb, failed_devices);
drivers/md/dm-writecache.c
1031
r = copy_mc_to_kernel(&sb_seq_count, &sb(wc)->seq_count,
drivers/md/dm-writecache.c
2160
for (b = 0; b < ARRAY_SIZE(sb(wc)->padding); b++)
drivers/md/dm-writecache.c
2161
pmem_assign(sb(wc)->padding[b], cpu_to_le64(0));
drivers/md/dm-writecache.c
2162
pmem_assign(sb(wc)->version, cpu_to_le32(MEMORY_SUPERBLOCK_VERSION));
drivers/md/dm-writecache.c
2163
pmem_assign(sb(wc)->block_size, cpu_to_le32(wc->block_size));
drivers/md/dm-writecache.c
2164
pmem_assign(sb(wc)->n_blocks, cpu_to_le64(wc->n_blocks));
drivers/md/dm-writecache.c
2165
pmem_assign(sb(wc)->seq_count, cpu_to_le64(0));
drivers/md/dm-writecache.c
2174
pmem_assign(sb(wc)->magic, cpu_to_le32(MEMORY_SUPERBLOCK_MAGIC));
drivers/md/dm-writecache.c
2175
writecache_flush_region(wc, &sb(wc)->magic, sizeof(sb(wc)->magic));
drivers/md/dm-writecache.c
2581
r = copy_mc_to_kernel(&s, sb(wc), sizeof(struct wc_memory_superblock));
drivers/md/dm-writecache.c
2592
r = copy_mc_to_kernel(&s, sb(wc),
drivers/md/dm-writecache.c
2621
if (offset / sizeof(struct wc_memory_entry) != le64_to_cpu(sb(wc)->n_blocks)) {
drivers/md/dm-writecache.c
2642
wc->block_start = (char *)sb(wc) + offset;
drivers/md/dm-writecache.c
394
return &sb(wc)->entries[e->index];
drivers/md/dm-writecache.c
826
pmem_assign(sb(wc)->seq_count, cpu_to_le64(wc->seq_count));
drivers/md/dm-zoned-metadata.c
1000
gen = le64_to_cpu(sb->gen);
drivers/md/dm-zoned-metadata.c
1001
stored_crc = le32_to_cpu(sb->crc);
drivers/md/dm-zoned-metadata.c
1002
sb->crc = 0;
drivers/md/dm-zoned-metadata.c
1003
crc = crc32_le(gen, (unsigned char *)sb, DMZ_BLOCK_SIZE);
drivers/md/dm-zoned-metadata.c
1010
sb_block = le64_to_cpu(sb->sb_block);
drivers/md/dm-zoned-metadata.c
1019
import_uuid(&sb_uuid, sb->dmz_uuid);
drivers/md/dm-zoned-metadata.c
1031
memcpy(zmd->label, sb->dmz_label, BDEVNAME_SIZE);
drivers/md/dm-zoned-metadata.c
1032
else if (memcmp(zmd->label, sb->dmz_label, BDEVNAME_SIZE)) {
drivers/md/dm-zoned-metadata.c
1034
sb->dmz_label, zmd->label);
drivers/md/dm-zoned-metadata.c
1037
import_uuid(&dev->uuid, sb->dev_uuid);
drivers/md/dm-zoned-metadata.c
1055
nr_meta_zones = (le32_to_cpu(sb->nr_meta_blocks) + zmd->zone_nr_blocks - 1)
drivers/md/dm-zoned-metadata.c
1064
if (!le32_to_cpu(sb->nr_reserved_seq) ||
drivers/md/dm-zoned-metadata.c
1065
le32_to_cpu(sb->nr_reserved_seq) >= (zmd->nr_useable_zones - nr_meta_zones)) {
drivers/md/dm-zoned-metadata.c
1071
(nr_meta_zones * 2 + le32_to_cpu(sb->nr_reserved_seq));
drivers/md/dm-zoned-metadata.c
1072
if (le32_to_cpu(sb->nr_chunks) > nr_data_zones) {
drivers/md/dm-zoned-metadata.c
1074
le32_to_cpu(sb->nr_chunks), nr_data_zones);
drivers/md/dm-zoned-metadata.c
1079
zmd->nr_meta_blocks = le32_to_cpu(sb->nr_meta_blocks);
drivers/md/dm-zoned-metadata.c
1080
zmd->nr_reserved_seq = le32_to_cpu(sb->nr_reserved_seq);
drivers/md/dm-zoned-metadata.c
1081
zmd->nr_chunks = le32_to_cpu(sb->nr_chunks);
drivers/md/dm-zoned-metadata.c
1082
zmd->nr_map_blocks = le32_to_cpu(sb->nr_map_blocks);
drivers/md/dm-zoned-metadata.c
1083
zmd->nr_bitmap_blocks = le32_to_cpu(sb->nr_bitmap_blocks);
drivers/md/dm-zoned-metadata.c
1093
static int dmz_read_sb(struct dmz_metadata *zmd, struct dmz_sb *sb, int set)
drivers/md/dm-zoned-metadata.c
1096
set, sb->dev->bdev, sb->block);
drivers/md/dm-zoned-metadata.c
1098
return dmz_rdwr_block(sb->dev, REQ_OP_READ,
drivers/md/dm-zoned-metadata.c
1099
sb->block, sb->mblk->page);
drivers/md/dm-zoned-metadata.c
1111
unsigned int zone_id = zmd->sb[0].zone->id;
drivers/md/dm-zoned-metadata.c
1119
zmd->sb[1].mblk = mblk;
drivers/md/dm-zoned-metadata.c
1120
zmd->sb[1].sb = mblk->data;
drivers/md/dm-zoned-metadata.c
1123
zmd->sb[1].block = zmd->sb[0].block + zone_nr_blocks;
drivers/md/dm-zoned-metadata.c
1124
zmd->sb[1].zone = dmz_get(zmd, zone_id + 1);
drivers/md/dm-zoned-metadata.c
1125
zmd->sb[1].dev = zmd->sb[0].dev;
drivers/md/dm-zoned-metadata.c
1127
if (dmz_read_sb(zmd, &zmd->sb[1], 1) != 0)
drivers/md/dm-zoned-metadata.c
1129
if (le32_to_cpu(zmd->sb[1].sb->magic) == DMZ_MAGIC)
drivers/md/dm-zoned-metadata.c
1131
zmd->sb[1].block += zone_nr_blocks;
drivers/md/dm-zoned-metadata.c
1132
zmd->sb[1].zone = dmz_get(zmd, zone_id + i);
drivers/md/dm-zoned-metadata.c
1136
zmd->sb[1].mblk = NULL;
drivers/md/dm-zoned-metadata.c
1137
zmd->sb[1].zone = NULL;
drivers/md/dm-zoned-metadata.c
1138
zmd->sb[1].dev = NULL;
drivers/md/dm-zoned-metadata.c
1146
static int dmz_get_sb(struct dmz_metadata *zmd, struct dmz_sb *sb, int set)
drivers/md/dm-zoned-metadata.c
1156
sb->mblk = mblk;
drivers/md/dm-zoned-metadata.c
1157
sb->sb = mblk->data;
drivers/md/dm-zoned-metadata.c
1160
ret = dmz_read_sb(zmd, sb, set);
drivers/md/dm-zoned-metadata.c
1163
sb->mblk = NULL;
drivers/md/dm-zoned-metadata.c
1179
dmz_dev_warn(zmd->sb[dst_set].dev,
drivers/md/dm-zoned-metadata.c
1183
zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone);
drivers/md/dm-zoned-metadata.c
1185
zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone);
drivers/md/dm-zoned-metadata.c
1193
ret = dmz_rdwr_block(zmd->sb[src_set].dev, REQ_OP_READ,
drivers/md/dm-zoned-metadata.c
1194
zmd->sb[src_set].block + i, page);
drivers/md/dm-zoned-metadata.c
1197
ret = dmz_rdwr_block(zmd->sb[dst_set].dev, REQ_OP_WRITE,
drivers/md/dm-zoned-metadata.c
1198
zmd->sb[dst_set].block + i, page);
drivers/md/dm-zoned-metadata.c
1204
if (!zmd->sb[dst_set].mblk) {
drivers/md/dm-zoned-metadata.c
1205
zmd->sb[dst_set].mblk = dmz_alloc_mblock(zmd, 0);
drivers/md/dm-zoned-metadata.c
1206
if (!zmd->sb[dst_set].mblk) {
drivers/md/dm-zoned-metadata.c
1210
zmd->sb[dst_set].sb = zmd->sb[dst_set].mblk->data;
drivers/md/dm-zoned-metadata.c
1229
if (!zmd->sb[0].zone) {
drivers/md/dm-zoned-metadata.c
1235
zmd->sb[0].block = dmz_start_block(zmd, zmd->sb[0].zone);
drivers/md/dm-zoned-metadata.c
1236
zmd->sb[0].dev = zmd->sb[0].zone->dev;
drivers/md/dm-zoned-metadata.c
1237
ret = dmz_get_sb(zmd, &zmd->sb[0], 0);
drivers/md/dm-zoned-metadata.c
1239
dmz_dev_err(zmd->sb[0].dev, "Read primary super block failed");
drivers/md/dm-zoned-metadata.c
1243
ret = dmz_check_sb(zmd, &zmd->sb[0], false);
drivers/md/dm-zoned-metadata.c
1248
if (!zmd->sb[1].zone) {
drivers/md/dm-zoned-metadata.c
1250
zmd->sb[0].zone->id + zmd->nr_meta_zones;
drivers/md/dm-zoned-metadata.c
1252
zmd->sb[1].zone = dmz_get(zmd, zone_id);
drivers/md/dm-zoned-metadata.c
1254
zmd->sb[1].block = dmz_start_block(zmd, zmd->sb[1].zone);
drivers/md/dm-zoned-metadata.c
1255
zmd->sb[1].dev = zmd->sb[0].dev;
drivers/md/dm-zoned-metadata.c
1256
ret = dmz_get_sb(zmd, &zmd->sb[1], 1);
drivers/md/dm-zoned-metadata.c
1261
dmz_dev_err(zmd->sb[1].dev, "Read secondary super block failed");
drivers/md/dm-zoned-metadata.c
1265
ret = dmz_check_sb(zmd, &zmd->sb[1], false);
drivers/md/dm-zoned-metadata.c
1276
sb_gen[0] = le64_to_cpu(zmd->sb[0].sb->gen);
drivers/md/dm-zoned-metadata.c
1280
dmz_dev_err(zmd->sb[0].dev,
drivers/md/dm-zoned-metadata.c
1287
sb_gen[1] = le64_to_cpu(zmd->sb[1].sb->gen);
drivers/md/dm-zoned-metadata.c
1292
dmz_dev_err(zmd->sb[1].dev,
drivers/md/dm-zoned-metadata.c
1306
dmz_dev_debug(zmd->sb[zmd->mblk_primary].dev,
drivers/md/dm-zoned-metadata.c
1312
struct dmz_sb *sb;
drivers/md/dm-zoned-metadata.c
1314
sb = kzalloc_obj(struct dmz_sb);
drivers/md/dm-zoned-metadata.c
1315
if (!sb)
drivers/md/dm-zoned-metadata.c
1318
sb->block = 0;
drivers/md/dm-zoned-metadata.c
1319
sb->zone = dmz_get(zmd, zmd->dev[i].zone_offset);
drivers/md/dm-zoned-metadata.c
1320
sb->dev = &zmd->dev[i];
drivers/md/dm-zoned-metadata.c
1321
if (!dmz_is_meta(sb->zone)) {
drivers/md/dm-zoned-metadata.c
1322
dmz_dev_err(sb->dev,
drivers/md/dm-zoned-metadata.c
1324
sb->zone->id);
drivers/md/dm-zoned-metadata.c
1328
ret = dmz_get_sb(zmd, sb, i + 1);
drivers/md/dm-zoned-metadata.c
1330
dmz_dev_err(sb->dev,
drivers/md/dm-zoned-metadata.c
1332
dmz_free_mblock(zmd, sb->mblk);
drivers/md/dm-zoned-metadata.c
1335
ret = dmz_check_sb(zmd, sb, true);
drivers/md/dm-zoned-metadata.c
1336
dmz_free_mblock(zmd, sb->mblk);
drivers/md/dm-zoned-metadata.c
1341
kfree(sb);
drivers/md/dm-zoned-metadata.c
136
struct dmz_super *sb;
drivers/md/dm-zoned-metadata.c
1402
if (zmd->nr_devs == 1 && !zmd->sb[0].zone) {
drivers/md/dm-zoned-metadata.c
1404
zmd->sb[0].zone = zone;
drivers/md/dm-zoned-metadata.c
1519
zmd->sb[0].zone = dmz_get(zmd, 0);
drivers/md/dm-zoned-metadata.c
177
struct dmz_sb sb[2];
drivers/md/dm-zoned-metadata.c
2791
if (zmd->sb[i].mblk) {
drivers/md/dm-zoned-metadata.c
2792
dmz_free_mblock(zmd, zmd->sb[i].mblk);
drivers/md/dm-zoned-metadata.c
2793
zmd->sb[i].mblk = NULL;
drivers/md/dm-zoned-metadata.c
2907
zone = dmz_get(zmd, zmd->sb[0].zone->id + i);
drivers/md/dm-zoned-metadata.c
536
sector_t block = zmd->sb[zmd->mblk_primary].block + mblk_no;
drivers/md/dm-zoned-metadata.c
537
struct dmz_dev *dev = zmd->sb[zmd->mblk_primary].dev;
drivers/md/dm-zoned-metadata.c
668
struct dmz_dev *dev = zmd->sb[zmd->mblk_primary].dev;
drivers/md/dm-zoned-metadata.c
711
struct dmz_dev *dev = zmd->sb[set].dev;
drivers/md/dm-zoned-metadata.c
712
sector_t block = zmd->sb[set].block + mblk->no;
drivers/md/dm-zoned-metadata.c
764
struct dmz_mblock *mblk = zmd->sb[set].mblk;
drivers/md/dm-zoned-metadata.c
765
struct dmz_super *sb = zmd->sb[set].sb;
drivers/md/dm-zoned-metadata.c
766
struct dmz_dev *dev = zmd->sb[set].dev;
drivers/md/dm-zoned-metadata.c
771
sb->magic = cpu_to_le32(DMZ_MAGIC);
drivers/md/dm-zoned-metadata.c
773
sb->version = cpu_to_le32(zmd->sb_version);
drivers/md/dm-zoned-metadata.c
776
export_uuid(sb->dmz_uuid, &zmd->uuid);
drivers/md/dm-zoned-metadata.c
777
memcpy(sb->dmz_label, zmd->label, BDEVNAME_SIZE);
drivers/md/dm-zoned-metadata.c
778
export_uuid(sb->dev_uuid, &dev->uuid);
drivers/md/dm-zoned-metadata.c
781
sb->gen = cpu_to_le64(sb_gen);
drivers/md/dm-zoned-metadata.c
788
sb_block = zmd->sb[set].zone->id << zmd->zone_nr_blocks_shift;
drivers/md/dm-zoned-metadata.c
789
sb->sb_block = cpu_to_le64(sb_block);
drivers/md/dm-zoned-metadata.c
790
sb->nr_meta_blocks = cpu_to_le32(zmd->nr_meta_blocks);
drivers/md/dm-zoned-metadata.c
791
sb->nr_reserved_seq = cpu_to_le32(zmd->nr_reserved_seq);
drivers/md/dm-zoned-metadata.c
792
sb->nr_chunks = cpu_to_le32(zmd->nr_chunks);
drivers/md/dm-zoned-metadata.c
794
sb->nr_map_blocks = cpu_to_le32(zmd->nr_map_blocks);
drivers/md/dm-zoned-metadata.c
795
sb->nr_bitmap_blocks = cpu_to_le32(zmd->nr_bitmap_blocks);
drivers/md/dm-zoned-metadata.c
797
sb->crc = 0;
drivers/md/dm-zoned-metadata.c
798
sb->crc = cpu_to_le32(crc32_le(sb_gen, (unsigned char *)sb, DMZ_BLOCK_SIZE));
drivers/md/dm-zoned-metadata.c
800
ret = dmz_rdwr_block(dev, REQ_OP_WRITE, zmd->sb[set].block,
drivers/md/dm-zoned-metadata.c
816
struct dmz_dev *dev = zmd->sb[set].dev;
drivers/md/dm-zoned-metadata.c
897
dev = zmd->sb[zmd->mblk_primary].dev;
drivers/md/dm-zoned-metadata.c
977
struct dmz_super *sb = dsb->sb;
drivers/md/dm-zoned-metadata.c
983
if (le32_to_cpu(sb->magic) != DMZ_MAGIC) {
drivers/md/dm-zoned-metadata.c
985
DMZ_MAGIC, le32_to_cpu(sb->magic));
drivers/md/dm-zoned-metadata.c
989
zmd->sb_version = le32_to_cpu(sb->version);
drivers/md/md-bitmap.c
1550
bitmap_super_t *sb;
drivers/md/md-bitmap.c
1553
sb = kmap_local_page(bitmap->storage.sb_page);
drivers/md/md-bitmap.c
1554
sb->events_cleared =
drivers/md/md-bitmap.c
1556
kunmap_local(sb);
drivers/md/md-bitmap.c
2341
bitmap_super_t *sb;
drivers/md/md-bitmap.c
2347
sb = kmap_local_page(bitmap->storage.sb_page);
drivers/md/md-bitmap.c
2348
stats->sync_size = le64_to_cpu(sb->sync_size);
drivers/md/md-bitmap.c
2349
kunmap_local(sb);
drivers/md/md-bitmap.c
681
bitmap_super_t *sb;
drivers/md/md-bitmap.c
690
sb = kmap_local_page(bitmap->storage.sb_page);
drivers/md/md-bitmap.c
691
sb->events = cpu_to_le64(bitmap->mddev->events);
drivers/md/md-bitmap.c
695
sb->events_cleared = cpu_to_le64(bitmap->events_cleared);
drivers/md/md-bitmap.c
700
sb->state = cpu_to_le32(bitmap->flags & ~BIT(BITMAP_WRITE_ERROR));
drivers/md/md-bitmap.c
702
sb->daemon_sleep = cpu_to_le32(bitmap->mddev->bitmap_info.daemon_sleep/HZ);
drivers/md/md-bitmap.c
703
sb->write_behind = cpu_to_le32(bitmap->mddev->bitmap_info.max_write_behind);
drivers/md/md-bitmap.c
705
sb->sync_size = cpu_to_le64(bitmap->mddev->resync_max_sectors);
drivers/md/md-bitmap.c
706
sb->chunksize = cpu_to_le32(bitmap->mddev->bitmap_info.chunksize);
drivers/md/md-bitmap.c
707
sb->nodes = cpu_to_le32(bitmap->mddev->bitmap_info.nodes);
drivers/md/md-bitmap.c
708
sb->sectors_reserved = cpu_to_le32(bitmap->mddev->
drivers/md/md-bitmap.c
710
kunmap_local(sb);
drivers/md/md-bitmap.c
721
bitmap_super_t *sb;
drivers/md/md-bitmap.c
725
sb = kmap_local_page(bitmap->storage.sb_page);
drivers/md/md-bitmap.c
727
pr_debug(" magic: %08x\n", le32_to_cpu(sb->magic));
drivers/md/md-bitmap.c
728
pr_debug(" version: %u\n", le32_to_cpu(sb->version));
drivers/md/md-bitmap.c
730
le32_to_cpu(*(__le32 *)(sb->uuid+0)),
drivers/md/md-bitmap.c
731
le32_to_cpu(*(__le32 *)(sb->uuid+4)),
drivers/md/md-bitmap.c
732
le32_to_cpu(*(__le32 *)(sb->uuid+8)),
drivers/md/md-bitmap.c
733
le32_to_cpu(*(__le32 *)(sb->uuid+12)));
drivers/md/md-bitmap.c
735
(unsigned long long) le64_to_cpu(sb->events));
drivers/md/md-bitmap.c
737
(unsigned long long) le64_to_cpu(sb->events_cleared));
drivers/md/md-bitmap.c
738
pr_debug(" state: %08x\n", le32_to_cpu(sb->state));
drivers/md/md-bitmap.c
739
pr_debug(" chunksize: %u B\n", le32_to_cpu(sb->chunksize));
drivers/md/md-bitmap.c
740
pr_debug(" daemon sleep: %us\n", le32_to_cpu(sb->daemon_sleep));
drivers/md/md-bitmap.c
742
(unsigned long long)le64_to_cpu(sb->sync_size)/2);
drivers/md/md-bitmap.c
743
pr_debug("max write behind: %u\n", le32_to_cpu(sb->write_behind));
drivers/md/md-bitmap.c
744
kunmap_local(sb);
drivers/md/md-bitmap.c
760
bitmap_super_t *sb;
drivers/md/md-bitmap.c
768
sb = kmap_local_page(bitmap->storage.sb_page);
drivers/md/md-bitmap.c
770
sb->magic = cpu_to_le32(BITMAP_MAGIC);
drivers/md/md-bitmap.c
771
sb->version = cpu_to_le32(BITMAP_MAJOR_HI);
drivers/md/md-bitmap.c
776
kunmap_local(sb);
drivers/md/md-bitmap.c
780
sb->chunksize = cpu_to_le32(chunksize);
drivers/md/md-bitmap.c
787
sb->daemon_sleep = cpu_to_le32(daemon_sleep);
drivers/md/md-bitmap.c
797
sb->write_behind = cpu_to_le32(write_behind);
drivers/md/md-bitmap.c
801
sb->sync_size = cpu_to_le64(bitmap->mddev->resync_max_sectors);
drivers/md/md-bitmap.c
803
memcpy(sb->uuid, bitmap->mddev->uuid, 16);
drivers/md/md-bitmap.c
806
sb->state = cpu_to_le32(bitmap->flags);
drivers/md/md-bitmap.c
808
sb->events_cleared = cpu_to_le64(bitmap->mddev->events);
drivers/md/md-bitmap.c
811
kunmap_local(sb);
drivers/md/md-bitmap.c
820
bitmap_super_t *sb;
drivers/md/md-bitmap.c
873
sb = kmap_local_page(sb_page);
drivers/md/md-bitmap.c
875
chunksize = le32_to_cpu(sb->chunksize);
drivers/md/md-bitmap.c
876
daemon_sleep = le32_to_cpu(sb->daemon_sleep) * HZ;
drivers/md/md-bitmap.c
877
write_behind = le32_to_cpu(sb->write_behind);
drivers/md/md-bitmap.c
878
sectors_reserved = le32_to_cpu(sb->sectors_reserved);
drivers/md/md-bitmap.c
881
if (sb->magic != cpu_to_le32(BITMAP_MAGIC))
drivers/md/md-bitmap.c
883
else if (le32_to_cpu(sb->version) < BITMAP_MAJOR_LO ||
drivers/md/md-bitmap.c
884
le32_to_cpu(sb->version) > BITMAP_MAJOR_CLUSTERED)
drivers/md/md-bitmap.c
904
if (sb->version == cpu_to_le32(BITMAP_MAJOR_CLUSTERED)) {
drivers/md/md-bitmap.c
905
nodes = le32_to_cpu(sb->nodes);
drivers/md/md-bitmap.c
907
sb->cluster_name, 64);
drivers/md/md-bitmap.c
911
sb->sync_size = cpu_to_le64(bitmap->mddev->resync_max_sectors);
drivers/md/md-bitmap.c
918
if (memcmp(sb->uuid, bitmap->mddev->uuid, 16)) {
drivers/md/md-bitmap.c
923
events = le64_to_cpu(sb->events);
drivers/md/md-bitmap.c
933
bitmap->flags |= le32_to_cpu(sb->state);
drivers/md/md-bitmap.c
934
if (le32_to_cpu(sb->version) == BITMAP_MAJOR_HOSTENDIAN)
drivers/md/md-bitmap.c
936
bitmap->events_cleared = le64_to_cpu(sb->events_cleared);
drivers/md/md-bitmap.c
940
kunmap_local(sb);
drivers/md/md-cluster.c
1466
struct mdp_superblock_1 *sb = page_address(rdev->sb_page);
drivers/md/md-cluster.c
1467
char *uuid = sb->device_uuid;
drivers/md/md-cluster.c
456
struct mdp_superblock_1 *sb = NULL;
drivers/md/md-cluster.c
474
sb = page_address(rdev->sb_page);
drivers/md/md-cluster.c
499
if (sb && !(le32_to_cpu(sb->feature_map) & MD_FEATURE_RESHAPE_ACTIVE))
drivers/md/md-llbitmap.c
1359
bitmap_super_t *sb;
drivers/md/md-llbitmap.c
1375
sb = kmap_local_page(sb_page);
drivers/md/md-llbitmap.c
1376
sb->events = cpu_to_le64(mddev->events);
drivers/md/md-llbitmap.c
1377
sb->state = cpu_to_le32(llbitmap->flags);
drivers/md/md-llbitmap.c
1378
sb->chunksize = cpu_to_le32(llbitmap->chunksize);
drivers/md/md-llbitmap.c
1379
sb->sync_size = cpu_to_le64(mddev->resync_max_sectors);
drivers/md/md-llbitmap.c
1380
sb->events_cleared = cpu_to_le64(llbitmap->events_cleared);
drivers/md/md-llbitmap.c
1381
sb->sectors_reserved = cpu_to_le32(mddev->bitmap_info.space);
drivers/md/md-llbitmap.c
1382
sb->daemon_sleep = cpu_to_le32(mddev->bitmap_info.daemon_sleep);
drivers/md/md-llbitmap.c
1384
kunmap_local(sb);
drivers/md/md-llbitmap.c
814
bitmap_super_t *sb;
drivers/md/md-llbitmap.c
829
sb = kmap_local_page(sb_page);
drivers/md/md-llbitmap.c
830
if (sb->magic != cpu_to_le32(BITMAP_MAGIC)) {
drivers/md/md-llbitmap.c
836
if (sb->version != cpu_to_le32(BITMAP_MAJOR_LOCKLESS)) {
drivers/md/md-llbitmap.c
842
if (memcmp(sb->uuid, mddev->uuid, 16)) {
drivers/md/md-llbitmap.c
849
int room = le32_to_cpu(sb->sectors_reserved);
drivers/md/md-llbitmap.c
856
llbitmap->flags = le32_to_cpu(sb->state);
drivers/md/md-llbitmap.c
862
chunksize = le32_to_cpu(sb->chunksize);
drivers/md/md-llbitmap.c
877
daemon_sleep = le32_to_cpu(sb->daemon_sleep);
drivers/md/md-llbitmap.c
884
events = le64_to_cpu(sb->events);
drivers/md/md-llbitmap.c
891
sb->sync_size = cpu_to_le64(mddev->resync_max_sectors);
drivers/md/md-llbitmap.c
903
kunmap_local(sb);
drivers/md/md.c
10544
struct mdp_superblock_1 *sb = page_address(rdev->sb_page);
drivers/md/md.c
10552
if (mddev->dev_sectors != le64_to_cpu(sb->size)) {
drivers/md/md.c
10553
ret = mddev->pers->resize(mddev, le64_to_cpu(sb->size));
drivers/md/md.c
10569
role = le16_to_cpu(sb->dev_roles[rdev2->desc_nr]);
drivers/md/md.c
10587
!(le32_to_cpu(sb->feature_map) &
drivers/md/md.c
10595
if ((le32_to_cpu(sb->feature_map)
drivers/md/md.c
10621
if (mddev->raid_disks != le32_to_cpu(sb->raid_disks)) {
drivers/md/md.c
10622
ret = update_raid_disks(mddev, le32_to_cpu(sb->raid_disks));
drivers/md/md.c
10632
(le32_to_cpu(sb->feature_map) & MD_FEATURE_RESHAPE_ACTIVE)) {
drivers/md/md.c
10637
mddev->reshape_position = le64_to_cpu(sb->reshape_position);
drivers/md/md.c
10644
!(le32_to_cpu(sb->feature_map) & MD_FEATURE_RESHAPE_ACTIVE)) {
drivers/md/md.c
10652
mddev->events = le64_to_cpu(sb->events);
drivers/md/md.c
10659
struct mdp_superblock_1 *sb;
drivers/md/md.c
10682
sb = page_address(rdev->sb_page);
drivers/md/md.c
10687
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_RECOVERY_OFFSET))
drivers/md/md.c
10688
rdev->recovery_offset = le64_to_cpu(sb->recovery_offset);
drivers/md/md.c
1255
static unsigned int calc_sb_csum(mdp_super_t *sb)
drivers/md/md.c
1258
u32 *sb32 = (u32*)sb;
drivers/md/md.c
1262
disk_csum = sb->sb_csum;
drivers/md/md.c
1263
sb->sb_csum = 0;
drivers/md/md.c
1278
sb->sb_csum = md_csum_fold(disk_csum);
drivers/md/md.c
1280
sb->sb_csum = disk_csum;
drivers/md/md.c
1355
mdp_super_t *sb;
drivers/md/md.c
1373
sb = page_address(rdev->sb_page);
drivers/md/md.c
1375
if (sb->md_magic != MD_SB_MAGIC) {
drivers/md/md.c
1381
if (sb->major_version != 0 ||
drivers/md/md.c
1382
sb->minor_version < 90 ||
drivers/md/md.c
1383
sb->minor_version > 91) {
drivers/md/md.c
1385
sb->major_version, sb->minor_version, rdev->bdev);
drivers/md/md.c
1389
if (sb->raid_disks <= 0)
drivers/md/md.c
1392
if (md_csum_fold(calc_sb_csum(sb)) != md_csum_fold(sb->sb_csum)) {
drivers/md/md.c
1397
rdev->preferred_minor = sb->md_minor;
drivers/md/md.c
1403
rdev->desc_nr = sb->this_disk.number;
drivers/md/md.c
1407
sb->disks[rdev->desc_nr].state & ((1<<MD_DISK_SYNC) | (1 << MD_DISK_ACTIVE)))
drivers/md/md.c
1418
if (!md_uuid_equal(refsb, sb)) {
drivers/md/md.c
1423
if (!md_sb_equal(refsb, sb)) {
drivers/md/md.c
1428
ev1 = md_event(sb);
drivers/md/md.c
1441
if ((u64)rdev->sectors >= (2ULL << 32) && sb->level >= 1)
drivers/md/md.c
1444
if (rdev->sectors < ((sector_t)sb->size) * 2 && sb->level >= 1)
drivers/md/md.c
1474
mdp_super_t *sb = page_address(rdev->sb_page);
drivers/md/md.c
1475
__u64 ev1 = md_event(sb);
drivers/md/md.c
1485
mddev->minor_version = sb->minor_version;
drivers/md/md.c
1486
mddev->patch_version = sb->patch_version;
drivers/md/md.c
1488
mddev->chunk_sectors = sb->chunk_size >> 9;
drivers/md/md.c
1489
mddev->ctime = sb->ctime;
drivers/md/md.c
1490
mddev->utime = sb->utime;
drivers/md/md.c
1491
mddev->level = sb->level;
drivers/md/md.c
1493
mddev->layout = sb->layout;
drivers/md/md.c
1494
mddev->raid_disks = sb->raid_disks;
drivers/md/md.c
1495
mddev->dev_sectors = ((sector_t)sb->size) * 2;
drivers/md/md.c
1505
mddev->reshape_position = sb->reshape_position;
drivers/md/md.c
1506
mddev->delta_disks = sb->delta_disks;
drivers/md/md.c
1507
mddev->new_level = sb->new_level;
drivers/md/md.c
1508
mddev->new_layout = sb->new_layout;
drivers/md/md.c
1509
mddev->new_chunk_sectors = sb->new_chunk >> 9;
drivers/md/md.c
1522
if (sb->state & (1<<MD_SB_CLEAN))
drivers/md/md.c
1525
if (sb->events_hi == sb->cp_events_hi &&
drivers/md/md.c
1526
sb->events_lo == sb->cp_events_lo) {
drivers/md/md.c
1527
mddev->resync_offset = sb->recovery_cp;
drivers/md/md.c
1532
memcpy(mddev->uuid+0, &sb->set_uuid0, 4);
drivers/md/md.c
1533
memcpy(mddev->uuid+4, &sb->set_uuid1, 4);
drivers/md/md.c
1534
memcpy(mddev->uuid+8, &sb->set_uuid2, 4);
drivers/md/md.c
1535
memcpy(mddev->uuid+12,&sb->set_uuid3, 4);
drivers/md/md.c
1539
if (sb->state & (1<<MD_SB_BITMAP_PRESENT) &&
drivers/md/md.c
1551
if (sb->disks[rdev->desc_nr].state & (
drivers/md/md.c
1569
desc = sb->disks + rdev->desc_nr;
drivers/md/md.c
1599
mdp_super_t *sb;
drivers/md/md.c
1618
sb = page_address(rdev->sb_page);
drivers/md/md.c
1620
memset(sb, 0, sizeof(*sb));
drivers/md/md.c
1622
sb->md_magic = MD_SB_MAGIC;
drivers/md/md.c
1623
sb->major_version = mddev->major_version;
drivers/md/md.c
1624
sb->patch_version = mddev->patch_version;
drivers/md/md.c
1625
sb->gvalid_words = 0; /* ignored */
drivers/md/md.c
1626
memcpy(&sb->set_uuid0, mddev->uuid+0, 4);
drivers/md/md.c
1627
memcpy(&sb->set_uuid1, mddev->uuid+4, 4);
drivers/md/md.c
1628
memcpy(&sb->set_uuid2, mddev->uuid+8, 4);
drivers/md/md.c
1629
memcpy(&sb->set_uuid3, mddev->uuid+12,4);
drivers/md/md.c
1631
sb->ctime = clamp_t(time64_t, mddev->ctime, 0, U32_MAX);
drivers/md/md.c
1632
sb->level = mddev->level;
drivers/md/md.c
1633
sb->size = mddev->dev_sectors / 2;
drivers/md/md.c
1634
sb->raid_disks = mddev->raid_disks;
drivers/md/md.c
1635
sb->md_minor = mddev->md_minor;
drivers/md/md.c
1636
sb->not_persistent = 0;
drivers/md/md.c
1637
sb->utime = clamp_t(time64_t, mddev->utime, 0, U32_MAX);
drivers/md/md.c
1638
sb->state = 0;
drivers/md/md.c
1639
sb->events_hi = (mddev->events>>32);
drivers/md/md.c
1640
sb->events_lo = (u32)mddev->events;
drivers/md/md.c
1643
sb->minor_version = 90;
drivers/md/md.c
1645
sb->minor_version = 91;
drivers/md/md.c
1646
sb->reshape_position = mddev->reshape_position;
drivers/md/md.c
1647
sb->new_level = mddev->new_level;
drivers/md/md.c
1648
sb->delta_disks = mddev->delta_disks;
drivers/md/md.c
1649
sb->new_layout = mddev->new_layout;
drivers/md/md.c
1650
sb->new_chunk = mddev->new_chunk_sectors << 9;
drivers/md/md.c
1652
mddev->minor_version = sb->minor_version;
drivers/md/md.c
1655
sb->recovery_cp = mddev->resync_offset;
drivers/md/md.c
1656
sb->cp_events_hi = (mddev->events>>32);
drivers/md/md.c
1657
sb->cp_events_lo = (u32)mddev->events;
drivers/md/md.c
1659
sb->state = (1<< MD_SB_CLEAN);
drivers/md/md.c
1661
sb->recovery_cp = 0;
drivers/md/md.c
1663
sb->layout = mddev->layout;
drivers/md/md.c
1664
sb->chunk_size = mddev->chunk_sectors << 9;
drivers/md/md.c
1667
sb->state |= (1<<MD_SB_BITMAP_PRESENT);
drivers/md/md.c
1669
sb->disks[0].state = (1<<MD_DISK_REMOVED);
drivers/md/md.c
1676
sb->minor_version >= 91)
drivers/md/md.c
1690
d = &sb->disks[rdev2->desc_nr];
drivers/md/md.c
1719
mdp_disk_t *d = &sb->disks[i];
drivers/md/md.c
1728
sb->nr_disks = nr_disks;
drivers/md/md.c
1729
sb->active_disks = active;
drivers/md/md.c
1730
sb->working_disks = working;
drivers/md/md.c
1731
sb->failed_disks = failed;
drivers/md/md.c
1732
sb->spare_disks = spare;
drivers/md/md.c
1734
sb->this_disk = sb->disks[rdev->desc_nr];
drivers/md/md.c
1735
sb->sb_csum = calc_sb_csum(sb);
drivers/md/md.c
1774
static __le32 calc_sb_1_csum(struct mdp_superblock_1 *sb)
drivers/md/md.c
1779
int size = 256 + le32_to_cpu(sb->max_dev)*2;
drivers/md/md.c
1780
__le32 *isuper = (__le32*)sb;
drivers/md/md.c
1782
disk_csum = sb->sb_csum;
drivers/md/md.c
1783
sb->sb_csum = 0;
drivers/md/md.c
1792
sb->sb_csum = disk_csum;
drivers/md/md.c
1798
struct mdp_superblock_1 *sb;
drivers/md/md.c
1835
sb = page_address(rdev->sb_page);
drivers/md/md.c
1837
if (sb->magic != cpu_to_le32(MD_SB_MAGIC) ||
drivers/md/md.c
1838
sb->major_version != cpu_to_le32(1) ||
drivers/md/md.c
1839
le32_to_cpu(sb->max_dev) > (4096-256)/2 ||
drivers/md/md.c
1840
le64_to_cpu(sb->super_offset) != rdev->sb_start ||
drivers/md/md.c
1841
(le32_to_cpu(sb->feature_map) & ~MD_FEATURE_ALL) != 0)
drivers/md/md.c
1844
if (calc_sb_1_csum(sb) != sb->sb_csum) {
drivers/md/md.c
1849
if (le64_to_cpu(sb->data_size) < 10) {
drivers/md/md.c
1854
if (sb->pad0 ||
drivers/md/md.c
1855
sb->pad3[0] ||
drivers/md/md.c
1856
memcmp(sb->pad3, sb->pad3+1, sizeof(sb->pad3) - sizeof(sb->pad3[1]))) {
drivers/md/md.c
1865
rdev->data_offset = le64_to_cpu(sb->data_offset);
drivers/md/md.c
1867
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_RESHAPE_ACTIVE) &&
drivers/md/md.c
1868
(le32_to_cpu(sb->feature_map) & MD_FEATURE_NEW_OFFSET))
drivers/md/md.c
1869
rdev->new_data_offset += (s32)le32_to_cpu(sb->new_offset);
drivers/md/md.c
1870
atomic_set(&rdev->corrected_errors, le32_to_cpu(sb->cnt_corrected_read));
drivers/md/md.c
1872
rdev->sb_size = le32_to_cpu(sb->max_dev) * 2 + 256;
drivers/md/md.c
1884
rdev->desc_nr = le32_to_cpu(sb->dev_number);
drivers/md/md.c
1891
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_BAD_BLOCKS) &&
drivers/md/md.c
1900
int sectors = le16_to_cpu(sb->bblog_size);
drivers/md/md.c
1903
offset = le32_to_cpu(sb->bblog_offset);
drivers/md/md.c
1911
rdev->badblocks.shift = sb->bblog_shift;
drivers/md/md.c
1916
sector <<= sb->bblog_shift;
drivers/md/md.c
1917
count <<= sb->bblog_shift;
drivers/md/md.c
1923
} else if (sb->bblog_offset != 0)
drivers/md/md.c
1926
if ((le32_to_cpu(sb->feature_map) &
drivers/md/md.c
1928
rdev->ppl.offset = (__s16)le16_to_cpu(sb->ppl.offset);
drivers/md/md.c
1929
rdev->ppl.size = le16_to_cpu(sb->ppl.size);
drivers/md/md.c
1933
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_RAID0_LAYOUT) &&
drivers/md/md.c
1934
sb->level != 0)
drivers/md/md.c
1938
if (rdev->desc_nr >= 0 && rdev->desc_nr < le32_to_cpu(sb->max_dev) &&
drivers/md/md.c
1939
(le16_to_cpu(sb->dev_roles[rdev->desc_nr]) < MD_DISK_ROLE_MAX ||
drivers/md/md.c
1940
le16_to_cpu(sb->dev_roles[rdev->desc_nr]) == MD_DISK_ROLE_JOURNAL))
drivers/md/md.c
1952
if (memcmp(sb->set_uuid, refsb->set_uuid, 16) != 0 ||
drivers/md/md.c
1953
sb->level != refsb->level ||
drivers/md/md.c
1954
sb->layout != refsb->layout ||
drivers/md/md.c
1955
sb->chunksize != refsb->chunksize) {
drivers/md/md.c
1961
ev1 = le64_to_cpu(sb->events);
drivers/md/md.c
1973
if (sectors < le64_to_cpu(sb->data_size))
drivers/md/md.c
1975
rdev->sectors = le64_to_cpu(sb->data_size);
drivers/md/md.c
1981
struct mdp_superblock_1 *sb = page_address(rdev->sb_page);
drivers/md/md.c
1982
__u64 ev1 = le64_to_cpu(sb->events);
drivers/md/md.c
1995
mddev->chunk_sectors = le32_to_cpu(sb->chunksize);
drivers/md/md.c
1996
mddev->ctime = le64_to_cpu(sb->ctime);
drivers/md/md.c
1997
mddev->utime = le64_to_cpu(sb->utime);
drivers/md/md.c
1998
mddev->level = le32_to_cpu(sb->level);
drivers/md/md.c
2000
mddev->layout = le32_to_cpu(sb->layout);
drivers/md/md.c
2001
mddev->raid_disks = le32_to_cpu(sb->raid_disks);
drivers/md/md.c
2002
mddev->dev_sectors = le64_to_cpu(sb->size);
drivers/md/md.c
2013
mddev->resync_offset = le64_to_cpu(sb->resync_offset);
drivers/md/md.c
2014
memcpy(mddev->uuid, sb->set_uuid, 16);
drivers/md/md.c
2019
mddev->logical_block_size = le32_to_cpu(sb->logical_block_size);
drivers/md/md.c
2021
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_BITMAP_OFFSET) &&
drivers/md/md.c
2024
(__s32)le32_to_cpu(sb->bitmap_offset);
drivers/md/md.c
2040
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_RESHAPE_ACTIVE)) {
drivers/md/md.c
2041
mddev->reshape_position = le64_to_cpu(sb->reshape_position);
drivers/md/md.c
2042
mddev->delta_disks = le32_to_cpu(sb->delta_disks);
drivers/md/md.c
2043
mddev->new_level = le32_to_cpu(sb->new_level);
drivers/md/md.c
2044
mddev->new_layout = le32_to_cpu(sb->new_layout);
drivers/md/md.c
2045
mddev->new_chunk_sectors = le32_to_cpu(sb->new_chunk);
drivers/md/md.c
2048
(le32_to_cpu(sb->feature_map)
drivers/md/md.c
2060
!(le32_to_cpu(sb->feature_map) & MD_FEATURE_RAID0_LAYOUT))
drivers/md/md.c
2063
if (le32_to_cpu(sb->feature_map) & MD_FEATURE_JOURNAL)
drivers/md/md.c
2066
if (le32_to_cpu(sb->feature_map) &
drivers/md/md.c
2068
if (le32_to_cpu(sb->feature_map) &
drivers/md/md.c
2071
if ((le32_to_cpu(sb->feature_map) & MD_FEATURE_PPL) &&
drivers/md/md.c
2072
(le32_to_cpu(sb->feature_map) &
drivers/md/md.c
2084
rdev->desc_nr < le32_to_cpu(sb->max_dev) &&
drivers/md/md.c
2085
(le16_to_cpu(sb->dev_roles[rdev->desc_nr]) < MD_DISK_ROLE_MAX ||
drivers/md/md.c
2086
le16_to_cpu(sb->dev_roles[rdev->desc_nr]) == MD_DISK_ROLE_JOURNAL))
drivers/md/md.c
2104
rdev->desc_nr >= le32_to_cpu(sb->max_dev)) {
drivers/md/md.c
2137
role = le16_to_cpu(sb->dev_roles[rdev->desc_nr]);
drivers/md/md.c
2146
if (!(le32_to_cpu(sb->feature_map) & MD_FEATURE_JOURNAL)) {
drivers/md/md.c
2152
rdev->journal_tail = le64_to_cpu(sb->journal_tail);
drivers/md/md.c
2157
if ((le32_to_cpu(sb->feature_map) &
drivers/md/md.c
2159
rdev->recovery_offset = le64_to_cpu(sb->recovery_offset);
drivers/md/md.c
2160
if (!(le32_to_cpu(sb->feature_map) &
drivers/md/md.c
2175
if (sb->devflags & WriteMostly1)
drivers/md/md.c
2177
if (sb->devflags & FailFast1)
drivers/md/md.c
2179
if (le32_to_cpu(sb->feature_map) & MD_FEATURE_REPLACEMENT)
drivers/md/md.c
2187
struct mdp_superblock_1 *sb;
drivers/md/md.c
2192
sb = page_address(rdev->sb_page);
drivers/md/md.c
2194
sb->feature_map = 0;
drivers/md/md.c
2195
sb->pad0 = 0;
drivers/md/md.c
2196
sb->recovery_offset = cpu_to_le64(0);
drivers/md/md.c
2197
memset(sb->pad3, 0, sizeof(sb->pad3));
drivers/md/md.c
2199
sb->utime = cpu_to_le64((__u64)mddev->utime);
drivers/md/md.c
2200
sb->events = cpu_to_le64(mddev->events);
drivers/md/md.c
2202
sb->resync_offset = cpu_to_le64(mddev->resync_offset);
drivers/md/md.c
2204
sb->resync_offset = cpu_to_le64(MaxSector);
drivers/md/md.c
2206
sb->resync_offset = cpu_to_le64(0);
drivers/md/md.c
2208
sb->cnt_corrected_read = cpu_to_le32(atomic_read(&rdev->corrected_errors));
drivers/md/md.c
2210
sb->raid_disks = cpu_to_le32(mddev->raid_disks);
drivers/md/md.c
2211
sb->size = cpu_to_le64(mddev->dev_sectors);
drivers/md/md.c
2212
sb->chunksize = cpu_to_le32(mddev->chunk_sectors);
drivers/md/md.c
2213
sb->level = cpu_to_le32(mddev->level);
drivers/md/md.c
2214
sb->layout = cpu_to_le32(mddev->layout);
drivers/md/md.c
2215
sb->logical_block_size = cpu_to_le32(mddev->logical_block_size);
drivers/md/md.c
2217
sb->devflags |= FailFast1;
drivers/md/md.c
2219
sb->devflags &= ~FailFast1;
drivers/md/md.c
2222
sb->devflags |= WriteMostly1;
drivers/md/md.c
2224
sb->devflags &= ~WriteMostly1;
drivers/md/md.c
2225
sb->data_offset = cpu_to_le64(rdev->data_offset);
drivers/md/md.c
2226
sb->data_size = cpu_to_le64(rdev->sectors);
drivers/md/md.c
2229
sb->bitmap_offset = cpu_to_le32((__u32)mddev->bitmap_info.offset);
drivers/md/md.c
2230
sb->feature_map = cpu_to_le32(MD_FEATURE_BITMAP_OFFSET);
drivers/md/md.c
2235
sb->feature_map |=
drivers/md/md.c
2237
sb->recovery_offset =
drivers/md/md.c
2240
sb->feature_map |=
drivers/md/md.c
2245
sb->journal_tail = cpu_to_le64(rdev->journal_tail);
drivers/md/md.c
2247
sb->feature_map |=
drivers/md/md.c
2251
sb->feature_map |= cpu_to_le32(MD_FEATURE_RESHAPE_ACTIVE);
drivers/md/md.c
2252
sb->reshape_position = cpu_to_le64(mddev->reshape_position);
drivers/md/md.c
2253
sb->new_layout = cpu_to_le32(mddev->new_layout);
drivers/md/md.c
2254
sb->delta_disks = cpu_to_le32(mddev->delta_disks);
drivers/md/md.c
2255
sb->new_level = cpu_to_le32(mddev->new_level);
drivers/md/md.c
2256
sb->new_chunk = cpu_to_le32(mddev->new_chunk_sectors);
drivers/md/md.c
2259
sb->feature_map
drivers/md/md.c
2262
sb->feature_map
drivers/md/md.c
2264
sb->new_offset = cpu_to_le32((__u32)(rdev->new_data_offset
drivers/md/md.c
2270
sb->feature_map |= cpu_to_le32(MD_FEATURE_CLUSTERED);
drivers/md/md.c
2274
else if (sb->bblog_offset == 0)
drivers/md/md.c
2281
sb->feature_map |= cpu_to_le32(MD_FEATURE_BAD_BLOCKS);
drivers/md/md.c
2301
(int)le32_to_cpu(sb->bblog_offset));
drivers/md/md.c
2302
bb->size = le16_to_cpu(sb->bblog_size);
drivers/md/md.c
2311
if (max_dev > le32_to_cpu(sb->max_dev)) {
drivers/md/md.c
2313
sb->max_dev = cpu_to_le32(max_dev);
drivers/md/md.c
2319
max_dev = le32_to_cpu(sb->max_dev);
drivers/md/md.c
2322
sb->dev_roles[i] = cpu_to_le16(MD_DISK_ROLE_SPARE);
drivers/md/md.c
2325
sb->feature_map |= cpu_to_le32(MD_FEATURE_JOURNAL);
drivers/md/md.c
2329
sb->feature_map |=
drivers/md/md.c
2332
sb->feature_map |= cpu_to_le32(MD_FEATURE_PPL);
drivers/md/md.c
2333
sb->ppl.offset = cpu_to_le16(rdev->ppl.offset);
drivers/md/md.c
2334
sb->ppl.size = cpu_to_le16(rdev->ppl.size);
drivers/md/md.c
2340
sb->dev_roles[i] = cpu_to_le16(MD_DISK_ROLE_FAULTY);
drivers/md/md.c
2342
sb->dev_roles[i] = cpu_to_le16(rdev2->raid_disk);
drivers/md/md.c
2344
sb->dev_roles[i] = cpu_to_le16(MD_DISK_ROLE_JOURNAL);
drivers/md/md.c
2346
sb->dev_roles[i] = cpu_to_le16(rdev2->raid_disk);
drivers/md/md.c
2348
sb->dev_roles[i] = cpu_to_le16(MD_DISK_ROLE_SPARE);
drivers/md/md.c
2351
sb->sb_csum = calc_sb_1_csum(sb);
drivers/md/md.c
2375
struct mdp_superblock_1 *sb;
drivers/md/md.c
2409
sb = page_address(rdev->sb_page);
drivers/md/md.c
2410
sb->data_size = cpu_to_le64(num_sectors);
drivers/md/md.c
2411
sb->super_offset = cpu_to_le64(rdev->sb_start);
drivers/md/md.c
2412
sb->sb_csum = calc_sb_1_csum(sb);
drivers/md/md.c
2742
struct mdp_superblock_1 *sb;
drivers/md/md.c
2756
sb = page_address(rdev->sb_page);
drivers/md/md.c
2759
role = le16_to_cpu(sb->dev_roles[rdev->desc_nr]);
drivers/md/md.c
2770
if ((mddev->dev_sectors != le64_to_cpu(sb->size)) ||
drivers/md/md.c
2771
(mddev->reshape_position != le64_to_cpu(sb->reshape_position)) ||
drivers/md/md.c
2772
(mddev->layout != le32_to_cpu(sb->layout)) ||
drivers/md/md.c
2773
(mddev->raid_disks != le32_to_cpu(sb->raid_disks)) ||
drivers/md/md.c
2774
(mddev->chunk_sectors != le32_to_cpu(sb->chunksize)))
drivers/md/raid10.c
4430
struct mdp_superblock_1 *sb = NULL;
drivers/md/raid10.c
4448
sb = page_address(rdev->sb_page);
drivers/md/raid10.c
4456
if ((sb && (le32_to_cpu(sb->feature_map) &
drivers/md/raid10.c
4739
struct mdp_superblock_1 *sb = NULL;
drivers/md/raid10.c
4744
sb = page_address(rdev->sb_page);
drivers/md/raid10.c
4745
if (sb) {
drivers/md/raid10.c
4746
sb_reshape_pos = le64_to_cpu(sb->reshape_position);
drivers/misc/ibmasm/ibmasmfs.c
108
static int ibmasmfs_fill_super(struct super_block *sb, struct fs_context *fc)
drivers/misc/ibmasm/ibmasmfs.c
112
sb->s_blocksize = PAGE_SIZE;
drivers/misc/ibmasm/ibmasmfs.c
113
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/misc/ibmasm/ibmasmfs.c
114
sb->s_magic = IBMASMFS_MAGIC;
drivers/misc/ibmasm/ibmasmfs.c
115
sb->s_op = &ibmasmfs_s_ops;
drivers/misc/ibmasm/ibmasmfs.c
116
sb->s_time_gran = 1;
drivers/misc/ibmasm/ibmasmfs.c
118
root = ibmasmfs_make_inode (sb, S_IFDIR | 0500);
drivers/misc/ibmasm/ibmasmfs.c
125
sb->s_root = d_make_root(root);
drivers/misc/ibmasm/ibmasmfs.c
126
if (!sb->s_root)
drivers/misc/ibmasm/ibmasmfs.c
129
ibmasmfs_create_files(sb);
drivers/misc/ibmasm/ibmasmfs.c
133
static struct inode *ibmasmfs_make_inode(struct super_block *sb, int mode)
drivers/misc/ibmasm/ibmasmfs.c
135
struct inode *ret = new_inode(sb);
drivers/misc/ibmasm/ibmasmfs.c
572
static void ibmasmfs_create_files (struct super_block *sb)
drivers/misc/ibmasm/ibmasmfs.c
581
dir = ibmasmfs_create_dir(sb->s_root, sp->dirname);
drivers/misc/ibmasm/ibmasmfs.c
76
static struct inode *ibmasmfs_make_inode(struct super_block *sb, int mode);
drivers/misc/ibmasm/ibmasmfs.c
77
static void ibmasmfs_create_files (struct super_block *sb);
drivers/misc/ibmasm/ibmasmfs.c
78
static int ibmasmfs_fill_super(struct super_block *sb, struct fs_context *fc);
drivers/mtd/mtdsuper.c
172
void kill_mtd_super(struct super_block *sb)
drivers/mtd/mtdsuper.c
174
generic_shutdown_super(sb);
drivers/mtd/mtdsuper.c
175
put_mtd_device(sb->s_mtd);
drivers/mtd/mtdsuper.c
176
sb->s_mtd = NULL;
drivers/mtd/mtdsuper.c
30
struct super_block *sb;
drivers/mtd/mtdsuper.c
33
sb = sget_dev(fc, MKDEV(MTD_BLOCK_MAJOR, mtd->index));
drivers/mtd/mtdsuper.c
34
if (IS_ERR(sb))
drivers/mtd/mtdsuper.c
35
return PTR_ERR(sb);
drivers/mtd/mtdsuper.c
37
if (sb->s_root) {
drivers/mtd/mtdsuper.c
54
sb->s_mtd = mtd;
drivers/mtd/mtdsuper.c
55
sb->s_bdi = bdi_get(mtd_bdi);
drivers/mtd/mtdsuper.c
57
ret = fill_super(sb, fc);
drivers/mtd/mtdsuper.c
61
sb->s_flags |= SB_ACTIVE;
drivers/mtd/mtdsuper.c
65
fc->root = dget(sb->s_root);
drivers/mtd/mtdsuper.c
69
deactivate_locked_super(sb);
drivers/mtd/mtdsuper.c
97
int (*fill_super)(struct super_block *sb,
drivers/mtd/nand/raw/marvell_nand.c
271
#define MARVELL_LAYOUT(ws, dc, ds, nc, fcc, db, sb, eb, ldb, lsb, leb) \
drivers/mtd/nand/raw/marvell_nand.c
279
.spare_bytes = sb, \
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4366
union host_hc_status_block *sb = &bnx2x_fp(bp, fp_index, status_blk);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4373
memset(sb, 0, sizeof(union host_hc_status_block));
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4378
BNX2X_PCI_FREE(sb->e2_sb,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4383
BNX2X_PCI_FREE(sb->e1x_sb,
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4449
(__le16 *)status_blk.e2_sb->sb.index_values;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4451
(__le16 *)status_blk.e2_sb->sb.running_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4454
(__le16 *)status_blk.e1x_sb->sb.index_values;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4456
(__le16 *)status_blk.e1x_sb->sb.running_index;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4518
union host_hc_status_block *sb;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4552
sb = &bnx2x_fp(bp, index, status_blk);
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4557
sb->e2_sb = BNX2X_PCI_ALLOC(&bnx2x_fp(bp, index, status_blk_mapping),
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4559
if (!sb->e2_sb)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4562
sb->e1x_sb = BNX2X_PCI_ALLOC(&bnx2x_fp(bp, index, status_blk_mapping),
drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c
4564
if (!sb->e1x_sb)
drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
3211
struct hc_status_block_e1x sb;
drivers/net/ethernet/broadcom/bnx2x/bnx2x_hsi.h
3228
struct hc_status_block_e2 sb;
drivers/net/ethernet/broadcom/cnic.c
4800
u32 sb = BNX2_L2CTX_L5_STATUSB_NUM(sb_id);
drivers/net/ethernet/broadcom/cnic.c
4807
cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_HOST_QIDX, sb);
drivers/net/ethernet/broadcom/cnic.c
4808
cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_HOST_QIDX, sb);
drivers/net/ethernet/broadcom/cnic.c
4925
struct host_sp_status_block *sb = cp->bnx2x_def_status_blk;
drivers/net/ethernet/broadcom/cnic.c
4980
&sb->sp_sb.index_values[HC_SP_INDEX_ETH_ISCSI_CQ_CONS];
drivers/net/ethernet/broadcom/cnic.c
4993
struct host_sp_status_block *sb = cp->bnx2x_def_status_blk;
drivers/net/ethernet/broadcom/cnic.c
5047
&sb->sp_sb.index_values[HC_SP_INDEX_ETH_ISCSI_RX_CQ_CONS];
drivers/net/ethernet/broadcom/cnic.c
5062
struct host_hc_status_block_e2 *sb = cp->status_blk.gen;
drivers/net/ethernet/broadcom/cnic.c
5065
&sb->sb.index_values[HC_INDEX_ISCSI_EQ_CONS];
drivers/net/ethernet/broadcom/cnic.c
5067
&sb->sb.running_index[SM_RX_ID];
drivers/net/ethernet/broadcom/cnic.c
5069
struct host_hc_status_block_e1x *sb = cp->status_blk.gen;
drivers/net/ethernet/broadcom/cnic.c
5072
&sb->sb.index_values[HC_INDEX_ISCSI_EQ_CONS];
drivers/net/ethernet/broadcom/cnic.c
5074
&sb->sb.running_index[SM_RX_ID];
drivers/net/ethernet/broadcom/cnic.c
5078
struct host_hc_status_block_e2 *sb = cp->status_blk.gen;
drivers/net/ethernet/broadcom/cnic.c
5084
&sb->sb.index_values[HC_INDEX_FCOE_EQ_CONS];
drivers/net/ethernet/broadcom/cnic.c
5086
&sb->sb.running_index[SM_RX_ID];
drivers/net/ethernet/broadcom/sb1250-mac.c
1057
struct sk_buff *sb;
drivers/net/ethernet/broadcom/sb1250-mac.c
1105
sb = d->sbdma_ctxtable[curidx];
drivers/net/ethernet/broadcom/sb1250-mac.c
1128
sbdma_add_rcvbuffer(sc, d, sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
1137
skb_put(sb,len);
drivers/net/ethernet/broadcom/sb1250-mac.c
1144
sb->protocol = eth_type_trans(sb,d->sbdma_eth->sbm_dev);
drivers/net/ethernet/broadcom/sb1250-mac.c
1149
sb->ip_summed = CHECKSUM_UNNECESSARY;
drivers/net/ethernet/broadcom/sb1250-mac.c
1152
skb_checksum_none_assert(sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
1155
prefetch(sb->data);
drivers/net/ethernet/broadcom/sb1250-mac.c
1156
prefetch((const void *)(((char *)sb->data)+32));
drivers/net/ethernet/broadcom/sb1250-mac.c
1158
dropped = netif_receive_skb(sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
1160
dropped = netif_rx(sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
1178
sbdma_add_rcvbuffer(sc, d, sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
1222
struct sk_buff *sb;
drivers/net/ethernet/broadcom/sb1250-mac.c
1262
sb = d->sbdma_ctxtable[curidx];
drivers/net/ethernet/broadcom/sb1250-mac.c
1269
dev->stats.tx_bytes += sb->len;
drivers/net/ethernet/broadcom/sb1250-mac.c
1276
dev_consume_skb_irq(sb);
drivers/net/ethernet/broadcom/sb1250-mac.c
770
struct sk_buff *sb)
drivers/net/ethernet/broadcom/sb1250-mac.c
812
if (sb == NULL) {
drivers/net/ethernet/broadcom/sb1250-mac.c
822
sb_new = sb;
drivers/net/ethernet/broadcom/sb1250-mac.c
885
static int sbdma_add_txbuffer(struct sbmacdma *d, struct sk_buff *sb)
drivers/net/ethernet/broadcom/sb1250-mac.c
914
length = sb->len;
drivers/net/ethernet/broadcom/sb1250-mac.c
923
phys = virt_to_phys(sb->data);
drivers/net/ethernet/broadcom/sb1250-mac.c
942
d->sbdma_ctxtable[dsc-d->sbdma_dscrtable] = sb;
drivers/net/ethernet/broadcom/sb1250-mac.c
977
struct sk_buff *sb;
drivers/net/ethernet/broadcom/sb1250-mac.c
980
sb = d->sbdma_ctxtable[idx];
drivers/net/ethernet/broadcom/sb1250-mac.c
981
if (sb) {
drivers/net/ethernet/broadcom/sb1250-mac.c
982
dev_kfree_skb(sb);
drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h
174
u32 sb;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2168
tid->sb = t4_read_reg(padap, LE_DB_SERVER_INDEX_A) / 4;
drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c
2172
tid->sb = t4_read_reg(padap, LE_DB_SRVR_START_INDEX_A);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3321
unsigned int sb;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3326
sb = t4_read_reg(adap, LE_DB_SERVER_INDEX_A) / 4;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3328
sb = t4_read_reg(adap, LE_DB_SRVR_START_INDEX_A);
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3330
if (sb) {
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
3332
sb - 1, adap->tids.hash_base,
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1034
unsigned int sb;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1045
sb = adapter->tids.stid_base;
drivers/net/ethernet/chelsio/cxgb4/cxgb4_filter.c
1046
for (i = 0; i < sb; i++) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
412
struct mlx5dr_ste_build *sb;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
417
sb = nic_matcher->ste_builder_arr[outer_ipv][inner_ipv];
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
466
mlx5dr_ste_build_general_purpose(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
470
mlx5dr_ste_build_register_0(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
474
mlx5dr_ste_build_register_1(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
480
mlx5dr_ste_build_src_gvmi_qpn(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
486
mlx5dr_ste_build_eth_l2_src_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
491
mlx5dr_ste_build_eth_l2_src(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
495
mlx5dr_ste_build_eth_l2_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
500
mlx5dr_ste_build_eth_l3_ipv6_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
504
mlx5dr_ste_build_eth_l3_ipv6_src(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
508
mlx5dr_ste_build_eth_ipv6_l3_l4(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
512
mlx5dr_ste_build_eth_l3_ipv4_5_tuple(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
517
mlx5dr_ste_build_eth_l3_ipv4_misc(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
522
mlx5dr_ste_build_tnl_vxlan_gpe(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
525
mlx5dr_ste_build_tnl_geneve(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
528
mlx5dr_ste_build_tnl_geneve_tlv_opt(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
532
mlx5dr_ste_build_tnl_geneve_tlv_opt_exist(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
537
mlx5dr_ste_build_tnl_gtpu_flex_parser_0(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
542
mlx5dr_ste_build_tnl_gtpu_flex_parser_1(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
547
mlx5dr_ste_build_tnl_gtpu(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
550
mlx5dr_ste_build_tnl_header_0_1(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
555
mlx5dr_ste_build_eth_l4_misc(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
559
mlx5dr_ste_build_mpls(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
563
mlx5dr_ste_build_tnl_mpls_over_gre(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
567
mlx5dr_ste_build_tnl_mpls_over_udp(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
572
mlx5dr_ste_build_icmp(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
577
mlx5dr_ste_build_tnl_gre(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
589
mlx5dr_ste_build_eth_l2_tnl(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
594
mlx5dr_ste_build_eth_l2_src_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
599
mlx5dr_ste_build_eth_l2_src(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
603
mlx5dr_ste_build_eth_l2_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
608
mlx5dr_ste_build_eth_l3_ipv6_dst(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
612
mlx5dr_ste_build_eth_l3_ipv6_src(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
616
mlx5dr_ste_build_eth_ipv6_l3_l4(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
620
mlx5dr_ste_build_eth_l3_ipv4_5_tuple(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
625
mlx5dr_ste_build_eth_l3_ipv4_misc(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
630
mlx5dr_ste_build_eth_l4_misc(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
634
mlx5dr_ste_build_mpls(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
638
mlx5dr_ste_build_tnl_mpls_over_gre(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
642
mlx5dr_ste_build_tnl_mpls_over_udp(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
649
mlx5dr_ste_build_flex_parser_0(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
653
mlx5dr_ste_build_flex_parser_1(ste_ctx, &sb[idx++],
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
660
mlx5dr_ste_build_empty_always_hit(&sb[idx++], rx);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_matcher.c
675
nic_matcher->ste_builder = sb;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1128
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1132
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1133
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1134
ste_ctx->build_eth_l2_src_dst_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1138
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1142
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1143
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1144
ste_ctx->build_eth_l3_ipv6_dst_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1148
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1152
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1153
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1154
ste_ctx->build_eth_l3_ipv6_src_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1158
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1162
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1163
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1164
ste_ctx->build_eth_l3_ipv4_5_tuple_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1168
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1172
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1173
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1174
ste_ctx->build_eth_l2_src_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1178
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1182
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1183
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1184
ste_ctx->build_eth_l2_dst_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1188
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1191
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1192
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1193
ste_ctx->build_eth_l2_tnl_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1197
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1201
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1202
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1203
ste_ctx->build_eth_l3_ipv4_misc_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1207
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1211
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1212
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1213
ste_ctx->build_eth_ipv6_l3_l4_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1217
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1223
void mlx5dr_ste_build_empty_always_hit(struct mlx5dr_ste_build *sb, bool rx)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1225
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1226
sb->lu_type = MLX5DR_STE_LU_TYPE_DONT_CARE;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1227
sb->byte_mask = 0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1228
sb->ste_build_tag_func = &dr_ste_build_empty_always_hit_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1232
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1236
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1237
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1238
ste_ctx->build_mpls_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1242
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1246
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1247
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1248
ste_ctx->build_tnl_gre_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1252
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1257
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1258
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1259
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1260
return ste_ctx->build_tnl_mpls_over_gre_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1264
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1269
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1270
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1271
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1272
return ste_ctx->build_tnl_mpls_over_udp_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1276
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1281
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1282
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1283
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1284
ste_ctx->build_icmp_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1288
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1292
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1293
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1294
ste_ctx->build_general_purpose_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1298
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1302
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1303
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1304
ste_ctx->build_eth_l4_misc_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1308
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1312
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1313
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1314
ste_ctx->build_tnl_vxlan_gpe_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1318
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1322
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1323
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1324
ste_ctx->build_tnl_geneve_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1328
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1333
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1334
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1335
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1336
ste_ctx->build_tnl_geneve_tlv_opt_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1340
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1348
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1349
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1350
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1351
ste_ctx->build_tnl_geneve_tlv_opt_exist_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1355
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1359
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1360
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1361
ste_ctx->build_tnl_gtpu_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1365
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1370
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1371
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1372
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1373
ste_ctx->build_tnl_gtpu_flex_parser_0_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1377
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1382
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1383
sb->caps = caps;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1384
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1385
ste_ctx->build_tnl_gtpu_flex_parser_1_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1389
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1393
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1394
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1395
ste_ctx->build_register_0_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1399
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1403
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1404
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1405
ste_ctx->build_register_1_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1409
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1415
sb->vhca_id_valid = mask->misc.source_eswitch_owner_vhca_id;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1417
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1418
sb->dmn = dmn;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1419
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1420
ste_ctx->build_src_gvmi_qpn_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1424
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1428
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1429
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1430
ste_ctx->build_flex_parser_0_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1434
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1438
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1439
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1440
ste_ctx->build_flex_parser_1_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1444
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1448
sb->rx = rx;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1449
sb->inner = inner;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
1450
ste_ctx->build_tnl_header_0_1_init(sb, mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
755
struct mlx5dr_ste_build *sb;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
763
sb = nic_matcher->ste_builder;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
766
sb->lu_type,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
770
mlx5dr_ste_set_bit_mask(ste_arr, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
772
ret = sb->ste_build_tag_func(value, sb, dr_ste_get_tag(ste_arr));
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
781
sb++;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
782
ste_ctx->set_next_lu_type(ste_arr, sb->lu_type);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.c
783
ste_ctx->set_byte_mask(ste_arr, sb->byte_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste.h
113
((*build_##fname##_init)(struct mlx5dr_ste_build *sb, \
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1008
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1011
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1016
return dr_ste_v0_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1020
dr_ste_v0_build_eth_l2_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1023
dr_ste_v0_build_eth_l2_src_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1024
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1025
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1026
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_src_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1031
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1034
struct mlx5dr_match_spec *mask = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1039
dr_ste_v0_build_eth_l2_src_or_dst_bit_mask(value, sb->inner, bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1044
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1047
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1052
return dr_ste_v0_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1056
dr_ste_v0_build_eth_l2_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1059
dr_ste_v0_build_eth_l2_dst_bit_mask(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1061
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1062
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1063
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1097
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1100
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1141
dr_ste_v0_build_eth_l2_tnl_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1144
dr_ste_v0_build_eth_l2_tnl_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1146
sb->lu_type = DR_STE_V0_LU_TYPE_ETHL2_TUNNELING_I;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1147
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1148
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_tnl_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1153
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1156
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1165
dr_ste_v0_build_eth_l3_ipv4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1168
dr_ste_v0_build_eth_l3_ipv4_misc_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1170
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_MISC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1171
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1172
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv4_misc_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1177
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1180
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1193
if (sb->inner)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1207
dr_ste_v0_build_eth_ipv6_l3_l4_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1210
dr_ste_v0_build_eth_ipv6_l3_l4_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1212
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1213
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1214
sb->ste_build_tag_func = &dr_ste_v0_build_eth_ipv6_l3_l4_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1219
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1224
if (sb->inner)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1233
dr_ste_v0_build_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1236
dr_ste_v0_build_mpls_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1238
sb->lu_type = DR_STE_CALC_LU_TYPE(MPLS_FIRST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1239
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1240
sb->ste_build_tag_func = &dr_ste_v0_build_mpls_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1245
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1264
dr_ste_v0_build_tnl_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1267
dr_ste_v0_build_tnl_gre_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1269
sb->lu_type = DR_STE_V0_LU_TYPE_GRE;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1270
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1271
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gre_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1276
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1307
dr_ste_v0_build_tnl_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1310
dr_ste_v0_build_tnl_mpls_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1312
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1313
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1314
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1319
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1336
parser_id = sb->caps->flex_parser_id_mpls_over_udp;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1344
dr_ste_v0_build_tnl_mpls_over_udp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1347
dr_ste_v0_build_tnl_mpls_over_udp_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1351
sb->lu_type = sb->caps->flex_parser_id_mpls_over_udp > DR_STE_MAX_FLEX_0_ID ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1355
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1356
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_over_udp_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1361
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1378
parser_id = sb->caps->flex_parser_id_mpls_over_gre;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1386
dr_ste_v0_build_tnl_mpls_over_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1389
dr_ste_v0_build_tnl_mpls_over_gre_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1394
sb->lu_type = sb->caps->flex_parser_id_mpls_over_gre > DR_STE_MAX_FLEX_0_ID ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1398
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1399
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_mpls_over_gre_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1407
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1425
dw0_location = sb->caps->flex_parser_id_icmp_dw0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1426
dw1_location = sb->caps->flex_parser_id_icmp_dw1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1431
dw0_location = sb->caps->flex_parser_id_icmpv6_dw0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1432
dw1_location = sb->caps->flex_parser_id_icmpv6_dw1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1450
dr_ste_v0_build_icmp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1456
dr_ste_v0_build_icmp_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1462
parser_id = is_ipv4 ? sb->caps->flex_parser_id_icmp_dw0 :
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1463
sb->caps->flex_parser_id_icmpv6_dw0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1464
sb->lu_type = parser_id > DR_STE_MAX_FLEX_0_ID ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1467
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1468
sb->ste_build_tag_func = &dr_ste_v0_build_icmp_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1473
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1485
dr_ste_v0_build_general_purpose_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1488
dr_ste_v0_build_general_purpose_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1490
sb->lu_type = DR_STE_V0_LU_TYPE_GENERAL_PURPOSE;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1491
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1492
sb->ste_build_tag_func = &dr_ste_v0_build_general_purpose_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1497
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1502
if (sb->inner) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1514
dr_ste_v0_build_eth_l4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1517
dr_ste_v0_build_eth_l4_misc_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1519
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL4_MISC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1520
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1521
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l4_misc_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1526
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1545
dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1548
dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1549
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1550
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1551
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_vxlan_gpe_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1556
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1574
dr_ste_v0_build_flex_parser_tnl_geneve_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1577
dr_ste_v0_build_flex_parser_tnl_geneve_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1578
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1579
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1580
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_geneve_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1585
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1599
dr_ste_v0_build_register_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1602
dr_ste_v0_build_register_0_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1604
sb->lu_type = DR_STE_V0_LU_TYPE_STEERING_REGISTERS_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1605
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1606
sb->ste_build_tag_func = &dr_ste_v0_build_register_0_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1611
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1625
dr_ste_v0_build_register_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1628
dr_ste_v0_build_register_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1630
sb->lu_type = DR_STE_V0_LU_TYPE_STEERING_REGISTERS_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1631
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1632
sb->ste_build_tag_func = &dr_ste_v0_build_register_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1648
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1654
struct mlx5dr_domain *dmn = sb->dmn;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1656
u8 *bit_mask = sb->bit_mask;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1662
if (sb->vhca_id_valid) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1697
dr_ste_v0_build_src_gvmi_qpn_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1700
dr_ste_v0_build_src_gvmi_qpn_bit_mask(mask, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1702
sb->lu_type = DR_STE_V0_LU_TYPE_SRC_GVMI_AND_QP;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1703
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1704
sb->ste_build_tag_func = &dr_ste_v0_build_src_gvmi_qpn_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1727
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1752
static void dr_ste_v0_build_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1755
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1756
dr_ste_v0_build_flex_parser_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1757
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1758
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1761
static void dr_ste_v0_build_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1764
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1765
dr_ste_v0_build_flex_parser_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1766
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1767
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1772
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1776
u8 parser_id = sb->caps->flex_parser_id_geneve_tlv_option_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1787
dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1790
dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1795
sb->lu_type = sb->caps->flex_parser_id_geneve_tlv_option_0 > 3 ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1799
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1800
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_geneve_tlv_opt_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1804
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1822
static void dr_ste_v0_build_flex_parser_tnl_gtpu_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1825
dr_ste_v0_build_flex_parser_tnl_gtpu_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1827
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1828
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1829
sb->ste_build_tag_func = &dr_ste_v0_build_flex_parser_tnl_gtpu_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1834
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1837
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1838
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1839
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_teid))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1840
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1841
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_2))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1842
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1843
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1844
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1849
dr_ste_v0_build_tnl_gtpu_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1852
dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1854
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1855
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1856
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gtpu_flex_parser_0_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1861
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1864
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1865
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1866
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_teid))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1867
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1868
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_2))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1869
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1870
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1871
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1876
dr_ste_v0_build_tnl_gtpu_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1879
dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1881
sb->lu_type = DR_STE_V0_LU_TYPE_FLEX_PARSER_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1882
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1883
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_gtpu_flex_parser_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1887
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1898
static void dr_ste_v0_build_tnl_header_0_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1901
sb->lu_type = DR_STE_V0_LU_TYPE_TUNNEL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1902
dr_ste_v0_build_tnl_header_0_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1903
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
1904
sb->ste_build_tag_func = &dr_ste_v0_build_tnl_header_0_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
739
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
742
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
783
dr_ste_v0_build_eth_l2_src_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
786
dr_ste_v0_build_eth_l2_src_dst_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
788
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL2_SRC_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
789
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
790
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l2_src_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
795
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
798
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
809
dr_ste_v0_build_eth_l3_ipv6_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
812
dr_ste_v0_build_eth_l3_ipv6_dst_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
814
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_DST, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
815
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
816
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv6_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
821
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
824
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
835
dr_ste_v0_build_eth_l3_ipv6_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
838
dr_ste_v0_build_eth_l3_ipv6_src_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
840
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV6_SRC, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
841
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
842
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv6_src_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
847
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
850
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
872
dr_ste_v0_build_eth_l3_ipv4_5_tuple_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
875
dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
877
sb->lu_type = DR_STE_CALC_LU_TYPE(ETHL3_IPV4_5_TUPLE, sb->rx, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
878
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v0.c
879
sb->ste_build_tag_func = &dr_ste_v0_build_eth_l3_ipv4_5_tuple_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1026
void dr_ste_v1_build_eth_l2_src_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1029
dr_ste_v1_build_eth_l2_src_dst_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1031
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL2_SRC_DST, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1032
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1033
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l2_src_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1037
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1040
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1050
void dr_ste_v1_build_eth_l3_ipv6_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1053
dr_ste_v1_build_eth_l3_ipv6_dst_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1055
sb->lu_type = DR_STE_CALC_DFNR_TYPE(IPV6_DES, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1056
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1057
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l3_ipv6_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1061
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1064
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1074
void dr_ste_v1_build_eth_l3_ipv6_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1077
dr_ste_v1_build_eth_l3_ipv6_src_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1079
sb->lu_type = DR_STE_CALC_DFNR_TYPE(IPV6_SRC, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1080
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1081
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l3_ipv6_src_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1085
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1088
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1109
void dr_ste_v1_build_eth_l3_ipv4_5_tuple_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1112
dr_ste_v1_build_eth_l3_ipv4_5_tuple_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1114
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL3_IPV4_5_TUPLE, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1115
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1116
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l3_ipv4_5_tuple_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1239
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1242
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1247
return dr_ste_v1_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1250
void dr_ste_v1_build_eth_l2_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1253
dr_ste_v1_build_eth_l2_src_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1255
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL2_SRC, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1256
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1257
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l2_src_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1272
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1275
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1280
return dr_ste_v1_build_eth_l2_src_or_dst_tag(value, sb->inner, tag);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1283
void dr_ste_v1_build_eth_l2_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1286
dr_ste_v1_build_eth_l2_dst_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1288
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL2, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1289
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1290
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l2_dst_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1322
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1325
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1363
void dr_ste_v1_build_eth_l2_tnl_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1366
dr_ste_v1_build_eth_l2_tnl_bit_mask(mask, sb->inner, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1368
sb->lu_type = DR_STE_V1_LU_TYPE_ETHL2_TNL;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1369
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1370
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l2_tnl_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1374
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1377
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1385
void dr_ste_v1_build_eth_l3_ipv4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1388
dr_ste_v1_build_eth_l3_ipv4_misc_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1390
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL3_IPV4_MISC, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1391
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1392
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l3_ipv4_misc_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1396
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1399
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1412
if (sb->inner)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1425
void dr_ste_v1_build_eth_ipv6_l3_l4_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1428
dr_ste_v1_build_eth_ipv6_l3_l4_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1430
sb->lu_type = DR_STE_CALC_DFNR_TYPE(ETHL4, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1431
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1432
sb->ste_build_tag_func = &dr_ste_v1_build_eth_ipv6_l3_l4_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1436
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1441
if (sb->inner)
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1449
void dr_ste_v1_build_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1452
dr_ste_v1_build_mpls_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1454
sb->lu_type = DR_STE_CALC_DFNR_TYPE(MPLS, sb->inner);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1455
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1456
sb->ste_build_tag_func = &dr_ste_v1_build_mpls_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1460
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1476
void dr_ste_v1_build_tnl_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1479
dr_ste_v1_build_tnl_gre_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1481
sb->lu_type = DR_STE_V1_LU_TYPE_GRE;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1482
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1483
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_gre_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1487
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1521
void dr_ste_v1_build_tnl_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1524
dr_ste_v1_build_tnl_mpls_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1526
sb->lu_type = DR_STE_V1_LU_TYPE_MPLS_I;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1527
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1528
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_mpls_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1532
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1549
parser_id = sb->caps->flex_parser_id_mpls_over_udp;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1556
void dr_ste_v1_build_tnl_mpls_over_udp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1559
dr_ste_v1_build_tnl_mpls_over_udp_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1564
sb->lu_type = sb->caps->flex_parser_id_mpls_over_udp > DR_STE_MAX_FLEX_0_ID ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1568
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1569
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_mpls_over_udp_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1573
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1590
parser_id = sb->caps->flex_parser_id_mpls_over_gre;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1597
void dr_ste_v1_build_tnl_mpls_over_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1600
dr_ste_v1_build_tnl_mpls_over_gre_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1605
sb->lu_type = sb->caps->flex_parser_id_mpls_over_gre > DR_STE_MAX_FLEX_0_ID ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1609
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1610
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_mpls_over_gre_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1614
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1644
void dr_ste_v1_build_icmp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1647
dr_ste_v1_build_icmp_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1649
sb->lu_type = DR_STE_V1_LU_TYPE_ETHL4_MISC_O;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1650
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1651
sb->ste_build_tag_func = &dr_ste_v1_build_icmp_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1655
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1666
void dr_ste_v1_build_general_purpose_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1669
dr_ste_v1_build_general_purpose_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1671
sb->lu_type = DR_STE_V1_LU_TYPE_GENERAL_PURPOSE;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1672
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1673
sb->ste_build_tag_func = &dr_ste_v1_build_general_purpose_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1677
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1682
if (sb->inner) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1693
void dr_ste_v1_build_eth_l4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1696
dr_ste_v1_build_eth_l4_misc_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1698
sb->lu_type = DR_STE_V1_LU_TYPE_ETHL4_MISC_O;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1699
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1700
sb->ste_build_tag_func = &dr_ste_v1_build_eth_l4_misc_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1705
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1723
void dr_ste_v1_build_flex_parser_tnl_vxlan_gpe_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1726
dr_ste_v1_build_flex_parser_tnl_vxlan_gpe_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1728
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1729
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1730
sb->ste_build_tag_func = &dr_ste_v1_build_flex_parser_tnl_vxlan_gpe_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1735
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1752
void dr_ste_v1_build_flex_parser_tnl_geneve_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1755
dr_ste_v1_build_flex_parser_tnl_geneve_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1757
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1758
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1759
sb->ste_build_tag_func = &dr_ste_v1_build_flex_parser_tnl_geneve_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1763
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1774
void dr_ste_v1_build_tnl_header_0_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1777
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1778
dr_ste_v1_build_tnl_header_0_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1779
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1780
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_header_0_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1784
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1797
void dr_ste_v1_build_register_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1800
dr_ste_v1_build_register_0_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1802
sb->lu_type = DR_STE_V1_LU_TYPE_STEERING_REGISTERS_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1803
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1804
sb->ste_build_tag_func = &dr_ste_v1_build_register_0_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1808
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1821
void dr_ste_v1_build_register_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1824
dr_ste_v1_build_register_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1826
sb->lu_type = DR_STE_V1_LU_TYPE_STEERING_REGISTERS_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1827
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1828
sb->ste_build_tag_func = &dr_ste_v1_build_register_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1842
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1848
struct mlx5dr_domain *dmn = sb->dmn;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1850
u8 *bit_mask = sb->bit_mask;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1855
if (sb->vhca_id_valid) {
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1887
void dr_ste_v1_build_src_gvmi_qpn_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1890
dr_ste_v1_build_src_gvmi_qpn_bit_mask(mask, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1892
sb->lu_type = DR_STE_V1_LU_TYPE_SRC_QP_GVMI;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1893
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1894
sb->ste_build_tag_func = &dr_ste_v1_build_src_gvmi_qpn_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1917
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1942
void dr_ste_v1_build_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1945
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1946
dr_ste_v1_build_felx_parser_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1947
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1948
sb->ste_build_tag_func = &dr_ste_v1_build_felx_parser_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1951
void dr_ste_v1_build_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1954
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1955
dr_ste_v1_build_felx_parser_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1956
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1957
sb->ste_build_tag_func = &dr_ste_v1_build_felx_parser_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1962
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1966
u8 parser_id = sb->caps->flex_parser_id_geneve_tlv_option_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1977
dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1980
dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1985
sb->lu_type = sb->caps->flex_parser_id_geneve_tlv_option_0 > 3 ?
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1989
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1990
sb->ste_build_tag_func = &dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1995
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
1998
u8 parser_id = sb->caps->flex_parser_id_geneve_tlv_option_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2010
dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_exist_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2013
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_OK;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2014
dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_exist_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2015
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2016
sb->ste_build_tag_func = &dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_exist_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2020
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2032
void dr_ste_v1_build_flex_parser_tnl_gtpu_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2035
dr_ste_v1_build_flex_parser_tnl_gtpu_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2037
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_TNL_HEADER;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2038
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2039
sb->ste_build_tag_func = &dr_ste_v1_build_flex_parser_tnl_gtpu_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2044
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2047
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2048
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2049
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_teid))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2050
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2051
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_dw_2))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2052
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2053
if (dr_is_flex_parser_0_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2054
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2059
dr_ste_v1_build_tnl_gtpu_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2062
dr_ste_v1_build_tnl_gtpu_flex_parser_0_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2064
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_0;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2065
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2066
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_gtpu_flex_parser_0_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2071
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2074
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2075
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2076
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_teid))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2077
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_teid, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2078
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_dw_2))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2079
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_dw_2, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2080
if (dr_is_flex_parser_1_id(sb->caps->flex_parser_id_gtpu_first_ext_dw_0))
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2081
DR_STE_SET_FLEX_PARSER_FIELD(tag, gtpu_first_ext_dw_0, sb->caps, &value->misc3);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2086
dr_ste_v1_build_tnl_gtpu_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2089
dr_ste_v1_build_tnl_gtpu_flex_parser_1_tag(mask, sb, sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2091
sb->lu_type = DR_STE_V1_LU_TYPE_FLEX_PARSER_1;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2092
sb->byte_mask = mlx5dr_ste_conv_bit_to_byte_mask(sb->bit_mask);
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
2093
sb->ste_build_tag_func = &dr_ste_v1_build_tnl_gtpu_flex_parser_1_tag;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
991
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.c
994
struct mlx5dr_match_spec *spec = sb->inner ? &value->inner : &value->outer;
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
179
void dr_ste_v1_build_eth_l2_src_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
181
void dr_ste_v1_build_eth_l3_ipv6_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
183
void dr_ste_v1_build_eth_l3_ipv6_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
185
void dr_ste_v1_build_eth_l3_ipv4_5_tuple_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
187
void dr_ste_v1_build_eth_l2_src_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
189
void dr_ste_v1_build_eth_l2_dst_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
191
void dr_ste_v1_build_eth_l2_tnl_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
193
void dr_ste_v1_build_eth_l3_ipv4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
195
void dr_ste_v1_build_eth_ipv6_l3_l4_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
197
void dr_ste_v1_build_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
199
void dr_ste_v1_build_tnl_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
201
void dr_ste_v1_build_tnl_mpls_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
203
void dr_ste_v1_build_tnl_mpls_over_udp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
205
void dr_ste_v1_build_tnl_mpls_over_gre_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
207
void dr_ste_v1_build_icmp_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
209
void dr_ste_v1_build_general_purpose_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
211
void dr_ste_v1_build_eth_l4_misc_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
213
void dr_ste_v1_build_flex_parser_tnl_vxlan_gpe_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
215
void dr_ste_v1_build_flex_parser_tnl_geneve_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
217
void dr_ste_v1_build_tnl_header_0_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
219
void dr_ste_v1_build_register_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
221
void dr_ste_v1_build_register_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
223
void dr_ste_v1_build_src_gvmi_qpn_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
225
void dr_ste_v1_build_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
227
void dr_ste_v1_build_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
229
void dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
231
void dr_ste_v1_build_flex_parser_tnl_geneve_tlv_opt_exist_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
233
void dr_ste_v1_build_flex_parser_tnl_gtpu_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
235
void dr_ste_v1_build_tnl_gtpu_flex_parser_0_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_ste_v1.h
237
void dr_ste_v1_build_tnl_gtpu_flex_parser_1_init(struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
224
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
396
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
400
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
404
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
408
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
412
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
416
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
420
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
424
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
428
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
432
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
436
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
440
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
445
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
450
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
455
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
459
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
463
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
468
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
473
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
477
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
482
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
487
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
491
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
495
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
499
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
503
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
508
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
512
struct mlx5dr_ste_build *sb,
drivers/net/ethernet/mellanox/mlx5/core/steering/sws/dr_types.h
515
void mlx5dr_ste_build_empty_always_hit(struct mlx5dr_ste_build *sb, bool rx);
drivers/net/ethernet/mellanox/mlxsw/spectrum.h
168
struct mlxsw_sp_sb *sb;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1263
mlxsw_sp->sb = kzalloc_obj(*mlxsw_sp->sb);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1264
if (!mlxsw_sp->sb)
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1266
mlxsw_sp->sb->cell_size = MLXSW_CORE_RES_GET(mlxsw_sp->core, CELL_SIZE);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1267
mlxsw_sp->sb->sb_size = MLXSW_CORE_RES_GET(mlxsw_sp->core,
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1272
mlxsw_sp->sb->max_headroom_cells = max_headroom_size /
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1273
mlxsw_sp->sb->cell_size;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1294
mlxsw_sp->sb->sb_size,
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
130
return mlxsw_sp->sb->cell_size * cells;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1311
kfree(mlxsw_sp->sb);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1319
kfree(mlxsw_sp->sb);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
135
return DIV_ROUND_UP(bytes, mlxsw_sp->sb->cell_size);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
1367
pool_info->cell_size = mlxsw_sp->sb->cell_size;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
151
return &mlxsw_sp->sb->prs[pool_index];
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
166
struct mlxsw_sp_sb_port *sb_port = &mlxsw_sp->sb->ports[local_port];
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
178
return &mlxsw_sp->sb->ports[local_port].pms[pool_index];
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
198
size = mlxsw_sp_bytes_cells(mlxsw_sp, mlxsw_sp->sb->sb_size);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
239
mlxsw_sp->sb->sb_size);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
538
return taken_headroom_cells <= mlxsw_sp->sb->max_headroom_cells;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
650
mlxsw_sp->sb->ports = kzalloc_objs(struct mlxsw_sp_sb_port, max_ports);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
651
if (!mlxsw_sp->sb->ports)
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
659
mlxsw_sp->sb->prs = prs;
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
662
err = mlxsw_sp_sb_port_init(mlxsw_sp, &mlxsw_sp->sb->ports[i]);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
671
mlxsw_sp_sb_port_fini(&mlxsw_sp->sb->ports[i]);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
672
kfree(mlxsw_sp->sb->prs);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
674
kfree(mlxsw_sp->sb->ports);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
684
mlxsw_sp_sb_port_fini(&mlxsw_sp->sb->ports[i]);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
685
kfree(mlxsw_sp->sb->prs);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
686
kfree(mlxsw_sp->sb->ports);
drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c
751
u32 sb_cells = div_u64(mlxsw_sp->sb->sb_size, mlxsw_sp->sb->cell_size);
drivers/net/ethernet/netronome/nfp/nfp_main.h
186
int nfp_shared_buf_pool_get(struct nfp_pf *pf, unsigned int sb, u16 pool_index,
drivers/net/ethernet/netronome/nfp/nfp_main.h
188
int nfp_shared_buf_pool_set(struct nfp_pf *pf, unsigned int sb,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
109
struct nfp_shared_buf *sb = &pf->shared_bufs[i];
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
112
memcpy_fromio(sb, sb_desc + i * entry_sz, sizeof(*sb));
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
115
le32_to_cpu(sb->id),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
116
le32_to_cpu(sb->size),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
117
le16_to_cpu(sb->ingress_pools_count),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
118
le16_to_cpu(sb->egress_pools_count),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
119
le16_to_cpu(sb->ingress_tc_count),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
120
le16_to_cpu(sb->egress_tc_count));
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
13
static u32 nfp_shared_buf_pool_unit(struct nfp_pf *pf, unsigned int sb)
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
15
__le32 sb_id = cpu_to_le32(sb);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
26
int nfp_shared_buf_pool_get(struct nfp_pf *pf, unsigned int sb, u16 pool_index,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
31
.shared_buf = cpu_to_le32(sb),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
37
unit_size = nfp_shared_buf_pool_unit(pf, sb);
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
56
int nfp_shared_buf_pool_set(struct nfp_pf *pf, unsigned int sb,
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
62
.shared_buf = cpu_to_le32(sb),
drivers/net/ethernet/netronome/nfp/nfp_shared_buf.c
69
unit_size = nfp_shared_buf_pool_unit(pf, sb);
drivers/net/ethernet/qlogic/qed/qed_main.c
2911
qed_get_sb_info(struct qed_dev *cdev, struct qed_sb_info *sb,
drivers/net/ethernet/qlogic/qed/qed_main.c
2928
rc = qed_int_get_sb_dbg(hwfn, ptt, sb, sb_dbg);
drivers/net/ethernet/qlogic/qed/qed_sp_commands.c
307
u16 sb = qed_int_get_sp_sb_id(p_hwfn);
drivers/net/ethernet/qlogic/qed/qed_sp_commands.c
331
p_ramrod->event_ring_sb_id = cpu_to_le16(sb);
drivers/net/ethernet/qlogic/qed/qed_sp_commands.c
418
sb, sb_index, outer_tag_config->outer_tag.tci);
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
345
ch_inf.sb = BRCMU_CHAN_SB_U;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
347
ch_inf.sb = BRCMU_CHAN_SB_L;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
352
ch_inf.sb = BRCMU_CHAN_SB_LL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
354
ch_inf.sb = BRCMU_CHAN_SB_LU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
356
ch_inf.sb = BRCMU_CHAN_SB_UL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
358
ch_inf.sb = BRCMU_CHAN_SB_UU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
363
ch_inf.sb = BRCMU_CHAN_SB_LLL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
365
ch_inf.sb = BRCMU_CHAN_SB_LLU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
367
ch_inf.sb = BRCMU_CHAN_SB_LUL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
369
ch_inf.sb = BRCMU_CHAN_SB_LUU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
371
ch_inf.sb = BRCMU_CHAN_SB_ULL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
373
ch_inf.sb = BRCMU_CHAN_SB_ULU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
375
ch_inf.sb = BRCMU_CHAN_SB_UUL;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
377
ch_inf.sb = BRCMU_CHAN_SB_UUU;
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
6977
if (ch->sb == BRCMU_CHAN_SB_U) {
drivers/net/wireless/broadcom/brcm80211/brcmfmac/cfg80211.c
7166
ch.sb = BRCMU_CHAN_SB_NONE;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
106
ch->sb = BRCMU_CHAN_SB_NONE;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
112
ch->sb = BRCMU_CHAN_SB_L;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
115
ch->sb = BRCMU_CHAN_SB_U;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
13
static u16 d11n_sb(enum brcmu_chan_sb sb)
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
147
ch->sb = BRCMU_CHAN_SB_NONE;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
15
switch (sb) {
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
153
ch->sb = BRCMU_CHAN_SB_L;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
156
ch->sb = BRCMU_CHAN_SB_U;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
164
ch->sb = brcmu_maskget16(ch->chspec, BRCMU_CHSPEC_D11AC_SB_MASK,
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
166
switch (ch->sb) {
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
186
ch->sb = brcmu_maskget16(ch->chspec, BRCMU_CHSPEC_D11AC_SB_MASK,
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
188
switch (ch->sb) {
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
44
ch->sb = BRCMU_CHAN_SB_NONE;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
50
0, d11n_sb(ch->sb));
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
79
if (ch->bw == BRCMU_CHAN_BW_20 || ch->sb == BRCMU_CHAN_SB_NONE)
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
80
ch->sb = BRCMU_CHAN_SB_L;
drivers/net/wireless/broadcom/brcm80211/brcmutil/d11.c
85
BRCMU_CHSPEC_D11AC_SB_SHIFT, ch->sb);
drivers/net/wireless/broadcom/brcm80211/include/brcmu_d11.h
136
enum brcmu_chan_sb sb;
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1678
struct iwl_stored_beacon_notif_common *sb = (void *)pkt->data;
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1682
u32 size = le32_to_cpu(sb->byte_count);
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1715
rx_status.mactime = le64_to_cpu(sb->tsf);
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1718
rx_status.device_timestamp = le32_to_cpu(sb->system_time);
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1720
(sb->band & cpu_to_le16(RX_RES_PHY_FLAGS_BAND_24)) ?
drivers/net/wireless/intel/iwlwifi/mvm/mac-ctxt.c
1723
ieee80211_channel_to_frequency(le16_to_cpu(sb->channel),
drivers/nvdimm/nd.h
540
u64 nd_sb_checksum(struct nd_gen_sb *sb);
drivers/platform/chrome/cros_ec_uart.c
210
struct acpi_resource_uart_serialbus *sb = &ares->data.uart_serial_bus;
drivers/platform/chrome/cros_ec_uart.c
213
sb->type == ACPI_RESOURCE_SERIAL_TYPE_UART) {
drivers/platform/chrome/cros_ec_uart.c
214
ec_uart->baudrate = sb->default_baud_rate;
drivers/platform/chrome/cros_ec_uart.c
217
ec_uart->flowcontrol = sb->flow_control;
drivers/platform/surface/surface3_power.c
381
struct acpi_resource_i2c_serialbus *sb;
drivers/platform/surface/surface3_power.c
393
if (!value64 || !i2c_acpi_get_i2c_resource(ares, &sb)) {
drivers/s390/block/dasd_diag.c
323
unsigned int sb, bsize;
drivers/s390/block/dasd_diag.c
449
for (sb = 512; sb < bsize; sb = sb << 1)
drivers/s390/block/dasd_eckd.c
2342
unsigned int sb, blk_per_trk;
drivers/s390/block/dasd_eckd.c
2418
for (sb = 512; sb < block->bp_block; sb = sb << 1)
drivers/s390/block/dasd_fba.c
185
int sb, rc;
drivers/s390/block/dasd_fba.c
196
for (sb = 512; sb < private->rdc_data.blk_size; sb = sb << 1)
drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h
351
#define h_copy_rdma(l, sa, sb, da, db) \
drivers/scsi/ibmvscsi_tgt/ibmvscsi_tgt.h
352
plpar_hcall_norets(H_COPY_RDMA, l, sa, sb, da, db)
drivers/scsi/lpfc/lpfc_init.c
3948
struct lpfc_io_buf *sb, *sb_next;
drivers/scsi/lpfc/lpfc_init.c
3958
list_for_each_entry_safe(sb, sb_next, &phba->lpfc_scsi_buf_list_put,
drivers/scsi/lpfc/lpfc_init.c
3960
list_del(&sb->list);
drivers/scsi/lpfc/lpfc_init.c
3961
dma_pool_free(phba->lpfc_sg_dma_buf_pool, sb->data,
drivers/scsi/lpfc/lpfc_init.c
3962
sb->dma_handle);
drivers/scsi/lpfc/lpfc_init.c
3963
kfree(sb);
drivers/scsi/lpfc/lpfc_init.c
3969
list_for_each_entry_safe(sb, sb_next, &phba->lpfc_scsi_buf_list_get,
drivers/scsi/lpfc/lpfc_init.c
3971
list_del(&sb->list);
drivers/scsi/lpfc/lpfc_init.c
3972
dma_pool_free(phba->lpfc_sg_dma_buf_pool, sb->data,
drivers/scsi/lpfc/lpfc_init.c
3973
sb->dma_handle);
drivers/scsi/lpfc/lpfc_init.c
3974
kfree(sb);
drivers/scsi/qedf/qedf_main.c
2208
struct status_block *sb = sb_info->sb_virt;
drivers/scsi/qedf/qedf_main.c
2218
prod_idx = sb->pi_array[QEDF_FCOE_PARAMS_GL_RQ_PI];
drivers/scsi/qedf/qedf_main.c
2235
struct status_block *sb = sb_info->sb_virt;
drivers/scsi/qedf/qedf_main.c
2247
prod_idx = sb->pi_array[QEDF_FCOE_PARAMS_GL_RQ_PI];
drivers/scsi/qedi/qedi_debugfs.c
135
struct status_block *sb = NULL;
drivers/scsi/qedi/qedi_debugfs.c
149
sb = sb_info->sb_virt;
drivers/scsi/qedi/qedi_debugfs.c
150
prod_idx = (sb->pi_array[QEDI_PROTO_CQ_PROD_IDX] &
drivers/scsi/qedi/qedi_main.c
1264
struct status_block *sb = sb_info->sb_virt;
drivers/scsi/qedi/qedi_main.c
1274
prod_idx = sb->pi_array[QEDI_PROTO_CQ_PROD_IDX];
drivers/scsi/qedi/qedi_main.c
1320
struct status_block *sb = sb_info->sb_virt;
drivers/scsi/qedi/qedi_main.c
1326
prod_idx = sb->pi_array[QEDI_PROTO_CQ_PROD_IDX];
drivers/scsi/scsi_debug.c
9338
int n, len, sb;
drivers/scsi/scsi_debug.c
9341
sb = (int)sizeof(b);
drivers/scsi/scsi_debug.c
9345
for (k = 0, n = 0; k < len && n < sb; ++k)
drivers/scsi/scsi_debug.c
9346
n += scnprintf(b + n, sb - n, "%02x ",
drivers/spi/spi-mpc52xx-psc.c
111
unsigned sb = 0; /* number of bytes sent */
drivers/spi/spi-mpc52xx-psc.c
129
send_at_once = t->len - sb;
drivers/spi/spi-mpc52xx-psc.c
135
for (; send_at_once; sb++, send_at_once--) {
drivers/spi/spi-mpc52xx-psc.c
141
out_8(&psc->mpc52xx_psc_buffer_8, tx_buf[sb]);
drivers/spi/spi-pl022.c
43
#define SSP_WRITE_BITS(reg, val, mask, sb) \
drivers/spi/spi-pl022.c
44
((reg) = (((reg) & ~(mask)) | (((val)<<(sb)) & (mask))))
drivers/spi/spi-pl022.c
51
#define GEN_MASK_BITS(val, mask, sb) \
drivers/spi/spi-pl022.c
52
(((val)<<(sb)) & (mask))
drivers/spi/spi.c
2742
struct acpi_resource_spi_serialbus *sb;
drivers/spi/spi.c
2748
sb = &ares->data.spi_serial_bus;
drivers/spi/spi.c
2749
if (sb->type != ACPI_RESOURCE_SERIAL_TYPE_SPI)
drivers/spi/spi.c
2815
struct acpi_resource_spi_serialbus *sb;
drivers/spi/spi.c
2819
sb = &ares->data.spi_serial_bus;
drivers/spi/spi.c
2820
if (sb->type == ACPI_RESOURCE_SERIAL_TYPE_SPI) {
drivers/spi/spi.c
2826
sb->resource_source.string_ptr,
drivers/spi/spi.c
2858
sb->device_selection);
drivers/spi/spi.c
2863
lookup->chip_select = sb->device_selection;
drivers/spi/spi.c
2866
lookup->max_speed_hz = sb->connection_speed;
drivers/spi/spi.c
2867
lookup->bits_per_word = sb->data_bit_length;
drivers/spi/spi.c
2869
if (sb->clock_phase == ACPI_SPI_SECOND_PHASE)
drivers/spi/spi.c
2871
if (sb->clock_polarity == ACPI_SPI_START_HIGH)
drivers/spi/spi.c
2873
if (sb->device_polarity == ACPI_SPI_ACTIVE_HIGH)
drivers/tty/hvc/hvc_iucv.c
390
struct iucv_tty_buffer *sb;
drivers/tty/hvc/hvc_iucv.c
404
sb = alloc_tty_buffer(priv->sndbuf_len, GFP_ATOMIC);
drivers/tty/hvc/hvc_iucv.c
405
if (!sb)
drivers/tty/hvc/hvc_iucv.c
408
memcpy(sb->mbuf->data, priv->sndbuf, priv->sndbuf_len);
drivers/tty/hvc/hvc_iucv.c
409
sb->mbuf->datalen = (u16) priv->sndbuf_len;
drivers/tty/hvc/hvc_iucv.c
410
sb->msg.length = MSG_SIZE(sb->mbuf->datalen);
drivers/tty/hvc/hvc_iucv.c
412
list_add_tail(&sb->list, &priv->tty_outqueue);
drivers/tty/hvc/hvc_iucv.c
414
rc = __iucv_message_send(priv->path, &sb->msg, 0, 0,
drivers/tty/hvc/hvc_iucv.c
415
(void *) sb->mbuf, sb->msg.length);
drivers/tty/hvc/hvc_iucv.c
419
list_del(&sb->list);
drivers/tty/hvc/hvc_iucv.c
420
destroy_tty_buffer(sb);
drivers/tty/serdev/core.c
575
struct acpi_resource_uart_serialbus *sb;
drivers/tty/serdev/core.c
580
sb = &ares->data.uart_serial_bus;
drivers/tty/serdev/core.c
581
if (sb->type != ACPI_RESOURCE_SERIAL_TYPE_UART)
drivers/tty/serdev/core.c
584
*uart = sb;
drivers/tty/serdev/core.c
592
struct acpi_resource_uart_serialbus *sb;
drivers/tty/serdev/core.c
595
if (!serdev_acpi_get_uart_resource(ares, &sb))
drivers/tty/serdev/core.c
602
sb->resource_source.string_ptr,
drivers/usb/gadget/function/f_fs.c
1854
ffs_sb_make_inode(struct super_block *sb, void *data,
drivers/usb/gadget/function/f_fs.c
1861
inode = new_inode(sb);
drivers/usb/gadget/function/f_fs.c
1883
static int ffs_sb_create_file(struct super_block *sb, const char *name,
drivers/usb/gadget/function/f_fs.c
1886
struct ffs_data *ffs = sb->s_fs_info;
drivers/usb/gadget/function/f_fs.c
1890
inode = ffs_sb_make_inode(sb, data, fops, NULL, &ffs->file_perms);
drivers/usb/gadget/function/f_fs.c
1893
dentry = simple_start_creating(sb->s_root, name);
drivers/usb/gadget/function/f_fs.c
1919
static int ffs_sb_fill(struct super_block *sb, struct fs_context *fc)
drivers/usb/gadget/function/f_fs.c
1925
ffs->sb = sb;
drivers/usb/gadget/function/f_fs.c
1927
sb->s_fs_info = ffs;
drivers/usb/gadget/function/f_fs.c
1928
sb->s_blocksize = PAGE_SIZE;
drivers/usb/gadget/function/f_fs.c
1929
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/usb/gadget/function/f_fs.c
1930
sb->s_magic = FUNCTIONFS_MAGIC;
drivers/usb/gadget/function/f_fs.c
1931
sb->s_op = &ffs_sb_operations;
drivers/usb/gadget/function/f_fs.c
1932
sb->s_time_gran = 1;
drivers/usb/gadget/function/f_fs.c
1936
inode = ffs_sb_make_inode(sb, NULL,
drivers/usb/gadget/function/f_fs.c
1940
sb->s_root = d_make_root(inode);
drivers/usb/gadget/function/f_fs.c
1941
if (!sb->s_root)
drivers/usb/gadget/function/f_fs.c
1945
return ffs_sb_create_file(sb, "ep0", ffs, &ffs_ep0_operations);
drivers/usb/gadget/function/f_fs.c
2086
ffs_fs_kill_sb(struct super_block *sb)
drivers/usb/gadget/function/f_fs.c
2088
kill_anon_super(sb);
drivers/usb/gadget/function/f_fs.c
2089
if (sb->s_fs_info) {
drivers/usb/gadget/function/f_fs.c
2090
struct ffs_data *ffs = sb->s_fs_info;
drivers/usb/gadget/function/f_fs.c
2172
ffs_epfiles_destroy(ffs->sb, epfiles,
drivers/usb/gadget/function/f_fs.c
2231
ffs_epfiles_destroy(ffs->sb, epfiles, ffs->eps_count);
drivers/usb/gadget/function/f_fs.c
2347
err = ffs_sb_create_file(ffs->sb, epfile->name,
drivers/usb/gadget/function/f_fs.c
2350
ffs_epfiles_destroy(ffs->sb, epfiles, i - 1);
drivers/usb/gadget/function/f_fs.c
2364
static void ffs_epfiles_destroy(struct super_block *sb,
drivers/usb/gadget/function/f_fs.c
2368
struct dentry *root = sb->s_root;
drivers/usb/gadget/function/f_fs.c
271
static void ffs_epfiles_destroy(struct super_block *sb,
drivers/usb/gadget/function/f_fs.c
274
static int ffs_sb_create_file(struct super_block *sb, const char *name,
drivers/usb/gadget/function/u_fs.h
267
struct super_block *sb;
drivers/usb/gadget/legacy/inode.c
152
struct super_block *sb;
drivers/usb/gadget/legacy/inode.c
1580
simple_remove_by_name(dev->sb->s_root, ep->name, NULL);
drivers/usb/gadget/legacy/inode.c
1590
static int gadgetfs_create_file (struct super_block *sb, char const *name,
drivers/usb/gadget/legacy/inode.c
1620
err = gadgetfs_create_file (dev->sb, data->name,
drivers/usb/gadget/legacy/inode.c
1965
gadgetfs_make_inode (struct super_block *sb,
drivers/usb/gadget/legacy/inode.c
1969
struct inode *inode = new_inode (sb);
drivers/usb/gadget/legacy/inode.c
1986
static int gadgetfs_create_file (struct super_block *sb, char const *name,
drivers/usb/gadget/legacy/inode.c
1992
inode = gadgetfs_make_inode (sb, data, fops,
drivers/usb/gadget/legacy/inode.c
1997
dentry = simple_start_creating(sb->s_root, name);
drivers/usb/gadget/legacy/inode.c
2015
gadgetfs_fill_super (struct super_block *sb, struct fs_context *fc)
drivers/usb/gadget/legacy/inode.c
2035
sb->s_blocksize = PAGE_SIZE;
drivers/usb/gadget/legacy/inode.c
2036
sb->s_blocksize_bits = PAGE_SHIFT;
drivers/usb/gadget/legacy/inode.c
2037
sb->s_magic = GADGETFS_MAGIC;
drivers/usb/gadget/legacy/inode.c
2038
sb->s_op = &gadget_fs_operations;
drivers/usb/gadget/legacy/inode.c
2039
sb->s_time_gran = 1;
drivers/usb/gadget/legacy/inode.c
2042
inode = gadgetfs_make_inode (sb,
drivers/usb/gadget/legacy/inode.c
2048
if (!(sb->s_root = d_make_root (inode)))
drivers/usb/gadget/legacy/inode.c
2058
dev->sb = sb;
drivers/usb/gadget/legacy/inode.c
2059
rc = gadgetfs_create_file(sb, CHIP, dev, &ep0_operations);
drivers/usb/gadget/legacy/inode.c
2099
gadgetfs_kill_sb (struct super_block *sb)
drivers/usb/gadget/legacy/inode.c
2102
kill_anon_super (sb);
drivers/usb/host/xhci-sideband.c
101
xhci_remove_secondary_interrupter(xhci_to_hcd(sb->xhci), sb->ir);
drivers/usb/host/xhci-sideband.c
102
sb->ir = NULL;
drivers/usb/host/xhci-sideband.c
118
void xhci_sideband_notify_ep_ring_free(struct xhci_sideband *sb,
drivers/usb/host/xhci-sideband.c
126
if (sb->notify_client)
drivers/usb/host/xhci-sideband.c
127
sb->notify_client(sb->intf, &evt);
drivers/usb/host/xhci-sideband.c
144
xhci_sideband_add_endpoint(struct xhci_sideband *sb,
drivers/usb/host/xhci-sideband.c
150
guard(mutex)(&sb->mutex);
drivers/usb/host/xhci-sideband.c
152
if (!sb->vdev)
drivers/usb/host/xhci-sideband.c
156
ep = &sb->vdev->eps[ep_index];
drivers/usb/host/xhci-sideband.c
169
if (sb->eps[ep_index] || ep->sideband)
drivers/usb/host/xhci-sideband.c
172
ep->sideband = sb;
drivers/usb/host/xhci-sideband.c
173
sb->eps[ep_index] = ep;
drivers/usb/host/xhci-sideband.c
18
xhci_ring_to_sgtable(struct xhci_sideband *sb, struct xhci_ring *ring)
drivers/usb/host/xhci-sideband.c
192
xhci_sideband_remove_endpoint(struct xhci_sideband *sb,
drivers/usb/host/xhci-sideband.c
198
guard(mutex)(&sb->mutex);
drivers/usb/host/xhci-sideband.c
201
ep = sb->eps[ep_index];
drivers/usb/host/xhci-sideband.c
203
if (!ep || !ep->sideband || ep->sideband != sb)
drivers/usb/host/xhci-sideband.c
206
__xhci_sideband_remove_endpoint(sb, ep);
drivers/usb/host/xhci-sideband.c
213
xhci_sideband_stop_endpoint(struct xhci_sideband *sb,
drivers/usb/host/xhci-sideband.c
220
ep = sb->eps[ep_index];
drivers/usb/host/xhci-sideband.c
222
if (!ep || !ep->sideband || ep->sideband != sb)
drivers/usb/host/xhci-sideband.c
225
return xhci_stop_endpoint_sync(sb->xhci, ep, 0, GFP_KERNEL);
drivers/usb/host/xhci-sideband.c
243
xhci_sideband_get_endpoint_buffer(struct xhci_sideband *sb,
drivers/usb/host/xhci-sideband.c
250
ep = sb->eps[ep_index];
drivers/usb/host/xhci-sideband.c
252
if (!ep || !ep->ring || !ep->sideband || ep->sideband != sb)
drivers/usb/host/xhci-sideband.c
255
return xhci_ring_to_sgtable(sb, ep->ring);
drivers/usb/host/xhci-sideband.c
272
xhci_sideband_get_event_buffer(struct xhci_sideband *sb)
drivers/usb/host/xhci-sideband.c
274
if (!sb || !sb->ir)
drivers/usb/host/xhci-sideband.c
277
return xhci_ring_to_sgtable(sb, sb->ir->event_ring);
drivers/usb/host/xhci-sideband.c
28
dev = xhci_to_hcd(sb->xhci)->self.sysdev;
drivers/usb/host/xhci-sideband.c
322
xhci_sideband_create_interrupter(struct xhci_sideband *sb, int num_seg,
drivers/usb/host/xhci-sideband.c
325
if (!sb || !sb->xhci)
drivers/usb/host/xhci-sideband.c
328
guard(mutex)(&sb->mutex);
drivers/usb/host/xhci-sideband.c
330
if (!sb->vdev)
drivers/usb/host/xhci-sideband.c
333
if (sb->ir)
drivers/usb/host/xhci-sideband.c
336
sb->ir = xhci_create_secondary_interrupter(xhci_to_hcd(sb->xhci),
drivers/usb/host/xhci-sideband.c
339
if (!sb->ir)
drivers/usb/host/xhci-sideband.c
342
sb->ir->ip_autoclear = ip_autoclear;
drivers/usb/host/xhci-sideband.c
356
xhci_sideband_remove_interrupter(struct xhci_sideband *sb)
drivers/usb/host/xhci-sideband.c
358
if (!sb)
drivers/usb/host/xhci-sideband.c
361
guard(mutex)(&sb->mutex);
drivers/usb/host/xhci-sideband.c
363
__xhci_sideband_remove_interrupter(sb);
drivers/usb/host/xhci-sideband.c
380
xhci_sideband_interrupter_id(struct xhci_sideband *sb)
drivers/usb/host/xhci-sideband.c
382
if (!sb || !sb->ir)
drivers/usb/host/xhci-sideband.c
385
return sb->ir->intr_num;
drivers/usb/host/xhci-sideband.c
407
struct xhci_sideband *sb;
drivers/usb/host/xhci-sideband.c
417
sb = kzalloc_node(sizeof(*sb), GFP_KERNEL, dev_to_node(hcd->self.sysdev));
drivers/usb/host/xhci-sideband.c
418
if (!sb)
drivers/usb/host/xhci-sideband.c
421
mutex_init(&sb->mutex);
drivers/usb/host/xhci-sideband.c
432
kfree(sb);
drivers/usb/host/xhci-sideband.c
436
sb->xhci = xhci;
drivers/usb/host/xhci-sideband.c
437
sb->vdev = vdev;
drivers/usb/host/xhci-sideband.c
438
sb->intf = intf;
drivers/usb/host/xhci-sideband.c
439
sb->type = type;
drivers/usb/host/xhci-sideband.c
440
sb->notify_client = notify_client;
drivers/usb/host/xhci-sideband.c
441
vdev->sideband = sb;
drivers/usb/host/xhci-sideband.c
445
return sb;
drivers/usb/host/xhci-sideband.c
460
xhci_sideband_unregister(struct xhci_sideband *sb)
drivers/usb/host/xhci-sideband.c
466
if (!sb)
drivers/usb/host/xhci-sideband.c
469
xhci = sb->xhci;
drivers/usb/host/xhci-sideband.c
471
scoped_guard(mutex, &sb->mutex) {
drivers/usb/host/xhci-sideband.c
472
vdev = sb->vdev;
drivers/usb/host/xhci-sideband.c
477
if (sb->eps[i])
drivers/usb/host/xhci-sideband.c
478
__xhci_sideband_remove_endpoint(sb, sb->eps[i]);
drivers/usb/host/xhci-sideband.c
480
__xhci_sideband_remove_interrupter(sb);
drivers/usb/host/xhci-sideband.c
482
sb->vdev = NULL;
drivers/usb/host/xhci-sideband.c
486
sb->xhci = NULL;
drivers/usb/host/xhci-sideband.c
490
kfree(sb);
drivers/usb/host/xhci-sideband.c
78
__xhci_sideband_remove_endpoint(struct xhci_sideband *sb, struct xhci_virt_ep *ep)
drivers/usb/host/xhci-sideband.c
80
lockdep_assert_held(&sb->mutex);
drivers/usb/host/xhci-sideband.c
86
xhci_stop_endpoint_sync(sb->xhci, ep, 0, GFP_KERNEL);
drivers/usb/host/xhci-sideband.c
89
sb->eps[ep->ep_index] = NULL;
drivers/usb/host/xhci-sideband.c
94
__xhci_sideband_remove_interrupter(struct xhci_sideband *sb)
drivers/usb/host/xhci-sideband.c
96
lockdep_assert_held(&sb->mutex);
drivers/usb/host/xhci-sideband.c
98
if (!sb->ir)
drivers/usb/storage/cypress_atacb.c
167
unsigned char *sb = srb->sense_buffer;
drivers/usb/storage/cypress_atacb.c
168
unsigned char *desc = sb + 8;
drivers/usb/storage/cypress_atacb.c
191
memset(sb, 0, SCSI_SENSE_BUFFERSIZE);
drivers/usb/storage/cypress_atacb.c
194
sb[1] = RECOVERED_ERROR;
drivers/usb/storage/cypress_atacb.c
195
sb[2] = 0; /* ATA PASS THROUGH INFORMATION AVAILABLE */
drivers/usb/storage/cypress_atacb.c
196
sb[3] = 0x1D;
drivers/usb/storage/cypress_atacb.c
206
sb[0] = 0x72;
drivers/usb/storage/cypress_atacb.c
210
sb[7] = 14;
drivers/xen/xenfs/super.c
47
static int xenfs_fill_super(struct super_block *sb, struct fs_context *fc)
drivers/xen/xenfs/super.c
68
return simple_fill_super(sb, XENFS_SUPER_MAGIC,
fs/9p/v9fs.h
189
struct super_block *sb, int new);
fs/9p/v9fs.h
196
struct super_block *sb, int new);
fs/9p/v9fs.h
235
struct super_block *sb)
fs/9p/v9fs.h
238
return v9fs_inode_from_fid_dotl(v9ses, fid, sb, 0);
fs/9p/v9fs.h
240
return v9fs_inode_from_fid(v9ses, fid, sb, 0);
fs/9p/v9fs.h
253
struct super_block *sb)
fs/9p/v9fs.h
256
return v9fs_inode_from_fid_dotl(v9ses, fid, sb, 1);
fs/9p/v9fs.h
258
return v9fs_inode_from_fid(v9ses, fid, sb, 1);
fs/9p/v9fs_vfs.h
41
struct inode *v9fs_alloc_inode(struct super_block *sb);
fs/9p/v9fs_vfs.h
54
struct super_block *sb, unsigned int flags);
fs/9p/vfs_inode.c
1107
struct super_block *sb, unsigned int flags)
fs/9p/vfs_inode.c
1110
struct v9fs_session_info *v9ses = sb->s_fs_info;
fs/9p/vfs_inode.c
227
struct inode *v9fs_alloc_inode(struct super_block *sb)
fs/9p/vfs_inode.c
231
v9inode = alloc_inode_sb(sb, v9fs_inode_cache, GFP_KERNEL);
fs/9p/vfs_inode.c
405
static struct inode *v9fs_qid_iget(struct super_block *sb,
fs/9p/vfs_inode.c
414
struct v9fs_session_info *v9ses = sb->s_fs_info;
fs/9p/vfs_inode.c
422
inode = iget5_locked(sb, QID2INO(qid), test, v9fs_set_inode, st);
fs/9p/vfs_inode.c
438
v9fs_stat2inode(st, inode, sb, 0);
fs/9p/vfs_inode.c
451
struct super_block *sb, int new)
fs/9p/vfs_inode.c
460
inode = v9fs_qid_iget(sb, &st->qid, st, new);
fs/9p/vfs_inode_dotl.c
104
struct v9fs_session_info *v9ses = sb->s_fs_info;
fs/9p/vfs_inode_dotl.c
112
inode = iget5_locked(sb, QID2INO(qid), test, v9fs_set_inode_dotl, st);
fs/9p/vfs_inode_dotl.c
145
struct super_block *sb, int new)
fs/9p/vfs_inode_dotl.c
154
inode = v9fs_qid_iget_dotl(sb, &st->qid, fid, st, new);
fs/9p/vfs_inode_dotl.c
96
static struct inode *v9fs_qid_iget_dotl(struct super_block *sb,
fs/9p/vfs_super.c
104
sb = sget_fc(fc, NULL, set_anon_super_fc);
fs/9p/vfs_super.c
105
if (IS_ERR(sb)) {
fs/9p/vfs_super.c
106
retval = PTR_ERR(sb);
fs/9p/vfs_super.c
109
retval = v9fs_fill_super(sb);
fs/9p/vfs_super.c
114
set_default_d_op(sb, &v9fs_cached_dentry_operations);
fs/9p/vfs_super.c
116
set_default_d_op(sb, &v9fs_dentry_operations);
fs/9p/vfs_super.c
117
sb->s_d_flags |= DCACHE_DONTCACHE;
fs/9p/vfs_super.c
120
inode = v9fs_get_new_inode_from_fid(v9ses, fid, sb);
fs/9p/vfs_super.c
131
sb->s_root = root;
fs/9p/vfs_super.c
138
fc->root = dget(sb->s_root);
fs/9p/vfs_super.c
156
deactivate_locked_super(sb);
fs/9p/vfs_super.c
182
v9fs_umount_begin(struct super_block *sb)
fs/9p/vfs_super.c
186
v9ses = sb->s_fs_info;
fs/9p/vfs_super.c
34
static int v9fs_fill_super(struct super_block *sb)
fs/9p/vfs_super.c
37
struct v9fs_session_info *v9ses = v9ses = sb->s_fs_info;
fs/9p/vfs_super.c
39
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/9p/vfs_super.c
40
sb->s_blocksize_bits = fls(v9ses->maxdata - 1);
fs/9p/vfs_super.c
41
sb->s_blocksize = 1 << sb->s_blocksize_bits;
fs/9p/vfs_super.c
42
sb->s_magic = V9FS_MAGIC;
fs/9p/vfs_super.c
44
sb->s_op = &v9fs_super_ops_dotl;
fs/9p/vfs_super.c
46
sb->s_xattr = v9fs_xattr_handlers;
fs/9p/vfs_super.c
48
sb->s_op = &v9fs_super_ops;
fs/9p/vfs_super.c
49
sb->s_time_max = U32_MAX;
fs/9p/vfs_super.c
52
sb->s_time_min = 0;
fs/9p/vfs_super.c
54
ret = super_setup_bdi(sb);
fs/9p/vfs_super.c
59
sb->s_bdi->ra_pages = 0;
fs/9p/vfs_super.c
60
sb->s_bdi->io_pages = 0;
fs/9p/vfs_super.c
62
sb->s_bdi->ra_pages = v9ses->maxdata >> PAGE_SHIFT;
fs/9p/vfs_super.c
63
sb->s_bdi->io_pages = v9ses->maxdata >> PAGE_SHIFT;
fs/9p/vfs_super.c
66
sb->s_flags |= SB_ACTIVE;
fs/9p/vfs_super.c
70
sb->s_flags |= SB_POSIXACL;
fs/9p/vfs_super.c
84
struct super_block *sb = NULL;
fs/adfs/adfs.h
126
int (*read)(struct super_block *sb, unsigned int indaddr,
fs/adfs/adfs.h
145
struct inode *adfs_iget(struct super_block *sb, struct object_info *obj);
fs/adfs/adfs.h
151
int adfs_map_lookup(struct super_block *sb, u32 frag_id, unsigned int offset);
fs/adfs/adfs.h
152
void adfs_map_statfs(struct super_block *sb, struct kstatfs *buf);
fs/adfs/adfs.h
153
struct adfs_discmap *adfs_read_map(struct super_block *sb, struct adfs_discrecord *dr);
fs/adfs/adfs.h
154
void adfs_free_map(struct super_block *sb);
fs/adfs/adfs.h
158
void __adfs_error(struct super_block *sb, const char *function,
fs/adfs/adfs.h
160
#define adfs_error(sb, fmt...) __adfs_error(sb, __func__, fmt)
fs/adfs/adfs.h
161
void adfs_msg(struct super_block *sb, const char *pfx, const char *fmt, ...);
fs/adfs/adfs.h
181
int adfs_dir_read_buffers(struct super_block *sb, u32 indaddr,
fs/adfs/adfs.h
184
extern int adfs_dir_update(struct super_block *sb, struct object_info *obj,
fs/adfs/adfs.h
206
static inline int __adfs_block_map(struct super_block *sb, u32 indaddr,
fs/adfs/adfs.h
213
block += off << ADFS_SB(sb)->s_log2sharesize;
fs/adfs/adfs.h
216
return adfs_map_lookup(sb, indaddr >> 8, block);
fs/adfs/adfs.h
82
static inline struct adfs_sb_info *ADFS_SB(struct super_block *sb)
fs/adfs/adfs.h
84
return sb->s_fs_info;
fs/adfs/adfs.h
91
struct super_block *sb;
fs/adfs/dir.c
105
num = ALIGN(size, sb->s_blocksize) >> sb->s_blocksize_bits;
fs/adfs/dir.c
122
block = __adfs_block_map(sb, indaddr, i);
fs/adfs/dir.c
124
adfs_error(sb, "dir %06x has a hole at offset %u",
fs/adfs/dir.c
129
dir->bhs[i] = sb_bread(sb, block);
fs/adfs/dir.c
131
adfs_error(sb,
fs/adfs/dir.c
147
static int adfs_dir_read(struct super_block *sb, u32 indaddr,
fs/adfs/dir.c
150
dir->sb = sb;
fs/adfs/dir.c
154
return ADFS_SB(sb)->s_dir->read(sb, indaddr, size, dir);
fs/adfs/dir.c
157
static int adfs_dir_read_inode(struct super_block *sb, struct inode *inode,
fs/adfs/dir.c
162
ret = adfs_dir_read(sb, ADFS_I(inode)->indaddr, inode->i_size, dir);
fs/adfs/dir.c
167
adfs_error(sb,
fs/adfs/dir.c
21
struct super_block *sb = dir->sb;
fs/adfs/dir.c
227
if (!(obj->attr & ADFS_NDA_DIRECTORY) && ADFS_SB(dir->sb)->s_ftsuffix) {
fs/adfs/dir.c
24
index = offset >> sb->s_blocksize_bits;
fs/adfs/dir.c
242
struct super_block *sb = inode->i_sb;
fs/adfs/dir.c
243
const struct adfs_dir_ops *ops = ADFS_SB(sb)->s_dir;
fs/adfs/dir.c
248
ret = adfs_dir_read_inode(sb, inode, &dir);
fs/adfs/dir.c
25
offset &= sb->s_blocksize - 1;
fs/adfs/dir.c
26
remain = sb->s_blocksize - offset;
fs/adfs/dir.c
276
adfs_dir_update(struct super_block *sb, struct object_info *obj, int wait)
fs/adfs/dir.c
278
const struct adfs_dir_ops *ops = ADFS_SB(sb)->s_dir;
fs/adfs/dir.c
289
ret = adfs_dir_read(sb, obj->parent_id, 0, &dir);
fs/adfs/dir.c
351
struct super_block *sb = inode->i_sb;
fs/adfs/dir.c
352
const struct adfs_dir_ops *ops = ADFS_SB(sb)->s_dir;
fs/adfs/dir.c
359
ret = adfs_dir_read_inode(sb, inode, &dir);
fs/adfs/dir.c
46
struct super_block *sb = dir->sb;
fs/adfs/dir.c
49
index = offset >> sb->s_blocksize_bits;
fs/adfs/dir.c
50
offset &= sb->s_blocksize - 1;
fs/adfs/dir.c
51
remain = sb->s_blocksize - offset;
fs/adfs/dir.c
75
dir->sb = NULL;
fs/adfs/dir.c
98
int adfs_dir_read_buffers(struct super_block *sb, u32 indaddr,
fs/adfs/dir_f.c
141
static int adfs_f_read(struct super_block *sb, u32 indaddr, unsigned int size,
fs/adfs/dir_f.c
144
const unsigned int blocksize_bits = sb->s_blocksize_bits;
fs/adfs/dir_f.c
150
ret = adfs_dir_read_buffers(sb, indaddr, ADFS_NEWDIR_SIZE, dir);
fs/adfs/dir_f.c
165
adfs_error(sb, "dir %06x is corrupted", indaddr);
fs/adfs/dir_f.c
282
adfs_error(dir->sb, "error reading directory entry");
fs/adfs/dir_f.c
286
adfs_error(dir->sb, "unable to locate entry to update");
fs/adfs/dir_f.c
312
adfs_msg(dir->sb, KERN_ERR, "error: update broke directory");
fs/adfs/dir_f.c
73
const int blocksize_bits = dir->sb->s_blocksize_bits;
fs/adfs/dir_fplus.c
103
static int adfs_fplus_read(struct super_block *sb, u32 indaddr,
fs/adfs/dir_fplus.c
112
ret = adfs_dir_read_buffers(sb, indaddr, sb->s_blocksize, dir);
fs/adfs/dir_fplus.c
119
adfs_error(sb, "dir %06x has malformed header", indaddr);
fs/adfs/dir_fplus.c
125
adfs_msg(sb, KERN_WARNING,
fs/adfs/dir_fplus.c
131
ret = adfs_dir_read_buffers(sb, indaddr, dirsize, dir);
fs/adfs/dir_fplus.c
136
(dir->bhs[dir->nr_buffers - 1]->b_data + (sb->s_blocksize - 8));
fs/adfs/dir_fplus.c
140
adfs_error(sb, "dir %06x has malformed tail", indaddr);
fs/adfs/dir_fplus.c
145
adfs_error(sb, "dir %06x checkbyte mismatch\n", indaddr);
fs/adfs/dir_fplus.c
242
adfs_error(dir->sb, "unable to locate entry to update");
fs/adfs/dir_fplus.c
247
adfs_error(dir->sb, "error reading directory entry");
fs/adfs/inode.c
135
static int adfs_mode2atts(struct super_block *sb, struct inode *inode,
fs/adfs/inode.c
138
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/inode.c
245
adfs_iget(struct super_block *sb, struct object_info *obj)
fs/adfs/inode.c
250
inode = new_inode(sb);
fs/adfs/inode.c
254
inode->i_uid = ADFS_SB(sb)->s_uid;
fs/adfs/inode.c
255
inode->i_gid = ADFS_SB(sb)->s_gid;
fs/adfs/inode.c
259
inode->i_blocks = (inode->i_size + sb->s_blocksize - 1) >>
fs/adfs/inode.c
260
sb->s_blocksize_bits;
fs/adfs/inode.c
274
inode->i_mode = adfs_atts2mode(sb, inode);
fs/adfs/inode.c
306
struct super_block *sb = inode->i_sb;
fs/adfs/inode.c
316
if ((ia_valid & ATTR_UID && !uid_eq(attr->ia_uid, ADFS_SB(sb)->s_uid)) ||
fs/adfs/inode.c
317
(ia_valid & ATTR_GID && !gid_eq(attr->ia_gid, ADFS_SB(sb)->s_gid)))
fs/adfs/inode.c
342
ADFS_I(inode)->attr = adfs_mode2atts(sb, inode, attr->ia_mode);
fs/adfs/inode.c
343
inode->i_mode = adfs_atts2mode(sb, inode);
fs/adfs/inode.c
363
struct super_block *sb = inode->i_sb;
fs/adfs/inode.c
374
return adfs_dir_update(sb, &obj, wbc->sync_mode == WB_SYNC_ALL);
fs/adfs/inode.c
92
adfs_atts2mode(struct super_block *sb, struct inode *inode)
fs/adfs/inode.c
96
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/map.c
196
void adfs_map_statfs(struct super_block *sb, struct kstatfs *buf)
fs/adfs/map.c
198
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/map.c
211
buf->f_blocks = adfs_disc_size(dr) >> sb->s_blocksize_bits;
fs/adfs/map.c
217
int adfs_map_lookup(struct super_block *sb, u32 frag_id, unsigned int offset)
fs/adfs/map.c
219
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/map.c
250
adfs_error(sb, "fragment 0x%04x at offset %d not found in map",
fs/adfs/map.c
255
adfs_error(sb, "invalid fragment 0x%04x (zone = %d, max = %d)",
fs/adfs/map.c
260
static unsigned char adfs_calczonecheck(struct super_block *sb, unsigned char *map)
fs/adfs/map.c
266
for (i = sb->s_blocksize - 4; i; i -= 4) {
fs/adfs/map.c
284
static int adfs_checkmap(struct super_block *sb, struct adfs_discmap *dm)
fs/adfs/map.c
289
for (i = 0; i < ADFS_SB(sb)->s_map_size; i++) {
fs/adfs/map.c
294
if (adfs_calczonecheck(sb, map) != map[0]) {
fs/adfs/map.c
295
adfs_error(sb, "zone %d fails zonecheck", i);
fs/adfs/map.c
301
adfs_error(sb, "crosscheck != 0xff");
fs/adfs/map.c
334
static int adfs_map_read(struct adfs_discmap *dm, struct super_block *sb,
fs/adfs/map.c
340
dm[zone].dm_bh = sb_bread(sb, map_addr + zone);
fs/adfs/map.c
356
struct adfs_discmap *adfs_read_map(struct super_block *sb, struct adfs_discrecord *dr)
fs/adfs/map.c
358
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/map.c
378
adfs_error(sb, "not enough memory");
fs/adfs/map.c
384
ret = adfs_map_read(dm, sb, map_addr, nzones);
fs/adfs/map.c
386
adfs_error(sb, "unable to read map");
fs/adfs/map.c
390
if (adfs_checkmap(sb, dm))
fs/adfs/map.c
393
adfs_error(sb, "map corrupted");
fs/adfs/map.c
401
void adfs_free_map(struct super_block *sb)
fs/adfs/map.c
403
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/super.c
178
struct super_block *sb = dentry->d_sb;
fs/adfs/super.c
179
struct adfs_sb_info *sbi = ADFS_SB(sb);
fs/adfs/super.c
180
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/adfs/super.c
182
adfs_map_statfs(sb, buf);
fs/adfs/super.c
186
buf->f_bsize = sb->s_blocksize;
fs/adfs/super.c
195
static struct inode *adfs_alloc_inode(struct super_block *sb)
fs/adfs/super.c
198
ei = alloc_inode_sb(sb, adfs_inode_cachep, GFP_KERNEL);
fs/adfs/super.c
254
static int adfs_probe(struct super_block *sb, unsigned int offset, int silent,
fs/adfs/super.c
255
int (*validate)(struct super_block *sb,
fs/adfs/super.c
259
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/super.c
26
void __adfs_error(struct super_block *sb, const char *function, const char *fmt, ...)
fs/adfs/super.c
267
if (sb->s_blocksize != blocksize &&
fs/adfs/super.c
268
!sb_set_blocksize(sb, blocksize)) {
fs/adfs/super.c
270
adfs_msg(sb, KERN_ERR,
fs/adfs/super.c
276
bh = sb_bread(sb, offset >> sb->s_blocksize_bits);
fs/adfs/super.c
278
adfs_msg(sb, KERN_ERR,
fs/adfs/super.c
280
offset >> sb->s_blocksize_bits, try);
fs/adfs/super.c
285
ret = validate(sb, bh, &dr);
fs/adfs/super.c
293
if (sb->s_blocksize == blocksize) {
fs/adfs/super.c
294
asb->s_map = adfs_read_map(sb, dr);
fs/adfs/super.c
305
static int adfs_validate_bblk(struct super_block *sb, struct buffer_head *bh,
fs/adfs/super.c
311
b_data = bh->b_data + (ADFS_DISCRECORD % sb->s_blocksize);
fs/adfs/super.c
324
static int adfs_validate_dr0(struct super_block *sb, struct buffer_head *bh,
fs/adfs/super.c
338
static int adfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/adfs/super.c
342
struct adfs_sb_info *asb = sb->s_fs_info;
fs/adfs/super.c
347
sb->s_flags |= ADFS_SB_FLAGS;
fs/adfs/super.c
349
sb->s_fs_info = asb;
fs/adfs/super.c
350
sb->s_magic = ADFS_SUPER_MAGIC;
fs/adfs/super.c
351
sb->s_time_gran = 10000000;
fs/adfs/super.c
354
ret = adfs_probe(sb, ADFS_DISCRECORD, 1, adfs_validate_bblk);
fs/adfs/super.c
356
ret = adfs_probe(sb, 0, silent, adfs_validate_dr0);
fs/adfs/super.c
359
adfs_msg(sb, KERN_ERR,
fs/adfs/super.c
36
sb->s_id, function ? ": " : "",
fs/adfs/super.c
361
sb->s_id);
fs/adfs/super.c
368
sb->s_op = &adfs_sops;
fs/adfs/super.c
400
set_default_d_op(sb, &adfs_dentry_operations);
fs/adfs/super.c
401
root = adfs_iget(sb, &root_obj);
fs/adfs/super.c
402
sb->s_root = d_make_root(root);
fs/adfs/super.c
403
if (!sb->s_root) {
fs/adfs/super.c
404
adfs_free_map(sb);
fs/adfs/super.c
405
adfs_error(sb, "get root inode failed\n");
fs/adfs/super.c
412
sb->s_fs_info = NULL;
fs/adfs/super.c
42
void adfs_msg(struct super_block *sb, const char *pfx, const char *fmt, ...)
fs/adfs/super.c
445
struct super_block *sb = fc->root->d_sb;
fs/adfs/super.c
446
struct adfs_sb_info *old_asb = ADFS_SB(sb);
fs/adfs/super.c
50
printk("%sADFS-fs (%s): %pV\n", pfx, sb->s_id, &vaf);
fs/adfs/super.c
93
static void adfs_put_super(struct super_block *sb)
fs/adfs/super.c
95
struct adfs_sb_info *asb = ADFS_SB(sb);
fs/adfs/super.c
97
adfs_free_map(sb);
fs/affs/affs.h
102
struct super_block *sb; /* the VFS superblock object */
fs/affs/affs.h
127
static inline struct affs_sb_info *AFFS_SB(struct super_block *sb)
fs/affs/affs.h
129
return sb->s_fs_info;
fs/affs/affs.h
132
void affs_mark_sb_dirty(struct super_block *sb);
fs/affs/affs.h
139
extern u32 affs_checksum_block(struct super_block *sb, struct buffer_head *bh);
fs/affs/affs.h
140
extern void affs_fix_checksum(struct super_block *sb, struct buffer_head *bh);
fs/affs/affs.h
145
extern void affs_error(struct super_block *sb, const char *function,
fs/affs/affs.h
148
extern void affs_warning(struct super_block *sb, const char *function,
fs/affs/affs.h
158
extern void affs_free_block(struct super_block *sb, u32 block);
fs/affs/affs.h
160
extern int affs_init_bitmap(struct super_block *sb, int *flags);
fs/affs/affs.h
161
extern void affs_free_bitmap(struct super_block *sb);
fs/affs/affs.h
166
extern int affs_hash_name(struct super_block *sb, const u8 *name, unsigned int len);
fs/affs/affs.h
17
#define AFFS_BLOCK(sb, bh, blk) (AFFS_HEAD(bh)->table[AFFS_SB(sb)->s_hashsize-1-(blk)])
fs/affs/affs.h
190
extern struct inode *affs_iget(struct super_block *sb,
fs/affs/affs.h
20
#define AFFS_TAIL(sb, bh) ((struct affs_tail *)((bh)->b_data+(sb)->s_blocksize-sizeof(struct affs_tail)))
fs/affs/affs.h
22
#define AFFS_ROOT_TAIL(sb, bh) ((struct affs_root_tail *)((bh)->b_data+(sb)->s_blocksize-sizeof(struct affs_root_tail)))
fs/affs/affs.h
222
static inline bool affs_validblock(struct super_block *sb, int block)
fs/affs/affs.h
224
return(block >= AFFS_SB(sb)->s_reserved &&
fs/affs/affs.h
225
block < AFFS_SB(sb)->s_partition_size);
fs/affs/affs.h
229
affs_set_blocksize(struct super_block *sb, int size)
fs/affs/affs.h
231
sb_set_blocksize(sb, size);
fs/affs/affs.h
234
affs_bread(struct super_block *sb, int block)
fs/affs/affs.h
237
if (affs_validblock(sb, block))
fs/affs/affs.h
238
return sb_bread(sb, block);
fs/affs/affs.h
242
affs_getblk(struct super_block *sb, int block)
fs/affs/affs.h
245
if (affs_validblock(sb, block))
fs/affs/affs.h
246
return sb_getblk(sb, block);
fs/affs/affs.h
250
affs_getzeroblk(struct super_block *sb, int block)
fs/affs/affs.h
254
if (affs_validblock(sb, block)) {
fs/affs/affs.h
255
bh = sb_getblk(sb, block);
fs/affs/affs.h
257
memset(bh->b_data, 0 , sb->s_blocksize);
fs/affs/affs.h
265
affs_getemptyblk(struct super_block *sb, int block)
fs/affs/affs.h
269
if (affs_validblock(sb, block)) {
fs/affs/affs.h
270
bh = sb_getblk(sb, block);
fs/affs/amigaffs.c
101
AFFS_TAIL(sb, bh)->hash_chain = ino;
fs/affs/amigaffs.c
104
AFFS_TAIL(sb, rem_bh)->parent = 0;
fs/affs/amigaffs.c
109
bh = affs_bread(sb, hash_ino);
fs/affs/amigaffs.c
112
hash_ino = be32_to_cpu(AFFS_TAIL(sb, bh)->hash_chain);
fs/affs/amigaffs.c
145
struct super_block *sb = inode->i_sb;
fs/affs/amigaffs.c
152
bh = affs_bread(sb, inode->i_ino);
fs/affs/amigaffs.c
161
link_ino = be32_to_cpu(AFFS_TAIL(sb, bh)->link_chain);
fs/affs/amigaffs.c
162
link_bh = affs_bread(sb, link_ino);
fs/affs/amigaffs.c
166
dir = affs_iget(sb, be32_to_cpu(AFFS_TAIL(sb, link_bh)->parent));
fs/affs/amigaffs.c
185
memcpy(AFFS_TAIL(sb, bh)->name, AFFS_TAIL(sb, link_bh)->name, 32);
fs/affs/amigaffs.c
196
link_bh = affs_bread(sb, link_ino);
fs/affs/amigaffs.c
201
while ((ino = be32_to_cpu(AFFS_TAIL(sb, bh)->link_chain)) != 0) {
fs/affs/amigaffs.c
203
__be32 ino2 = AFFS_TAIL(sb, link_bh)->link_chain;
fs/affs/amigaffs.c
204
AFFS_TAIL(sb, bh)->link_chain = ino2;
fs/affs/amigaffs.c
209
switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) {
fs/affs/amigaffs.c
214
if (!AFFS_TAIL(sb, bh)->link_chain)
fs/affs/amigaffs.c
217
affs_free_block(sb, link_ino);
fs/affs/amigaffs.c
221
bh = affs_bread(sb, ino);
fs/affs/amigaffs.c
236
struct super_block *sb = inode->i_sb;
fs/affs/amigaffs.c
241
bh = affs_bread(sb, inode->i_ino);
fs/affs/amigaffs.c
246
for (size = AFFS_SB(sb)->s_hashsize - 1; size >= 0; size--)
fs/affs/amigaffs.c
269
struct super_block *sb;
fs/affs/amigaffs.c
275
sb = dir->i_sb;
fs/affs/amigaffs.c
28
struct super_block *sb = dir->i_sb;
fs/affs/amigaffs.c
284
bh = affs_bread(sb, (u32)(long)dentry->d_fsdata);
fs/affs/amigaffs.c
290
switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) {
fs/affs/amigaffs.c
339
affs_checksum_block(struct super_block *sb, struct buffer_head *bh)
fs/affs/amigaffs.c
34
offset = affs_hash_name(sb, AFFS_TAIL(sb, bh)->name + 1, AFFS_TAIL(sb, bh)->name[0]);
fs/affs/amigaffs.c
346
for (bsize = sb->s_blocksize / sizeof(__be32); bsize > 0; bsize--)
fs/affs/amigaffs.c
357
affs_fix_checksum(struct super_block *sb, struct buffer_head *bh)
fs/affs/amigaffs.c
359
int cnt = sb->s_blocksize / sizeof(__be32);
fs/affs/amigaffs.c
38
dir_bh = affs_bread(sb, dir->i_ino);
fs/affs/amigaffs.c
45
dir_bh = affs_bread(sb, hash_ino);
fs/affs/amigaffs.c
473
affs_error(struct super_block *sb, const char *function, const char *fmt, ...)
fs/affs/amigaffs.c
48
hash_ino = be32_to_cpu(AFFS_TAIL(sb, dir_bh)->hash_chain);
fs/affs/amigaffs.c
481
pr_crit("error (device %s): %s(): %pV\n", sb->s_id, function, &vaf);
fs/affs/amigaffs.c
482
if (!sb_rdonly(sb))
fs/affs/amigaffs.c
484
sb->s_flags |= SB_RDONLY;
fs/affs/amigaffs.c
489
affs_warning(struct super_block *sb, const char *function, const char *fmt, ...)
fs/affs/amigaffs.c
497
pr_warn("(device %s): %s(): %pV\n", sb->s_id, function, &vaf);
fs/affs/amigaffs.c
50
AFFS_TAIL(sb, bh)->parent = cpu_to_be32(dir->i_ino);
fs/affs/amigaffs.c
51
AFFS_TAIL(sb, bh)->hash_chain = 0;
fs/affs/amigaffs.c
52
affs_fix_checksum(sb, bh);
fs/affs/amigaffs.c
57
AFFS_TAIL(sb, dir_bh)->hash_chain = cpu_to_be32(ino);
fs/affs/amigaffs.c
77
struct super_block *sb;
fs/affs/amigaffs.c
83
sb = dir->i_sb;
fs/affs/amigaffs.c
85
offset = affs_hash_name(sb, AFFS_TAIL(sb, rem_bh)->name+1, AFFS_TAIL(sb, rem_bh)->name[0]);
fs/affs/amigaffs.c
89
bh = affs_bread(sb, dir->i_ino);
fs/affs/amigaffs.c
97
ino = AFFS_TAIL(sb, rem_bh)->hash_chain;
fs/affs/bitmap.c
102
affs_error(sb, "affs_free_block","Block %u outside partition", block);
fs/affs/bitmap.c
117
struct super_block *sb;
fs/affs/bitmap.c
125
sb = inode->i_sb;
fs/affs/bitmap.c
126
sbi = AFFS_SB(sb);
fs/affs/bitmap.c
138
affs_warning(sb, "affs_balloc", "invalid goal %d", goal);
fs/affs/bitmap.c
15
affs_count_free_blocks(struct super_block *sb)
fs/affs/bitmap.c
174
bh = affs_bread(sb, bm->bm_key);
fs/affs/bitmap.c
184
enddata = (__be32 *)((u8 *)bh->b_data + sb->s_blocksize);
fs/affs/bitmap.c
227
affs_mark_sb_dirty(sb);
fs/affs/bitmap.c
23
if (sb_rdonly(sb))
fs/affs/bitmap.c
235
affs_error(sb,"affs_read_block","Cannot read bitmap block %u", bm->bm_key);
fs/affs/bitmap.c
244
int affs_init_bitmap(struct super_block *sb, int *flags)
fs/affs/bitmap.c
251
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/bitmap.c
256
if (!AFFS_ROOT_TAIL(sb, sbi->s_root_bh)->bm_flag) {
fs/affs/bitmap.c
257
pr_notice("Bitmap invalid - mounting %s read only\n", sb->s_id);
fs/affs/bitmap.c
26
mutex_lock(&AFFS_SB(sb)->s_bmlock);
fs/affs/bitmap.c
264
sbi->s_bmap_bits = sb->s_blocksize * 8 - 32;
fs/affs/bitmap.c
275
blk = sb->s_blocksize / 4 - 49;
fs/affs/bitmap.c
28
bm = AFFS_SB(sb)->s_bitmap;
fs/affs/bitmap.c
282
bh = affs_bread(sb, bm->bm_key);
fs/affs/bitmap.c
288
if (affs_checksum_block(sb, bh)) {
fs/affs/bitmap.c
290
bm->bm_key, sb->s_id);
fs/affs/bitmap.c
295
bm->bm_free = memweight(bh->b_data + 4, sb->s_blocksize - 4);
fs/affs/bitmap.c
30
for (i = AFFS_SB(sb)->s_bmap_count; i > 0; bm++, i--)
fs/affs/bitmap.c
304
bmap_bh = affs_bread(sb, be32_to_cpu(bmap_blk[blk]));
fs/affs/bitmap.c
312
end = sb->s_blocksize / 4 - 1;
fs/affs/bitmap.c
33
mutex_unlock(&AFFS_SB(sb)->s_bmlock);
fs/affs/bitmap.c
337
while (++offset < sb->s_blocksize / 4)
fs/affs/bitmap.c
340
((__be32 *)bh->b_data)[0] = cpu_to_be32(-affs_checksum_block(sb, bh));
fs/affs/bitmap.c
345
bm->bm_free = memweight(bh->b_data + 4, sb->s_blocksize - 4);
fs/affs/bitmap.c
353
void affs_free_bitmap(struct super_block *sb)
fs/affs/bitmap.c
355
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/bitmap.c
39
affs_free_block(struct super_block *sb, u32 block)
fs/affs/bitmap.c
41
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/bitmap.c
62
bh = affs_bread(sb, bm->bm_key);
fs/affs/bitmap.c
83
affs_mark_sb_dirty(sb);
fs/affs/bitmap.c
90
affs_warning(sb,"affs_free_block","Trying to free block %u which is already free", block);
fs/affs/bitmap.c
95
affs_error(sb,"affs_free_block","Cannot read bitmap block %u", bm->bm_key);
fs/affs/dir.c
105
affs_warning(sb, "readdir", "More than 65535 entries in chain");
fs/affs/dir.c
110
dir_bh = affs_bread(sb, inode->i_ino);
fs/affs/dir.c
125
fh_bh = affs_bread(sb, ino);
fs/affs/dir.c
127
affs_error(sb, "readdir","Cannot read block %d", i);
fs/affs/dir.c
131
ino = be32_to_cpu(AFFS_TAIL(sb, fh_bh)->hash_chain);
fs/affs/dir.c
139
for (; hash_pos < AFFS_SB(sb)->s_hashsize; hash_pos++) {
fs/affs/dir.c
146
fh_bh = affs_bread(sb, ino);
fs/affs/dir.c
148
affs_error(sb, "readdir",
fs/affs/dir.c
153
namelen = min(AFFS_TAIL(sb, fh_bh)->name[0],
fs/affs/dir.c
155
name = AFFS_TAIL(sb, fh_bh)->name + 1;
fs/affs/dir.c
162
ino = be32_to_cpu(AFFS_TAIL(sb, fh_bh)->hash_chain);
fs/affs/dir.c
82
struct super_block *sb = inode->i_sb;
fs/affs/file.c
103
bh = affs_bread(sb, key);
fs/affs/file.c
106
key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension);
fs/affs/file.c
123
struct super_block *sb = inode->i_sb;
fs/affs/file.c
131
new_bh = affs_getzeroblk(sb, blocknr);
fs/affs/file.c
133
affs_free_block(sb, blocknr);
fs/affs/file.c
139
AFFS_TAIL(sb, new_bh)->stype = cpu_to_be32(ST_FILE);
fs/affs/file.c
140
AFFS_TAIL(sb, new_bh)->parent = cpu_to_be32(inode->i_ino);
fs/affs/file.c
141
affs_fix_checksum(sb, new_bh);
fs/affs/file.c
145
tmp = be32_to_cpu(AFFS_TAIL(sb, bh)->extension);
fs/affs/file.c
147
affs_warning(sb, "alloc_ext", "previous extension set (%x)", tmp);
fs/affs/file.c
148
AFFS_TAIL(sb, bh)->extension = cpu_to_be32(blocknr);
fs/affs/file.c
175
struct super_block *sb = inode->i_sb;
fs/affs/file.c
184
ext_key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension);
fs/affs/file.c
261
bh = affs_bread(sb, ext_key);
fs/affs/file.c
264
ext_key = be32_to_cpu(AFFS_TAIL(sb, bh)->extension);
fs/affs/file.c
278
bh = affs_bread(sb, ext_key);
fs/affs/file.c
300
struct super_block *sb = inode->i_sb;
fs/affs/file.c
318
ext = (u32)block / AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
319
block -= ext * AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
323
map_bh(bh_result, sb, (sector_t)be32_to_cpu(AFFS_BLOCK(sb, ext_bh, block)));
fs/affs/file.c
330
AFFS_I(inode)->mmu_private += AFFS_SB(sb)->s_data_blksize;
fs/affs/file.c
335
affs_warning(sb, "get_block",
fs/affs/file.c
338
AFFS_BLOCK(sb, ext_bh, block) = cpu_to_be32(blocknr);
fs/affs/file.c
347
affs_warning(sb, "get_block", "first block already set (%d)", tmp);
fs/affs/file.c
53
struct super_block *sb = inode->i_sb;
fs/affs/file.c
531
struct super_block *sb = inode->i_sb;
fs/affs/file.c
540
bsize = AFFS_SB(sb)->s_data_blksize;
fs/affs/file.c
563
struct super_block *sb = inode->i_sb;
fs/affs/file.c
570
bsize = AFFS_SB(sb)->s_data_blksize;
fs/affs/file.c
583
affs_fix_checksum(sb, bh);
fs/affs/file.c
604
affs_fix_checksum(sb, bh);
fs/affs/file.c
611
affs_warning(sb, "extent_file_ofs",
fs/affs/file.c
697
struct super_block *sb = inode->i_sb;
fs/affs/file.c
715
bsize = AFFS_SB(sb)->s_data_blksize;
fs/affs/file.c
734
affs_fix_checksum(sb, bh);
fs/affs/file.c
763
affs_warning(sb, "commit_write_ofs",
fs/affs/file.c
772
affs_fix_checksum(sb, bh);
fs/affs/file.c
797
affs_warning(sb, "commit_write_ofs",
fs/affs/file.c
807
affs_fix_checksum(sb, bh);
fs/affs/file.c
855
struct super_block *sb = inode->i_sb;
fs/affs/file.c
861
affs_free_block(sb, ++AFFS_I(inode)->i_lastalloc);
fs/affs/file.c
870
struct super_block *sb = inode->i_sb;
fs/affs/file.c
883
last_blk = ((u32)inode->i_size - 1) / AFFS_SB(sb)->s_data_blksize;
fs/affs/file.c
884
ext = last_blk / AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
907
affs_warning(sb, "truncate",
fs/affs/file.c
925
ext_key = be32_to_cpu(AFFS_TAIL(sb, ext_bh)->extension);
fs/affs/file.c
931
i = last_blk % AFFS_SB(sb)->s_hashsize + 1;
fs/affs/file.c
936
size = AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
940
affs_free_block(sb, be32_to_cpu(AFFS_BLOCK(sb, ext_bh, i)));
fs/affs/file.c
941
AFFS_BLOCK(sb, ext_bh, i) = 0;
fs/affs/file.c
943
AFFS_TAIL(sb, ext_bh)->extension = 0;
fs/affs/file.c
944
affs_fix_checksum(sb, ext_bh);
fs/affs/file.c
951
if (affs_test_opt(AFFS_SB(sb)->s_flags, SF_OFS)) {
fs/affs/file.c
955
affs_warning(sb, "truncate",
fs/affs/file.c
973
ext_bh = affs_bread(sb, ext_key);
fs/affs/file.c
974
size = AFFS_SB(sb)->s_hashsize;
fs/affs/file.c
978
affs_free_block(sb, be32_to_cpu(AFFS_BLOCK(sb, ext_bh, i)));
fs/affs/file.c
979
affs_free_block(sb, ext_key);
fs/affs/file.c
980
ext_key = be32_to_cpu(AFFS_TAIL(sb, ext_bh)->extension);
fs/affs/inode.c
116
affs_warning(sb, "read_inode", "inode is LINKDIR");
fs/affs/inode.c
124
affs_warning(sb, "read_inode", "inode is LINKFILE");
fs/affs/inode.c
168
struct super_block *sb = inode->i_sb;
fs/affs/inode.c
179
bh = affs_bread(sb, inode->i_ino);
fs/affs/inode.c
18
struct inode *affs_iget(struct super_block *sb, unsigned long ino)
fs/affs/inode.c
181
affs_error(sb,"write_inode","Cannot read block %lu",inode->i_ino);
fs/affs/inode.c
184
tail = AFFS_TAIL(sb, bh);
fs/affs/inode.c
187
&AFFS_ROOT_TAIL(sb, bh)->root_change);
fs/affs/inode.c
193
if (!(inode->i_ino == AFFS_SB(sb)->s_root_block)) {
fs/affs/inode.c
196
if (affs_test_opt(AFFS_SB(sb)->s_flags, SF_MUFS)) {
fs/affs/inode.c
20
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/inode.c
202
if (!affs_test_opt(AFFS_SB(sb)->s_flags, SF_SETUID))
fs/affs/inode.c
204
if (!affs_test_opt(AFFS_SB(sb)->s_flags, SF_SETGID))
fs/affs/inode.c
208
affs_fix_checksum(sb, bh);
fs/affs/inode.c
29
inode = iget_locked(sb, ino);
fs/affs/inode.c
293
struct super_block *sb = dir->i_sb;
fs/affs/inode.c
298
if (!(inode = new_inode(sb)))
fs/affs/inode.c
304
bh = affs_getzeroblk(sb, block);
fs/affs/inode.c
335
affs_free_block(sb, block);
fs/affs/inode.c
350
struct super_block *sb = dir->i_sb;
fs/affs/inode.c
360
bh = affs_bread(sb, inode->i_ino);
fs/affs/inode.c
374
bh = affs_getzeroblk(sb, block);
fs/affs/inode.c
38
bh = affs_bread(sb, block);
fs/affs/inode.c
384
affs_copy_name(AFFS_TAIL(sb, bh)->name, dentry);
fs/affs/inode.c
385
AFFS_TAIL(sb, bh)->stype = cpu_to_be32(type);
fs/affs/inode.c
386
AFFS_TAIL(sb, bh)->parent = cpu_to_be32(dir->i_ino);
fs/affs/inode.c
390
chain = AFFS_TAIL(sb, inode_bh)->link_chain;
fs/affs/inode.c
391
AFFS_TAIL(sb, bh)->original = cpu_to_be32(inode->i_ino);
fs/affs/inode.c
392
AFFS_TAIL(sb, bh)->link_chain = chain;
fs/affs/inode.c
393
AFFS_TAIL(sb, inode_bh)->link_chain = cpu_to_be32(block);
fs/affs/inode.c
399
affs_fix_checksum(sb, bh);
fs/affs/inode.c
40
affs_warning(sb, "read_inode", "Cannot read block %d", block);
fs/affs/inode.c
416
affs_free_block(sb, block);
fs/affs/inode.c
43
if (affs_checksum_block(sb, bh) || be32_to_cpu(AFFS_HEAD(bh)->ptype) != T_SHORT) {
fs/affs/inode.c
44
affs_warning(sb,"read_inode",
fs/affs/inode.c
50
tail = AFFS_TAIL(sb, bh);
fs/affs/namei.c
157
affs_hash_name(struct super_block *sb, const u8 *name, unsigned int len)
fs/affs/namei.c
159
toupper_t fn = affs_get_toupper(sb);
fs/affs/namei.c
166
return hash % AFFS_SB(sb)->s_hashsize;
fs/affs/namei.c
172
struct super_block *sb = dir->i_sb;
fs/affs/namei.c
174
toupper_t fn = affs_get_toupper(sb);
fs/affs/namei.c
179
bh = affs_bread(sb, dir->i_ino);
fs/affs/namei.c
183
key = be32_to_cpu(AFFS_HEAD(bh)->table[affs_hash_name(sb, dentry->d_name.name, dentry->d_name.len)]);
fs/affs/namei.c
189
bh = affs_bread(sb, key);
fs/affs/namei.c
192
if (affs_match(dentry, AFFS_TAIL(sb, bh)->name, fn))
fs/affs/namei.c
194
key = be32_to_cpu(AFFS_TAIL(sb, bh)->hash_chain);
fs/affs/namei.c
201
struct super_block *sb = dir->i_sb;
fs/affs/namei.c
219
switch (be32_to_cpu(AFFS_TAIL(sb, bh)->stype)) {
fs/affs/namei.c
223
ino = be32_to_cpu(AFFS_TAIL(sb, bh)->original);
fs/affs/namei.c
226
inode = affs_iget(sb, ino);
fs/affs/namei.c
248
struct super_block *sb = dir->i_sb;
fs/affs/namei.c
265
inode->i_mapping->a_ops = affs_test_opt(AFFS_SB(sb)->s_flags, SF_OFS) ?
fs/affs/namei.c
319
struct super_block *sb = dir->i_sb;
fs/affs/namei.c
329
maxlen = AFFS_SB(sb)->s_hashsize * sizeof(u32) - 1;
fs/affs/namei.c
341
bh = affs_bread(sb, inode->i_ino);
fs/affs/namei.c
348
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/namei.c
36
affs_get_toupper(struct super_block *sb)
fs/affs/namei.c
38
return affs_test_opt(AFFS_SB(sb)->s_flags, SF_INTL) ?
fs/affs/namei.c
408
struct super_block *sb = old_dir->i_sb;
fs/affs/namei.c
426
bh = affs_bread(sb, d_inode(old_dentry)->i_ino);
fs/affs/namei.c
438
affs_copy_name(AFFS_TAIL(sb, bh)->name, new_dentry);
fs/affs/namei.c
439
affs_fix_checksum(sb, bh);
fs/affs/namei.c
456
struct super_block *sb = old_dir->i_sb;
fs/affs/namei.c
461
bh_old = affs_bread(sb, d_inode(old_dentry)->i_ino);
fs/affs/namei.c
465
bh_new = affs_bread(sb, d_inode(new_dentry)->i_ino);
fs/affs/namei.c
486
affs_copy_name(AFFS_TAIL(sb, bh_old)->name, new_dentry);
fs/affs/namei.c
487
affs_fix_checksum(sb, bh_old);
fs/affs/namei.c
493
affs_copy_name(AFFS_TAIL(sb, bh_new)->name, old_dentry);
fs/affs/namei.c
494
affs_fix_checksum(sb, bh_new);
fs/affs/namei.c
538
static struct inode *affs_nfs_get_inode(struct super_block *sb, u64 ino,
fs/affs/namei.c
543
if (!affs_validblock(sb, ino))
fs/affs/namei.c
546
inode = affs_iget(sb, ino);
fs/affs/namei.c
553
static struct dentry *affs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/affs/namei.c
556
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/affs/namei.c
560
static struct dentry *affs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/affs/namei.c
563
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/affs/super.c
103
i = alloc_inode_sb(sb, affs_inode_cachep, GFP_KERNEL);
fs/affs/super.c
270
struct super_block *sb = root->d_sb;
fs/affs/super.c
271
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
273
if (sb->s_blocksize)
fs/affs/super.c
274
seq_printf(m, ",bs=%lu", sb->s_blocksize);
fs/affs/super.c
306
static int affs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/affs/super.c
322
sb->s_magic = AFFS_SUPER_MAGIC;
fs/affs/super.c
323
sb->s_op = &affs_sops;
fs/affs/super.c
324
sb->s_flags |= SB_NODIRATIME;
fs/affs/super.c
326
sb->s_time_gran = NSEC_PER_SEC;
fs/affs/super.c
327
sb->s_time_min = sys_tz.tz_minuteswest * 60 + AFFS_EPOCH_DELTA;
fs/affs/super.c
328
sb->s_time_max = 86400LL * U32_MAX + 86400 + sb->s_time_min;
fs/affs/super.c
33
affs_commit_super(struct super_block *sb, int wait)
fs/affs/super.c
334
sb->s_fs_info = sbi;
fs/affs/super.c
335
sbi->sb = sb;
fs/affs/super.c
35
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
357
size = bdev_nr_sectors(sb->s_bdev);
fs/affs/super.c
360
affs_set_blocksize(sb, PAGE_SIZE);
fs/affs/super.c
363
i = bdev_logical_block_size(sb->s_bdev);
fs/affs/super.c
37
struct affs_root_tail *tail = AFFS_ROOT_TAIL(sb, bh);
fs/affs/super.c
376
affs_set_blocksize(sb, blocksize);
fs/affs/super.c
392
sb->s_id,
fs/affs/super.c
395
root_bh = affs_bread(sb, sbi->s_root_block + num_bm);
fs/affs/super.c
398
if (!affs_checksum_block(sb, root_bh) &&
fs/affs/super.c
400
be32_to_cpu(AFFS_ROOT_TAIL(sb, root_bh)->stype) == ST_ROOT) {
fs/affs/super.c
41
affs_fix_checksum(sb, bh);
fs/affs/super.c
410
pr_err("No valid root block on device %s\n", sb->s_id);
fs/affs/super.c
420
boot_bh = sb_bread(sb, 0);
fs/affs/super.c
434
|| chksum == MUFS_DCOFS) && !sb_rdonly(sb)) {
fs/affs/super.c
435
pr_notice("Dircache FS - mounting %s read only\n", sb->s_id);
fs/affs/super.c
436
sb->s_flags |= SB_RDONLY;
fs/affs/super.c
458
sb->s_flags |= SB_NOEXEC;
fs/affs/super.c
468
sb->s_flags |= SB_NOEXEC;
fs/affs/super.c
472
sb->s_id, chksum);
fs/affs/super.c
477
u8 len = AFFS_ROOT_TAIL(sb, root_bh)->disk_name[0];
fs/affs/super.c
480
AFFS_ROOT_TAIL(sb, root_bh)->disk_name + 1,
fs/affs/super.c
484
sb->s_flags |= SB_NODEV | SB_NOSUID;
fs/affs/super.c
486
sbi->s_data_blksize = sb->s_blocksize;
fs/affs/super.c
490
tmp_flags = sb->s_flags;
fs/affs/super.c
491
ret = affs_init_bitmap(sb, &tmp_flags);
fs/affs/super.c
494
sb->s_flags = tmp_flags;
fs/affs/super.c
498
root_inode = affs_iget(sb, ctx->root_block);
fs/affs/super.c
50
affs_put_super(struct super_block *sb)
fs/affs/super.c
502
if (affs_test_opt(AFFS_SB(sb)->s_flags, SF_INTL))
fs/affs/super.c
503
set_default_d_op(sb, &affs_intl_dentry_operations);
fs/affs/super.c
505
set_default_d_op(sb, &affs_dentry_operations);
fs/affs/super.c
507
sb->s_root = d_make_root(root_inode);
fs/affs/super.c
508
if (!sb->s_root) {
fs/affs/super.c
513
sb->s_export_op = &affs_export_ops;
fs/affs/super.c
514
pr_debug("s_flags=%lX\n", sb->s_flags);
fs/affs/super.c
52
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
520
struct super_block *sb = fc->root->d_sb;
fs/affs/super.c
522
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
525
sync_filesystem(sb);
fs/affs/super.c
548
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb))
fs/affs/super.c
552
affs_free_bitmap(sb);
fs/affs/super.c
554
res = affs_init_bitmap(sb, &fc->sb_flags);
fs/affs/super.c
562
struct super_block *sb = dentry->d_sb;
fs/affs/super.c
564
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/affs/super.c
567
__func__, AFFS_SB(sb)->s_partition_size,
fs/affs/super.c
568
AFFS_SB(sb)->s_reserved);
fs/affs/super.c
570
free = affs_count_free_blocks(sb);
fs/affs/super.c
572
buf->f_bsize = sb->s_blocksize;
fs/affs/super.c
573
buf->f_blocks = AFFS_SB(sb)->s_partition_size - AFFS_SB(sb)->s_reserved;
fs/affs/super.c
586
static void affs_kill_sb(struct super_block *sb)
fs/affs/super.c
588
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
589
kill_block_super(sb);
fs/affs/super.c
59
affs_sync_fs(struct super_block *sb, int wait)
fs/affs/super.c
591
affs_free_bitmap(sb);
fs/affs/super.c
61
affs_commit_super(sb, wait);
fs/affs/super.c
623
struct super_block *sb = fc->root->d_sb;
fs/affs/super.c
624
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
68
struct super_block *sb;
fs/affs/super.c
71
sb = sbi->sb;
fs/affs/super.c
77
affs_commit_super(sb, 1);
fs/affs/super.c
80
void affs_mark_sb_dirty(struct super_block *sb)
fs/affs/super.c
82
struct affs_sb_info *sbi = AFFS_SB(sb);
fs/affs/super.c
85
if (sb_rdonly(sb))
fs/affs/super.c
99
static struct inode *affs_alloc_inode(struct super_block *sb)
fs/afs/callback.c
192
struct super_block *sb;
fs/afs/callback.c
200
sb = rcu_dereference(volume->sb);
fs/afs/callback.c
201
if (!sb)
fs/afs/callback.c
204
inode = find_inode_rcu(sb, fid->vnode, afs_ilookup5_test_by_fid, fid);
fs/afs/dynroot.c
369
struct inode *afs_dynroot_iget_root(struct super_block *sb)
fs/afs/dynroot.c
371
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/dynroot.c
379
inode = iget5_locked(sb, fid.vnode,
fs/afs/dynroot.c
47
static struct inode *afs_iget_pseudo_dir(struct super_block *sb, ino_t ino)
fs/afs/dynroot.c
55
inode = iget5_locked(sb, fid.vnode,
fs/afs/inode.c
563
struct super_block *sb = dvp->vnode->netfs.inode.i_sb;
fs/afs/inode.c
570
inode = iget5_locked(sb, vp->fid.vnode, afs_iget5_test, afs_iget5_set, vp);
fs/afs/inode.c
624
struct inode *afs_root_iget(struct super_block *sb, struct key *key)
fs/afs/inode.c
626
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/inode.c
634
inode = iget5_locked(sb, 1, NULL, afs_iget5_set_root, NULL);
fs/afs/internal.h
1144
struct inode *afs_dynroot_iget_root(struct super_block *sb);
fs/afs/internal.h
1276
static inline struct afs_net *afs_sb2net(struct super_block *sb)
fs/afs/internal.h
1278
return afs_net(AFS_FS_S(sb)->net_ns);
fs/afs/internal.h
257
static inline struct afs_super_info *AFS_FS_S(struct super_block *sb)
fs/afs/internal.h
259
return sb->s_fs_info;
fs/afs/internal.h
632
struct super_block __rcu *sb; /* Superblock on which inodes reside */
fs/afs/super.c
33
static void afs_kill_super(struct super_block *sb);
fs/afs/super.c
34
static struct inode *afs_alloc_inode(struct super_block *sb);
fs/afs/super.c
420
static int afs_test_super(struct super_block *sb, struct fs_context *fc)
fs/afs/super.c
423
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/super.c
432
static int afs_dynroot_test_super(struct super_block *sb, struct fs_context *fc)
fs/afs/super.c
434
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/super.c
440
static int afs_set_super(struct super_block *sb, struct fs_context *fc)
fs/afs/super.c
442
return set_anon_super(sb, NULL);
fs/afs/super.c
448
static int afs_fill_super(struct super_block *sb, struct afs_fs_context *ctx)
fs/afs/super.c
450
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/super.c
457
sb->s_blocksize = PAGE_SIZE;
fs/afs/super.c
458
sb->s_blocksize_bits = PAGE_SHIFT;
fs/afs/super.c
459
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/afs/super.c
460
sb->s_magic = AFS_FS_MAGIC;
fs/afs/super.c
461
sb->s_op = &afs_super_ops;
fs/afs/super.c
463
sb->s_xattr = afs_xattr_handlers;
fs/afs/super.c
464
ret = super_setup_bdi(sb);
fs/afs/super.c
470
inode = afs_dynroot_iget_root(sb);
fs/afs/super.c
472
sprintf(sb->s_id, "%llu", as->volume->vid);
fs/afs/super.c
474
inode = afs_root_iget(sb, ctx->key);
fs/afs/super.c
481
sb->s_root = d_make_root(inode);
fs/afs/super.c
482
if (!sb->s_root)
fs/afs/super.c
486
set_default_d_op(sb, &afs_dynroot_dentry_operations);
fs/afs/super.c
488
set_default_d_op(sb, &afs_fs_dentry_operations);
fs/afs/super.c
489
rcu_assign_pointer(as->volume->sb, sb);
fs/afs/super.c
530
static void afs_kill_super(struct super_block *sb)
fs/afs/super.c
532
struct afs_super_info *as = AFS_FS_S(sb);
fs/afs/super.c
538
rcu_assign_pointer(as->volume->sb, NULL);
fs/afs/super.c
539
kill_anon_super(sb);
fs/afs/super.c
551
struct super_block *sb;
fs/afs/super.c
569
sb = sget_fc(fc,
fs/afs/super.c
572
if (IS_ERR(sb)) {
fs/afs/super.c
573
ret = PTR_ERR(sb);
fs/afs/super.c
577
if (!sb->s_root) {
fs/afs/super.c
580
ret = afs_fill_super(sb, ctx);
fs/afs/super.c
583
sb->s_flags |= SB_ACTIVE;
fs/afs/super.c
586
ASSERTCMP(sb->s_flags, &, SB_ACTIVE);
fs/afs/super.c
589
fc->root = dget(sb->s_root);
fs/afs/super.c
591
_leave(" = 0 [%p]", sb);
fs/afs/super.c
595
deactivate_locked_super(sb);
fs/afs/super.c
670
static struct inode *afs_alloc_inode(struct super_block *sb)
fs/afs/super.c
674
vnode = alloc_inode_sb(sb, afs_inode_cachep, GFP_KERNEL);
fs/anon_inodes.c
113
struct inode *anon_inode_make_secure_inode(struct super_block *sb, const char *name,
fs/anon_inodes.c
119
inode = alloc_anon_inode(sb);
fs/autofs/autofs_i.h
129
struct super_block *sb;
fs/autofs/autofs_i.h
140
static inline struct autofs_sb_info *autofs_sbi(struct super_block *sb)
fs/autofs/autofs_i.h
142
return (struct autofs_sb_info *)(sb->s_fs_info);
fs/autofs/autofs_i.h
174
int autofs_do_expire_multi(struct super_block *sb, struct vfsmount *mnt,
fs/autofs/autofs_i.h
261
return new_encode_dev(sbi->sb->s_dev);
fs/autofs/autofs_i.h
266
return d_inode(sbi->sb->s_root)->i_ino;
fs/autofs/dev-ioctl.c
500
devid = sbi->sb->s_dev;
fs/autofs/dev-ioctl.c
538
return autofs_do_expire_multi(sbi->sb, mnt, sbi, how);
fs/autofs/dev-ioctl.c
607
dev_t dev = sbi->sb->s_dev;
fs/autofs/dev-ioctl.c
716
struct super_block *sb;
fs/autofs/dev-ioctl.c
726
sb = file_inode(fp)->i_sb;
fs/autofs/dev-ioctl.c
727
if (sb->s_type != &autofs_fs_type) {
fs/autofs/dev-ioctl.c
732
sbi = autofs_sbi(sb);
fs/autofs/expire.c
271
static struct dentry *autofs_expire_direct(struct super_block *sb,
fs/autofs/expire.c
276
struct dentry *root = dget(sb->s_root);
fs/autofs/expire.c
417
static struct dentry *autofs_expire_indirect(struct super_block *sb,
fs/autofs/expire.c
423
struct dentry *root = sb->s_root;
fs/autofs/expire.c
536
int autofs_expire_run(struct super_block *sb,
fs/autofs/expire.c
551
dentry = autofs_expire_indirect(sb, mnt, sbi, 0);
fs/autofs/expire.c
575
int autofs_do_expire_multi(struct super_block *sb, struct vfsmount *mnt,
fs/autofs/expire.c
582
dentry = autofs_expire_direct(sb, mnt, sbi, how);
fs/autofs/expire.c
584
dentry = autofs_expire_indirect(sb, mnt, sbi, how);
fs/autofs/expire.c
611
int autofs_expire_multi(struct super_block *sb, struct vfsmount *mnt,
fs/autofs/expire.c
619
return autofs_do_expire_multi(sb, mnt, sbi, how);
fs/autofs/inode.c
310
sbi->sb = s;
fs/autofs/inode.c
41
void autofs_kill_sb(struct super_block *sb)
fs/autofs/inode.c
43
struct autofs_sb_info *sbi = autofs_sbi(sb);
fs/autofs/inode.c
430
struct inode *autofs_get_inode(struct super_block *sb, umode_t mode)
fs/autofs/inode.c
432
struct inode *inode = new_inode(sb);
fs/autofs/inode.c
438
if (sb->s_root) {
fs/autofs/inode.c
439
inode->i_uid = d_inode(sb->s_root)->i_uid;
fs/autofs/inode.c
440
inode->i_gid = d_inode(sb->s_root)->i_gid;
fs/autofs/inode.c
58
kill_anon_super(sb);
fs/backing-file.c
152
struct super_block *sb = file_inode(iocb->ki_filp)->i_sb;
fs/backing-file.c
154
if (sb->s_dio_done_wq)
fs/backing-file.c
157
return sb_init_dio_done_wq(sb);
fs/befs/befs.h
100
void befs_dump_super_block(const struct super_block *sb, befs_super_block *);
fs/befs/befs.h
101
void befs_dump_inode(const struct super_block *sb, befs_inode *);
fs/befs/befs.h
102
void befs_dump_index_entry(const struct super_block *sb, befs_disk_btree_super *);
fs/befs/befs.h
103
void befs_dump_index_node(const struct super_block *sb, befs_btree_nodehead *);
fs/befs/befs.h
123
iaddr2blockno(struct super_block *sb, const befs_inode_addr *iaddr)
fs/befs/befs.h
125
return ((iaddr->allocation_group << BEFS_SB(sb)->ag_shift) +
fs/befs/befs.h
130
blockno2iaddr(struct super_block *sb, befs_blocknr_t blockno)
fs/befs/befs.h
134
iaddr.allocation_group = blockno >> BEFS_SB(sb)->ag_shift;
fs/befs/befs.h
136
blockno - (iaddr.allocation_group << BEFS_SB(sb)->ag_shift);
fs/befs/befs.h
143
befs_iaddrs_per_block(struct super_block *sb)
fs/befs/befs.h
145
return BEFS_SB(sb)->block_size / sizeof(befs_disk_inode_addr);
fs/befs/befs.h
94
void befs_error(const struct super_block *sb, const char *fmt, ...);
fs/befs/befs.h
96
void befs_warning(const struct super_block *sb, const char *fmt, ...);
fs/befs/befs.h
98
void befs_debug(const struct super_block *sb, const char *fmt, ...);
fs/befs/btree.c
111
static int befs_find_key(struct super_block *sb,
fs/befs/btree.c
115
static char *befs_bt_get_key(struct super_block *sb,
fs/befs/btree.c
134
befs_bt_read_super(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
140
befs_debug(sb, "---> %s", __func__);
fs/befs/btree.c
142
bh = befs_read_datastream(sb, ds, 0, NULL);
fs/befs/btree.c
145
befs_error(sb, "Couldn't read index header.");
fs/befs/btree.c
149
befs_dump_index_entry(sb, od_sup);
fs/befs/btree.c
151
sup->magic = fs32_to_cpu(sb, od_sup->magic);
fs/befs/btree.c
152
sup->node_size = fs32_to_cpu(sb, od_sup->node_size);
fs/befs/btree.c
153
sup->max_depth = fs32_to_cpu(sb, od_sup->max_depth);
fs/befs/btree.c
154
sup->data_type = fs32_to_cpu(sb, od_sup->data_type);
fs/befs/btree.c
155
sup->root_node_ptr = fs64_to_cpu(sb, od_sup->root_node_ptr);
fs/befs/btree.c
159
befs_error(sb, "Index header has bad magic.");
fs/befs/btree.c
163
befs_debug(sb, "<--- %s", __func__);
fs/befs/btree.c
167
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
191
befs_bt_read_node(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
196
befs_debug(sb, "---> %s", __func__);
fs/befs/btree.c
201
node->bh = befs_read_datastream(sb, ds, node_off, &off);
fs/befs/btree.c
203
befs_error(sb, "%s failed to read "
fs/befs/btree.c
205
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
212
befs_dump_index_node(sb, node->od_node);
fs/befs/btree.c
214
node->head.left = fs64_to_cpu(sb, node->od_node->left);
fs/befs/btree.c
215
node->head.right = fs64_to_cpu(sb, node->od_node->right);
fs/befs/btree.c
216
node->head.overflow = fs64_to_cpu(sb, node->od_node->overflow);
fs/befs/btree.c
218
fs16_to_cpu(sb, node->od_node->all_key_count);
fs/befs/btree.c
220
fs16_to_cpu(sb, node->od_node->all_key_length);
fs/befs/btree.c
222
befs_debug(sb, "<--- %s", __func__);
fs/befs/btree.c
245
befs_btree_find(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
253
befs_debug(sb, "---> %s Key: %s", __func__, key);
fs/befs/btree.c
255
if (befs_bt_read_super(sb, ds, &bt_super) != BEFS_OK) {
fs/befs/btree.c
256
befs_error(sb,
fs/befs/btree.c
263
befs_error(sb, "befs_btree_find() failed to allocate %zu "
fs/befs/btree.c
272
if (befs_bt_read_node(sb, ds, this_node, node_off) != BEFS_OK) {
fs/befs/btree.c
273
befs_error(sb, "befs_btree_find() failed to read "
fs/befs/btree.c
279
res = befs_find_key(sb, this_node, key, &node_off);
fs/befs/btree.c
283
if (befs_bt_read_node(sb, ds, this_node, node_off) != BEFS_OK) {
fs/befs/btree.c
284
befs_error(sb, "befs_btree_find() failed to read "
fs/befs/btree.c
291
res = befs_find_key(sb, this_node, key, value);
fs/befs/btree.c
297
befs_error(sb, "<--- %s Key %s not found", __func__, key);
fs/befs/btree.c
298
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
302
befs_debug(sb, "<--- %s Found key %s, value %llu", __func__,
fs/befs/btree.c
310
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
329
befs_find_key(struct super_block *sb, struct befs_btree_node *node,
fs/befs/btree.c
339
befs_debug(sb, "---> %s %s", __func__, findkey);
fs/befs/btree.c
345
thiskey = befs_bt_get_key(sb, node, last, &keylen);
fs/befs/btree.c
349
befs_debug(sb, "<--- node can't contain %s", findkey);
fs/befs/btree.c
360
befs_debug(sb, "first: %d, last: %d, mid: %d", first, last,
fs/befs/btree.c
362
thiskey = befs_bt_get_key(sb, node, mid, &keylen);
fs/befs/btree.c
367
befs_debug(sb, "<--- %s found %s at %d",
fs/befs/btree.c
370
*value = fs64_to_cpu(sb, valarray[mid]);
fs/befs/btree.c
381
*value = fs64_to_cpu(sb, valarray[mid + 1]);
fs/befs/btree.c
383
*value = fs64_to_cpu(sb, valarray[mid]);
fs/befs/btree.c
384
befs_error(sb, "<--- %s %s not found", __func__, findkey);
fs/befs/btree.c
385
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
410
befs_btree_read(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
425
befs_debug(sb, "---> %s", __func__);
fs/befs/btree.c
427
if (befs_bt_read_super(sb, ds, &bt_super) != BEFS_OK) {
fs/befs/btree.c
428
befs_error(sb,
fs/befs/btree.c
435
befs_error(sb, "befs_btree_read() failed to allocate %zu "
fs/befs/btree.c
444
res = befs_btree_seekleaf(sb, ds, &bt_super, this_node, &node_off);
fs/befs/btree.c
450
befs_debug(sb, "<--- %s Tree is EMPTY", __func__);
fs/befs/btree.c
464
befs_debug(sb,
fs/befs/btree.c
476
if (befs_bt_read_node(sb, ds, this_node, node_off) != BEFS_OK) {
fs/befs/btree.c
477
befs_error(sb, "%s failed to read node at %llu",
fs/befs/btree.c
489
keystart = befs_bt_get_key(sb, this_node, cur_key, &keylen);
fs/befs/btree.c
491
befs_debug(sb, "Read [%llu,%d]: keysize %d",
fs/befs/btree.c
496
befs_error(sb, "%s keybuf too small (%zu) "
fs/befs/btree.c
503
*value = fs64_to_cpu(sb, valarray[cur_key]);
fs/befs/btree.c
506
befs_debug(sb, "Read [%llu,%d]: Key \"%.*s\", Value %llu", node_off,
fs/befs/btree.c
512
befs_debug(sb, "<--- %s", __func__);
fs/befs/btree.c
522
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
541
befs_btree_seekleaf(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
547
befs_debug(sb, "---> %s", __func__);
fs/befs/btree.c
549
if (befs_bt_read_node(sb, ds, this_node, *node_off) != BEFS_OK) {
fs/befs/btree.c
550
befs_error(sb, "%s failed to read "
fs/befs/btree.c
554
befs_debug(sb, "Seekleaf to root node %llu", *node_off);
fs/befs/btree.c
557
befs_debug(sb, "<--- %s Tree is EMPTY", __func__);
fs/befs/btree.c
564
befs_debug(sb, "%s encountered "
fs/befs/btree.c
571
*node_off = fs64_to_cpu(sb, valarray[0]);
fs/befs/btree.c
573
if (befs_bt_read_node(sb, ds, this_node, *node_off) != BEFS_OK) {
fs/befs/btree.c
574
befs_error(sb, "%s failed to read "
fs/befs/btree.c
579
befs_debug(sb, "Seekleaf to child node %llu", *node_off);
fs/befs/btree.c
581
befs_debug(sb, "Node %llu is a leaf node", *node_off);
fs/befs/btree.c
586
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/btree.c
674
befs_bt_get_key(struct super_block *sb, struct befs_btree_node *node,
fs/befs/btree.c
692
prev_key_end = fs16_to_cpu(sb, keylen_index[index - 1]);
fs/befs/btree.c
694
*keylen = fs16_to_cpu(sb, keylen_index[index]) - prev_key_end;
fs/befs/btree.c
91
static int befs_btree_seekleaf(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
96
static int befs_bt_read_super(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.c
99
static int befs_bt_read_node(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.h
10
int befs_btree_read(struct super_block *sb, const befs_data_stream *ds,
fs/befs/btree.h
7
int befs_btree_find(struct super_block *sb, const befs_data_stream *ds,
fs/befs/datastream.c
102
err = befs_find_brun_direct(sb, data, fblock, run);
fs/befs/datastream.c
105
err = befs_find_brun_indirect(sb, data, fblock, run);
fs/befs/datastream.c
108
err = befs_find_brun_dblindirect(sb, data, fblock, run);
fs/befs/datastream.c
111
befs_error(sb,
fs/befs/datastream.c
130
befs_read_lsymlink(struct super_block *sb, const befs_data_stream *ds,
fs/befs/datastream.c
137
befs_debug(sb, "---> %s length: %llu", __func__, len);
fs/befs/datastream.c
140
bh = befs_read_datastream(sb, ds, bytes_read, NULL);
fs/befs/datastream.c
142
befs_error(sb, "BeFS: Error reading datastream block "
fs/befs/datastream.c
144
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/datastream.c
148
plen = ((bytes_read + BEFS_SB(sb)->block_size) < len) ?
fs/befs/datastream.c
149
BEFS_SB(sb)->block_size : len - bytes_read;
fs/befs/datastream.c
155
befs_debug(sb, "<--- %s read %u bytes", __func__, (unsigned int)
fs/befs/datastream.c
172
befs_count_blocks(struct super_block *sb, const befs_data_stream *ds)
fs/befs/datastream.c
177
struct befs_sb_info *befs_sb = BEFS_SB(sb);
fs/befs/datastream.c
179
befs_debug(sb, "---> %s", __func__);
fs/befs/datastream.c
209
indirblocks = dbl_bruns / befs_iaddrs_per_block(sb);
fs/befs/datastream.c
216
befs_debug(sb, "<--- %s %u blocks", __func__, (unsigned int)blocks);
fs/befs/datastream.c
24
static int befs_find_brun_direct(struct super_block *sb,
fs/befs/datastream.c
251
befs_find_brun_direct(struct super_block *sb, const befs_data_stream *data,
fs/befs/datastream.c
258
befs_debug(sb, "---> %s, find %lu", __func__, (unsigned long)blockno);
fs/befs/datastream.c
269
befs_debug(sb, "---> %s, "
fs/befs/datastream.c
276
befs_error(sb, "%s failed to find file block %lu", __func__,
fs/befs/datastream.c
278
befs_debug(sb, "---> %s ERROR", __func__);
fs/befs/datastream.c
28
static int befs_find_brun_indirect(struct super_block *sb,
fs/befs/datastream.c
305
befs_find_brun_indirect(struct super_block *sb,
fs/befs/datastream.c
318
befs_blocknr_t indirblockno = iaddr2blockno(sb, &indirect);
fs/befs/datastream.c
319
int arraylen = befs_iaddrs_per_block(sb);
fs/befs/datastream.c
321
befs_debug(sb, "---> %s, find %lu", __func__, (unsigned long)blockno);
fs/befs/datastream.c
323
indir_start_blk = data->max_direct_range >> BEFS_SB(sb)->block_shift;
fs/befs/datastream.c
328
indirblock = sb_bread(sb, indirblockno + i);
fs/befs/datastream.c
33
static int befs_find_brun_dblindirect(struct super_block *sb,
fs/befs/datastream.c
330
befs_error(sb, "---> %s failed to read "
fs/befs/datastream.c
333
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/datastream.c
340
int len = fs16_to_cpu(sb, array[j].len);
fs/befs/datastream.c
345
fs32_to_cpu(sb, array[j].allocation_group);
fs/befs/datastream.c
347
fs16_to_cpu(sb, array[j].start) + offset;
fs/befs/datastream.c
349
fs16_to_cpu(sb, array[j].len) - offset;
fs/befs/datastream.c
352
befs_debug(sb,
fs/befs/datastream.c
366
befs_error(sb, "BeFS: %s failed to find "
fs/befs/datastream.c
369
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/datastream.c
414
befs_find_brun_dblindirect(struct super_block *sb,
fs/befs/datastream.c
434
data->max_indirect_range >> BEFS_SB(sb)->block_shift;
fs/befs/datastream.c
446
size_t diblklen = iblklen * befs_iaddrs_per_block(sb)
fs/befs/datastream.c
449
befs_debug(sb, "---> %s find %lu", __func__, (unsigned long)blockno);
fs/befs/datastream.c
462
dbl_which_block = dblindir_indx / befs_iaddrs_per_block(sb);
fs/befs/datastream.c
464
befs_error(sb, "The double-indirect index calculated by "
fs/befs/datastream.c
472
sb_bread(sb, iaddr2blockno(sb, &data->double_indirect) +
fs/befs/datastream.c
475
befs_error(sb, "%s couldn't read the "
fs/befs/datastream.c
478
iaddr2blockno(sb, &data->double_indirect) +
fs/befs/datastream.c
484
dblindir_indx - (dbl_which_block * befs_iaddrs_per_block(sb));
fs/befs/datastream.c
486
indir_run = fsrun_to_cpu(sb, iaddr_array[dbl_block_indx]);
fs/befs/datastream.c
49
befs_read_datastream(struct super_block *sb, const befs_data_stream *ds,
fs/befs/datastream.c
490
which_block = indir_indx / befs_iaddrs_per_block(sb);
fs/befs/datastream.c
492
befs_error(sb, "The indirect index calculated by "
fs/befs/datastream.c
499
sb_bread(sb, iaddr2blockno(sb, &indir_run) + which_block);
fs/befs/datastream.c
501
befs_error(sb, "%s couldn't read the indirect block "
fs/befs/datastream.c
503
iaddr2blockno(sb, &indir_run) + which_block);
fs/befs/datastream.c
507
block_indx = indir_indx - (which_block * befs_iaddrs_per_block(sb));
fs/befs/datastream.c
509
*run = fsrun_to_cpu(sb, iaddr_array[block_indx]);
fs/befs/datastream.c
520
befs_debug(sb, "Found file block %lu in double_indirect[%d][%d],"
fs/befs/datastream.c
56
befs_debug(sb, "---> %s %llu", __func__, pos);
fs/befs/datastream.c
57
block = pos >> BEFS_SB(sb)->block_shift;
fs/befs/datastream.c
59
*off = pos - (block << BEFS_SB(sb)->block_shift);
fs/befs/datastream.c
61
if (befs_fblock2brun(sb, ds, block, &run) != BEFS_OK) {
fs/befs/datastream.c
62
befs_error(sb, "BeFS: Error finding disk addr of block %lu",
fs/befs/datastream.c
64
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/datastream.c
67
bh = befs_bread_iaddr(sb, run);
fs/befs/datastream.c
69
befs_error(sb, "BeFS: Error reading block %lu from datastream",
fs/befs/datastream.c
74
befs_debug(sb, "<--- %s read data, starting at %llu", __func__, pos);
fs/befs/datastream.c
95
befs_fblock2brun(struct super_block *sb, const befs_data_stream *data,
fs/befs/datastream.c
99
befs_off_t pos = fblock << BEFS_SB(sb)->block_shift;
fs/befs/datastream.h
11
int befs_fblock2brun(struct super_block *sb, const befs_data_stream *data,
fs/befs/datastream.h
14
size_t befs_read_lsymlink(struct super_block *sb, const befs_data_stream *data,
fs/befs/datastream.h
17
befs_blocknr_t befs_count_blocks(struct super_block *sb,
fs/befs/datastream.h
7
struct buffer_head *befs_read_datastream(struct super_block *sb,
fs/befs/debug.c
100
befs_debug(sb, " attributes [%u, %hu, %hu]",
fs/befs/debug.c
103
befs_debug(sb, " type %08x", fs32_to_cpu(sb, inode->type));
fs/befs/debug.c
104
befs_debug(sb, " inode_size %u", fs32_to_cpu(sb, inode->inode_size));
fs/befs/debug.c
106
if (S_ISLNK(fs32_to_cpu(sb, inode->mode))) {
fs/befs/debug.c
107
befs_debug(sb, " Symbolic link [%s]", inode->data.symlink);
fs/befs/debug.c
113
fsrun_to_cpu(sb, inode->data.datastream.direct[i]);
fs/befs/debug.c
114
befs_debug(sb, " direct %d [%u, %hu, %hu]", i,
fs/befs/debug.c
118
befs_debug(sb, " max_direct_range %llu",
fs/befs/debug.c
119
fs64_to_cpu(sb,
fs/befs/debug.c
123
tmp_run = fsrun_to_cpu(sb, inode->data.datastream.indirect);
fs/befs/debug.c
124
befs_debug(sb, " indirect [%u, %hu, %hu]",
fs/befs/debug.c
128
befs_debug(sb, " max_indirect_range %llu",
fs/befs/debug.c
129
fs64_to_cpu(sb,
fs/befs/debug.c
134
fsrun_to_cpu(sb, inode->data.datastream.double_indirect);
fs/befs/debug.c
135
befs_debug(sb, " double indirect [%u, %hu, %hu]",
fs/befs/debug.c
139
befs_debug(sb, " max_double_indirect_range %llu",
fs/befs/debug.c
140
fs64_to_cpu(sb,
fs/befs/debug.c
144
befs_debug(sb, " size %llu",
fs/befs/debug.c
145
fs64_to_cpu(sb, inode->data.datastream.size));
fs/befs/debug.c
156
befs_dump_super_block(const struct super_block *sb, befs_super_block *sup)
fs/befs/debug.c
162
befs_debug(sb, "befs_super_block information");
fs/befs/debug.c
164
befs_debug(sb, " name %s", sup->name);
fs/befs/debug.c
165
befs_debug(sb, " magic1 %08x", fs32_to_cpu(sb, sup->magic1));
fs/befs/debug.c
166
befs_debug(sb, " fs_byte_order %08x",
fs/befs/debug.c
167
fs32_to_cpu(sb, sup->fs_byte_order));
fs/befs/debug.c
169
befs_debug(sb, " block_size %u", fs32_to_cpu(sb, sup->block_size));
fs/befs/debug.c
170
befs_debug(sb, " block_shift %u", fs32_to_cpu(sb, sup->block_shift));
fs/befs/debug.c
172
befs_debug(sb, " num_blocks %llu", fs64_to_cpu(sb, sup->num_blocks));
fs/befs/debug.c
173
befs_debug(sb, " used_blocks %llu", fs64_to_cpu(sb, sup->used_blocks));
fs/befs/debug.c
174
befs_debug(sb, " inode_size %u", fs32_to_cpu(sb, sup->inode_size));
fs/befs/debug.c
176
befs_debug(sb, " magic2 %08x", fs32_to_cpu(sb, sup->magic2));
fs/befs/debug.c
177
befs_debug(sb, " blocks_per_ag %u",
fs/befs/debug.c
178
fs32_to_cpu(sb, sup->blocks_per_ag));
fs/befs/debug.c
179
befs_debug(sb, " ag_shift %u", fs32_to_cpu(sb, sup->ag_shift));
fs/befs/debug.c
180
befs_debug(sb, " num_ags %u", fs32_to_cpu(sb, sup->num_ags));
fs/befs/debug.c
182
befs_debug(sb, " flags %08x", fs32_to_cpu(sb, sup->flags));
fs/befs/debug.c
184
tmp_run = fsrun_to_cpu(sb, sup->log_blocks);
fs/befs/debug.c
185
befs_debug(sb, " log_blocks %u, %hu, %hu",
fs/befs/debug.c
188
befs_debug(sb, " log_start %lld", fs64_to_cpu(sb, sup->log_start));
fs/befs/debug.c
189
befs_debug(sb, " log_end %lld", fs64_to_cpu(sb, sup->log_end));
fs/befs/debug.c
191
befs_debug(sb, " magic3 %08x", fs32_to_cpu(sb, sup->magic3));
fs/befs/debug.c
193
tmp_run = fsrun_to_cpu(sb, sup->root_dir);
fs/befs/debug.c
194
befs_debug(sb, " root_dir %u, %hu, %hu",
fs/befs/debug.c
197
tmp_run = fsrun_to_cpu(sb, sup->indices);
fs/befs/debug.c
198
befs_debug(sb, " indices %u, %hu, %hu",
fs/befs/debug.c
207
befs_dump_small_data(const struct super_block *sb, befs_small_data *sd)
fs/befs/debug.c
213
befs_dump_run(const struct super_block *sb, befs_disk_block_run run)
fs/befs/debug.c
217
befs_block_run n = fsrun_to_cpu(sb, run);
fs/befs/debug.c
219
befs_debug(sb, "[%u, %hu, %hu]", n.allocation_group, n.start, n.len);
fs/befs/debug.c
226
befs_dump_index_entry(const struct super_block *sb,
fs/befs/debug.c
231
befs_debug(sb, "Btree super structure");
fs/befs/debug.c
232
befs_debug(sb, " magic %08x", fs32_to_cpu(sb, super->magic));
fs/befs/debug.c
233
befs_debug(sb, " node_size %u", fs32_to_cpu(sb, super->node_size));
fs/befs/debug.c
234
befs_debug(sb, " max_depth %08x", fs32_to_cpu(sb, super->max_depth));
fs/befs/debug.c
236
befs_debug(sb, " data_type %08x", fs32_to_cpu(sb, super->data_type));
fs/befs/debug.c
237
befs_debug(sb, " root_node_pointer %016LX",
fs/befs/debug.c
238
fs64_to_cpu(sb, super->root_node_ptr));
fs/befs/debug.c
239
befs_debug(sb, " free_node_pointer %016LX",
fs/befs/debug.c
240
fs64_to_cpu(sb, super->free_node_ptr));
fs/befs/debug.c
241
befs_debug(sb, " maximum size %016LX",
fs/befs/debug.c
242
fs64_to_cpu(sb, super->max_size));
fs/befs/debug.c
248
befs_dump_index_node(const struct super_block *sb, befs_btree_nodehead *node)
fs/befs/debug.c
252
befs_debug(sb, "Btree node structure");
fs/befs/debug.c
253
befs_debug(sb, " left %016LX", fs64_to_cpu(sb, node->left));
fs/befs/debug.c
254
befs_debug(sb, " right %016LX", fs64_to_cpu(sb, node->right));
fs/befs/debug.c
255
befs_debug(sb, " overflow %016LX", fs64_to_cpu(sb, node->overflow));
fs/befs/debug.c
256
befs_debug(sb, " all_key_count %hu",
fs/befs/debug.c
257
fs16_to_cpu(sb, node->all_key_count));
fs/befs/debug.c
258
befs_debug(sb, " all_key_length %hu",
fs/befs/debug.c
259
fs16_to_cpu(sb, node->all_key_length));
fs/befs/debug.c
29
befs_error(const struct super_block *sb, const char *fmt, ...)
fs/befs/debug.c
37
pr_err("(%s): %pV\n", sb->s_id, &vaf);
fs/befs/debug.c
42
befs_warning(const struct super_block *sb, const char *fmt, ...)
fs/befs/debug.c
50
pr_warn("(%s): %pV\n", sb->s_id, &vaf);
fs/befs/debug.c
55
befs_debug(const struct super_block *sb, const char *fmt, ...)
fs/befs/debug.c
65
pr_debug("(%s): %pV\n", sb->s_id, &vaf);
fs/befs/debug.c
72
befs_dump_inode(const struct super_block *sb, befs_inode *inode)
fs/befs/debug.c
78
befs_debug(sb, "befs_inode information");
fs/befs/debug.c
80
befs_debug(sb, " magic1 %08x", fs32_to_cpu(sb, inode->magic1));
fs/befs/debug.c
82
tmp_run = fsrun_to_cpu(sb, inode->inode_num);
fs/befs/debug.c
83
befs_debug(sb, " inode_num %u, %hu, %hu",
fs/befs/debug.c
86
befs_debug(sb, " uid %u", fs32_to_cpu(sb, inode->uid));
fs/befs/debug.c
87
befs_debug(sb, " gid %u", fs32_to_cpu(sb, inode->gid));
fs/befs/debug.c
88
befs_debug(sb, " mode %08x", fs32_to_cpu(sb, inode->mode));
fs/befs/debug.c
89
befs_debug(sb, " flags %08x", fs32_to_cpu(sb, inode->flags));
fs/befs/debug.c
90
befs_debug(sb, " create_time %llu",
fs/befs/debug.c
91
fs64_to_cpu(sb, inode->create_time));
fs/befs/debug.c
92
befs_debug(sb, " last_modified_time %llu",
fs/befs/debug.c
93
fs64_to_cpu(sb, inode->last_modified_time));
fs/befs/debug.c
95
tmp_run = fsrun_to_cpu(sb, inode->parent);
fs/befs/debug.c
96
befs_debug(sb, " parent [%u, %hu, %hu]",
fs/befs/debug.c
99
tmp_run = fsrun_to_cpu(sb, inode->attributes);
fs/befs/endian.h
106
fsds_to_cpu(const struct super_block *sb, const befs_disk_data_stream *n)
fs/befs/endian.h
112
data.direct[i] = fsrun_to_cpu(sb, n->direct[i]);
fs/befs/endian.h
114
data.max_direct_range = fs64_to_cpu(sb, n->max_direct_range);
fs/befs/endian.h
115
data.indirect = fsrun_to_cpu(sb, n->indirect);
fs/befs/endian.h
116
data.max_indirect_range = fs64_to_cpu(sb, n->max_indirect_range);
fs/befs/endian.h
117
data.double_indirect = fsrun_to_cpu(sb, n->double_indirect);
fs/befs/endian.h
118
data.max_double_indirect_range = fs64_to_cpu(sb,
fs/befs/endian.h
121
data.size = fs64_to_cpu(sb, n->size);
fs/befs/endian.h
16
fs64_to_cpu(const struct super_block *sb, fs64 n)
fs/befs/endian.h
18
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
25
cpu_to_fs64(const struct super_block *sb, u64 n)
fs/befs/endian.h
27
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
34
fs32_to_cpu(const struct super_block *sb, fs32 n)
fs/befs/endian.h
36
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
43
cpu_to_fs32(const struct super_block *sb, u32 n)
fs/befs/endian.h
45
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
52
fs16_to_cpu(const struct super_block *sb, fs16 n)
fs/befs/endian.h
54
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
61
cpu_to_fs16(const struct super_block *sb, u16 n)
fs/befs/endian.h
63
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE)
fs/befs/endian.h
72
fsrun_to_cpu(const struct super_block *sb, befs_disk_block_run n)
fs/befs/endian.h
76
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE) {
fs/befs/endian.h
89
cpu_to_fsrun(const struct super_block *sb, befs_block_run n)
fs/befs/endian.h
93
if (BEFS_SB(sb)->byte_order == BEFS_BYTESEX_LE) {
fs/befs/inode.c
19
befs_check_inode(struct super_block *sb, befs_inode *raw_inode,
fs/befs/inode.c
22
u32 magic1 = fs32_to_cpu(sb, raw_inode->magic1);
fs/befs/inode.c
23
befs_inode_addr ino_num = fsrun_to_cpu(sb, raw_inode->inode_num);
fs/befs/inode.c
24
u32 flags = fs32_to_cpu(sb, raw_inode->flags);
fs/befs/inode.c
28
befs_error(sb,
fs/befs/inode.c
37
if (inode != iaddr2blockno(sb, &ino_num)) {
fs/befs/inode.c
38
befs_error(sb, "inode blocknr field disagrees with vfs "
fs/befs/inode.c
40
inode, (unsigned long)iaddr2blockno(sb, &ino_num));
fs/befs/inode.c
49
befs_error(sb, "inode is not used - inode = %lu",
fs/befs/inode.h
6
int befs_check_inode(struct super_block *sb, befs_inode *raw_inode,
fs/befs/io.c
27
befs_bread_iaddr(struct super_block *sb, befs_inode_addr iaddr)
fs/befs/io.c
31
struct befs_sb_info *befs_sb = BEFS_SB(sb);
fs/befs/io.c
33
befs_debug(sb, "---> Enter %s "
fs/befs/io.c
38
befs_error(sb, "BEFS: Invalid allocation group %u, max is %u",
fs/befs/io.c
43
block = iaddr2blockno(sb, &iaddr);
fs/befs/io.c
45
befs_debug(sb, "%s: offset = %lu", __func__, (unsigned long)block);
fs/befs/io.c
47
bh = sb_bread(sb, block);
fs/befs/io.c
50
befs_error(sb, "Failed to read block %lu",
fs/befs/io.c
55
befs_debug(sb, "<--- %s", __func__);
fs/befs/io.c
59
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/io.h
5
struct buffer_head *befs_bread_iaddr(struct super_block *sb,
fs/befs/linuxvfs.c
137
struct super_block *sb = inode->i_sb;
fs/befs/linuxvfs.c
143
befs_debug(sb, "---> befs_get_block() for inode %lu, block %ld",
fs/befs/linuxvfs.c
146
befs_error(sb, "befs_get_block() was asked to write to "
fs/befs/linuxvfs.c
152
res = befs_fblock2brun(sb, ds, block, &run);
fs/befs/linuxvfs.c
154
befs_error(sb,
fs/befs/linuxvfs.c
161
disk_off = (ulong) iaddr2blockno(sb, &run);
fs/befs/linuxvfs.c
165
befs_debug(sb, "<--- %s for inode %lu, block %ld, disk address %lu",
fs/befs/linuxvfs.c
176
struct super_block *sb = dir->i_sb;
fs/befs/linuxvfs.c
184
befs_debug(sb, "---> %s name %pd inode %ld", __func__,
fs/befs/linuxvfs.c
188
if (BEFS_SB(sb)->nls) {
fs/befs/linuxvfs.c
190
befs_nls2utf(sb, name, strlen(name), &utfname, &utfnamelen);
fs/befs/linuxvfs.c
192
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/linuxvfs.c
195
ret = befs_btree_find(sb, ds, utfname, &offset);
fs/befs/linuxvfs.c
199
ret = befs_btree_find(sb, ds, name, &offset);
fs/befs/linuxvfs.c
203
befs_debug(sb, "<--- %s %pd not found", __func__, dentry);
fs/befs/linuxvfs.c
206
befs_error(sb, "<--- %s Error", __func__);
fs/befs/linuxvfs.c
211
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
220
struct super_block *sb = inode->i_sb;
fs/befs/linuxvfs.c
227
befs_debug(sb, "---> %s name %pD, inode %ld, ctx->pos %lld",
fs/befs/linuxvfs.c
231
result = befs_btree_read(sb, ds, ctx->pos, BEFS_NAME_LEN + 1,
fs/befs/linuxvfs.c
235
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/linuxvfs.c
236
befs_error(sb, "IO error reading %pD (inode %lu)",
fs/befs/linuxvfs.c
241
befs_debug(sb, "<--- %s END", __func__);
fs/befs/linuxvfs.c
245
befs_debug(sb, "<--- %s Empty directory", __func__);
fs/befs/linuxvfs.c
250
if (BEFS_SB(sb)->nls) {
fs/befs/linuxvfs.c
255
befs_utf2nls(sb, keybuf, keysize, &nlsname,
fs/befs/linuxvfs.c
258
befs_debug(sb, "<--- %s ERROR", __func__);
fs/befs/linuxvfs.c
277
befs_alloc_inode(struct super_block *sb)
fs/befs/linuxvfs.c
281
bi = alloc_inode_sb(sb, befs_inode_cachep, GFP_KERNEL);
fs/befs/linuxvfs.c
299
static struct inode *befs_iget(struct super_block *sb, unsigned long ino)
fs/befs/linuxvfs.c
303
struct befs_sb_info *befs_sb = BEFS_SB(sb);
fs/befs/linuxvfs.c
307
befs_debug(sb, "---> %s inode = %lu", __func__, ino);
fs/befs/linuxvfs.c
309
inode = iget_locked(sb, ino);
fs/befs/linuxvfs.c
318
befs_ino->i_inode_num = blockno2iaddr(sb, inode->i_ino);
fs/befs/linuxvfs.c
320
befs_debug(sb, " real inode number [%u, %hu, %hu]",
fs/befs/linuxvfs.c
324
bh = sb_bread(sb, inode->i_ino);
fs/befs/linuxvfs.c
326
befs_error(sb, "unable to read inode block - "
fs/befs/linuxvfs.c
333
befs_dump_inode(sb, raw_inode);
fs/befs/linuxvfs.c
335
if (befs_check_inode(sb, raw_inode, inode->i_ino) != BEFS_OK) {
fs/befs/linuxvfs.c
336
befs_error(sb, "Bad inode: %lu", inode->i_ino);
fs/befs/linuxvfs.c
340
inode->i_mode = (umode_t) fs32_to_cpu(sb, raw_inode->mode);
fs/befs/linuxvfs.c
349
make_kuid(&init_user_ns, fs32_to_cpu(sb, raw_inode->uid));
fs/befs/linuxvfs.c
352
make_kgid(&init_user_ns, fs32_to_cpu(sb, raw_inode->gid));
fs/befs/linuxvfs.c
366
fs64_to_cpu(sb, raw_inode->last_modified_time) >> 16,
fs/befs/linuxvfs.c
371
befs_ino->i_inode_num = fsrun_to_cpu(sb, raw_inode->inode_num);
fs/befs/linuxvfs.c
372
befs_ino->i_parent = fsrun_to_cpu(sb, raw_inode->parent);
fs/befs/linuxvfs.c
373
befs_ino->i_attribute = fsrun_to_cpu(sb, raw_inode->attributes);
fs/befs/linuxvfs.c
374
befs_ino->i_flags = fs32_to_cpu(sb, raw_inode->flags);
fs/befs/linuxvfs.c
385
fsds_to_cpu(sb, &raw_inode->data.datastream);
fs/befs/linuxvfs.c
387
num_blks = befs_count_blocks(sb, &befs_ino->i_data.ds);
fs/befs/linuxvfs.c
410
befs_error(sb, "Inode %lu is not a regular file, "
fs/befs/linuxvfs.c
417
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
426
befs_debug(sb, "<--- %s - Bad inode", __func__);
fs/befs/linuxvfs.c
474
struct super_block *sb = inode->i_sb;
fs/befs/linuxvfs.c
482
befs_error(sb, "Long symlink with illegal length");
fs/befs/linuxvfs.c
485
befs_debug(sb, "Follow long symlink");
fs/befs/linuxvfs.c
487
if (befs_read_lsymlink(sb, data, link, len) != len) {
fs/befs/linuxvfs.c
488
befs_error(sb, "Failed to read entire long symlink");
fs/befs/linuxvfs.c
50
static struct inode *befs_alloc_inode(struct super_block *sb);
fs/befs/linuxvfs.c
504
befs_utf2nls(struct super_block *sb, const char *in,
fs/befs/linuxvfs.c
507
struct nls_table *nls = BEFS_SB(sb)->nls;
fs/befs/linuxvfs.c
518
befs_debug(sb, "---> %s", __func__);
fs/befs/linuxvfs.c
521
befs_error(sb, "%s called with no NLS table loaded", __func__);
fs/befs/linuxvfs.c
54
static int befs_utf2nls(struct super_block *sb, const char *in, int in_len,
fs/befs/linuxvfs.c
546
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
551
befs_error(sb, "Name using character set %s contains a character that "
fs/befs/linuxvfs.c
553
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
56
static int befs_nls2utf(struct super_block *sb, const char *in, int in_len,
fs/befs/linuxvfs.c
581
befs_nls2utf(struct super_block *sb, const char *in,
fs/befs/linuxvfs.c
584
struct nls_table *nls = BEFS_SB(sb)->nls;
fs/befs/linuxvfs.c
596
befs_debug(sb, "---> %s\n", __func__);
fs/befs/linuxvfs.c
599
befs_error(sb, "%s called with no NLS table loaded.",
fs/befs/linuxvfs.c
61
static struct dentry *befs_fh_to_dentry(struct super_block *sb,
fs/befs/linuxvfs.c
626
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
63
static struct dentry *befs_fh_to_parent(struct super_block *sb,
fs/befs/linuxvfs.c
631
befs_error(sb, "Name using character set %s contains a character that "
fs/befs/linuxvfs.c
633
befs_debug(sb, "<--- %s", __func__);
fs/befs/linuxvfs.c
638
static struct inode *befs_nfs_get_inode(struct super_block *sb, uint64_t ino,
fs/befs/linuxvfs.c
642
return befs_iget(sb, ino);
fs/befs/linuxvfs.c
648
static struct dentry *befs_fh_to_dentry(struct super_block *sb,
fs/befs/linuxvfs.c
651
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/befs/linuxvfs.c
658
static struct dentry *befs_fh_to_parent(struct super_block *sb,
fs/befs/linuxvfs.c
661
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/befs/linuxvfs.c
749
befs_put_super(struct super_block *sb)
fs/befs/linuxvfs.c
751
kfree(BEFS_SB(sb)->mount_opts.iocharset);
fs/befs/linuxvfs.c
752
BEFS_SB(sb)->mount_opts.iocharset = NULL;
fs/befs/linuxvfs.c
753
unload_nls(BEFS_SB(sb)->nls);
fs/befs/linuxvfs.c
754
kfree(sb->s_fs_info);
fs/befs/linuxvfs.c
755
sb->s_fs_info = NULL;
fs/befs/linuxvfs.c
780
befs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/befs/linuxvfs.c
793
sb->s_fs_info = kzalloc_obj(*befs_sb);
fs/befs/linuxvfs.c
794
if (sb->s_fs_info == NULL)
fs/befs/linuxvfs.c
797
befs_sb = BEFS_SB(sb);
fs/befs/linuxvfs.c
801
befs_debug(sb, "---> %s", __func__);
fs/befs/linuxvfs.c
803
if (!sb_rdonly(sb)) {
fs/befs/linuxvfs.c
804
befs_warning(sb,
fs/befs/linuxvfs.c
806
sb->s_flags |= SB_RDONLY;
fs/befs/linuxvfs.c
817
blocksize = sb_min_blocksize(sb, 1024);
fs/befs/linuxvfs.c
820
befs_error(sb, "unable to set blocksize");
fs/befs/linuxvfs.c
824
bh = sb_bread(sb, sb_block);
fs/befs/linuxvfs.c
827
befs_error(sb, "unable to read superblock");
fs/befs/linuxvfs.c
835
befs_debug(sb, "Using PPC superblock location");
fs/befs/linuxvfs.c
837
befs_debug(sb, "Using x86 superblock location");
fs/befs/linuxvfs.c
842
if ((befs_load_sb(sb, disk_sb) != BEFS_OK) ||
fs/befs/linuxvfs.c
843
(befs_check_sb(sb) != BEFS_OK))
fs/befs/linuxvfs.c
846
befs_dump_super_block(sb, disk_sb);
fs/befs/linuxvfs.c
852
befs_error(sb, "blocks count: %llu is larger than the host can use",
fs/befs/linuxvfs.c
861
sb->s_magic = BEFS_SUPER_MAGIC;
fs/befs/linuxvfs.c
863
sb_set_blocksize(sb, (ulong) befs_sb->block_size);
fs/befs/linuxvfs.c
864
sb->s_op = &befs_sops;
fs/befs/linuxvfs.c
865
sb->s_export_op = &befs_export_operations;
fs/befs/linuxvfs.c
866
sb->s_time_min = 0;
fs/befs/linuxvfs.c
867
sb->s_time_max = 0xffffffffffffll;
fs/befs/linuxvfs.c
868
root = befs_iget(sb, iaddr2blockno(sb, &(befs_sb->root_dir)));
fs/befs/linuxvfs.c
873
sb->s_root = d_make_root(root);
fs/befs/linuxvfs.c
874
if (!sb->s_root) {
fs/befs/linuxvfs.c
876
befs_error(sb, "get root inode failed");
fs/befs/linuxvfs.c
882
befs_debug(sb, "Loading nls: %s",
fs/befs/linuxvfs.c
886
befs_warning(sb, "Cannot load nls %s"
fs/befs/linuxvfs.c
893
befs_debug(sb, "Loading default nls");
fs/befs/linuxvfs.c
904
kfree(sb->s_fs_info);
fs/befs/linuxvfs.c
905
sb->s_fs_info = NULL;
fs/befs/linuxvfs.c
923
struct super_block *sb = dentry->d_sb;
fs/befs/linuxvfs.c
924
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/befs/linuxvfs.c
926
befs_debug(sb, "---> %s", __func__);
fs/befs/linuxvfs.c
929
buf->f_bsize = sb->s_blocksize;
fs/befs/linuxvfs.c
930
buf->f_blocks = BEFS_SB(sb)->num_blocks;
fs/befs/linuxvfs.c
931
buf->f_bfree = BEFS_SB(sb)->num_blocks - BEFS_SB(sb)->used_blocks;
fs/befs/linuxvfs.c
938
befs_debug(sb, "<--- %s", __func__);
fs/befs/super.c
107
befs_error(sb, "ag_shift disagrees with blocks_per_ag.");
fs/befs/super.c
111
befs_error(sb, "Filesystem not clean! There are blocks in the "
fs/befs/super.c
21
befs_load_sb(struct super_block *sb, befs_super_block *disk_sb)
fs/befs/super.c
23
struct befs_sb_info *befs_sb = BEFS_SB(sb);
fs/befs/super.c
31
befs_sb->magic1 = fs32_to_cpu(sb, disk_sb->magic1);
fs/befs/super.c
32
befs_sb->magic2 = fs32_to_cpu(sb, disk_sb->magic2);
fs/befs/super.c
33
befs_sb->magic3 = fs32_to_cpu(sb, disk_sb->magic3);
fs/befs/super.c
34
befs_sb->block_size = fs32_to_cpu(sb, disk_sb->block_size);
fs/befs/super.c
35
befs_sb->block_shift = fs32_to_cpu(sb, disk_sb->block_shift);
fs/befs/super.c
36
befs_sb->num_blocks = fs64_to_cpu(sb, disk_sb->num_blocks);
fs/befs/super.c
37
befs_sb->used_blocks = fs64_to_cpu(sb, disk_sb->used_blocks);
fs/befs/super.c
38
befs_sb->inode_size = fs32_to_cpu(sb, disk_sb->inode_size);
fs/befs/super.c
40
befs_sb->blocks_per_ag = fs32_to_cpu(sb, disk_sb->blocks_per_ag);
fs/befs/super.c
41
befs_sb->ag_shift = fs32_to_cpu(sb, disk_sb->ag_shift);
fs/befs/super.c
42
befs_sb->num_ags = fs32_to_cpu(sb, disk_sb->num_ags);
fs/befs/super.c
44
befs_sb->flags = fs32_to_cpu(sb, disk_sb->flags);
fs/befs/super.c
46
befs_sb->log_blocks = fsrun_to_cpu(sb, disk_sb->log_blocks);
fs/befs/super.c
47
befs_sb->log_start = fs64_to_cpu(sb, disk_sb->log_start);
fs/befs/super.c
48
befs_sb->log_end = fs64_to_cpu(sb, disk_sb->log_end);
fs/befs/super.c
50
befs_sb->root_dir = fsrun_to_cpu(sb, disk_sb->root_dir);
fs/befs/super.c
51
befs_sb->indices = fsrun_to_cpu(sb, disk_sb->indices);
fs/befs/super.c
58
befs_check_sb(struct super_block *sb)
fs/befs/super.c
60
struct befs_sb_info *befs_sb = BEFS_SB(sb);
fs/befs/super.c
66
befs_error(sb, "invalid magic header");
fs/befs/super.c
80
befs_error(sb, "invalid blocksize: %u", befs_sb->block_size);
fs/befs/super.c
85
befs_error(sb, "blocksize(%u) cannot be larger "
fs/befs/super.c
97
befs_error(sb, "block_shift disagrees with block_size. "
fs/befs/super.h
5
int befs_load_sb(struct super_block *sb, befs_super_block *disk_sb);
fs/befs/super.h
6
int befs_check_sb(struct super_block *sb);
fs/bfs/bfs.h
41
static inline struct bfs_sb_info *BFS_SB(struct super_block *sb)
fs/bfs/bfs.h
43
return sb->s_fs_info;
fs/bfs/bfs.h
56
extern struct inode *bfs_iget(struct super_block *sb, unsigned long ino);
fs/bfs/file.c
110
map_bh(bh_result, sb, phys);
fs/bfs/file.c
148
map_bh(bh_result, sb, phys);
fs/bfs/file.c
35
struct super_block *sb)
fs/bfs/file.c
39
bh = sb_bread(sb, from);
fs/bfs/file.c
42
new = sb_getblk(sb, to);
fs/bfs/file.c
50
static int bfs_move_blocks(struct super_block *sb, unsigned long start,
fs/bfs/file.c
57
if(bfs_move_block(i, where + i, sb)) {
fs/bfs/file.c
70
struct super_block *sb = inode->i_sb;
fs/bfs/file.c
71
struct bfs_sb_info *info = BFS_SB(sb);
fs/bfs/file.c
79
map_bh(bh_result, sb, phys);
fs/bfs/file.c
91
map_bh(bh_result, sb, phys);
fs/bfs/inode.c
114
static struct bfs_inode *find_inode(struct super_block *sb, u16 ino, struct buffer_head **p)
fs/bfs/inode.c
116
if ((ino < BFS_ROOT_INO) || (ino > BFS_SB(sb)->si_lasti)) {
fs/bfs/inode.c
117
printf("Bad inode number %s:%08x\n", sb->s_id, ino);
fs/bfs/inode.c
123
*p = sb_bread(sb, 1 + ino / BFS_INODES_PER_BLOCK);
fs/bfs/inode.c
125
printf("Unable to read inode %s:%08x\n", sb->s_id, ino);
fs/bfs/inode.c
254
static struct inode *bfs_alloc_inode(struct super_block *sb)
fs/bfs/inode.c
257
bi = alloc_inode_sb(sb, bfs_inode_cachep, GFP_KERNEL);
fs/bfs/inode.c
35
struct inode *bfs_iget(struct super_block *sb, unsigned long ino)
fs/bfs/inode.c
42
inode = iget_locked(sb, ino);
fs/binfmt_misc.c
619
static struct inode *bm_get_inode(struct super_block *sb, int mode)
fs/binfmt_misc.c
621
struct inode *inode = new_inode(sb);
fs/binfmt_misc.c
770
static int add_entry(Node *e, struct super_block *sb)
fs/binfmt_misc.c
772
struct dentry *dentry = simple_start_creating(sb->s_root, e->name);
fs/binfmt_misc.c
779
inode = bm_get_inode(sb, S_IFREG | 0644);
fs/binfmt_misc.c
803
struct super_block *sb = file_inode(file)->i_sb;
fs/binfmt_misc.c
831
err = add_entry(e, sb);
fs/binfmt_misc.c
913
static void bm_put_super(struct super_block *sb)
fs/binfmt_misc.c
915
struct user_namespace *user_ns = sb->s_fs_info;
fs/binfmt_misc.c
917
sb->s_fs_info = NULL;
fs/binfmt_misc.c
927
static int bm_fill_super(struct super_block *sb, struct fs_context *fc)
fs/binfmt_misc.c
930
struct user_namespace *user_ns = sb->s_user_ns;
fs/binfmt_misc.c
988
err = simple_fill_super(sb, BINFMTFS_MAGIC, bm_files);
fs/binfmt_misc.c
990
sb->s_op = &s_ops;
fs/btrfs/bio.c
952
if (unlikely(sb_rdonly(fs_info->sb)))
fs/btrfs/block-group.c
1906
guard(super_write)(fs_info->sb);
fs/btrfs/block-group.c
3088
if (sb_rdonly(fs_info->sb)) {
fs/btrfs/btrfs_inode.h
603
struct inode *btrfs_alloc_inode(struct super_block *sb);
fs/btrfs/defrag.c
258
scoped_guard(super_write, fs_info->sb)
fs/btrfs/dev-replace.c
1085
if (sb_rdonly(fs_info->sb))
fs/btrfs/dev-replace.c
251
fs_info->sb, &fs_holder_ops);
fs/btrfs/discard.c
89
return (!(fs_info->sb->s_flags & SB_RDONLY) &&
fs/btrfs/disk-io.c
180
if (sb_rdonly(fs_info->sb))
fs/btrfs/disk-io.c
1867
static int btrfs_init_btree_inode(struct super_block *sb)
fs/btrfs/disk-io.c
1869
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/disk-io.c
1874
inode = new_inode(sb);
fs/btrfs/disk-io.c
2043
if (sb_rdonly(fs_info->sb)) {
fs/btrfs/disk-io.c
2286
const struct btrfs_super_block *sb)
fs/btrfs/disk-io.c
2293
const u32 sectorsize = btrfs_super_sectorsize(sb);
fs/btrfs/disk-io.c
2294
u32 sys_array_size = btrfs_super_sys_array_size(sb);
fs/btrfs/disk-io.c
2311
disk_key = (struct btrfs_disk_key *)(sb->sys_chunk_array + cur);
fs/btrfs/disk-io.c
2325
chunk = (struct btrfs_chunk *)(sb->sys_chunk_array + cur);
fs/btrfs/disk-io.c
2361
const struct btrfs_super_block *sb, int mirror_num)
fs/btrfs/disk-io.c
2363
u64 nodesize = btrfs_super_nodesize(sb);
fs/btrfs/disk-io.c
2364
u64 sectorsize = btrfs_super_sectorsize(sb);
fs/btrfs/disk-io.c
2368
if (btrfs_super_magic(sb) != BTRFS_MAGIC) {
fs/btrfs/disk-io.c
2372
if ((btrfs_super_flags(sb) & ~BTRFS_SUPER_FLAG_SUPP)) {
fs/btrfs/disk-io.c
2376
btrfs_super_flags(sb) & ~BTRFS_SUPER_FLAG_SUPP);
fs/btrfs/disk-io.c
2381
btrfs_super_flags(sb) & ~BTRFS_SUPER_FLAG_SUPP);
fs/btrfs/disk-io.c
2384
if (btrfs_super_root_level(sb) >= BTRFS_MAX_LEVEL) {
fs/btrfs/disk-io.c
2386
btrfs_super_root_level(sb), BTRFS_MAX_LEVEL);
fs/btrfs/disk-io.c
2389
if (btrfs_super_chunk_root_level(sb) >= BTRFS_MAX_LEVEL) {
fs/btrfs/disk-io.c
2391
btrfs_super_chunk_root_level(sb), BTRFS_MAX_LEVEL);
fs/btrfs/disk-io.c
2394
if (btrfs_super_log_root_level(sb) >= BTRFS_MAX_LEVEL) {
fs/btrfs/disk-io.c
2396
btrfs_super_log_root_level(sb), BTRFS_MAX_LEVEL);
fs/btrfs/disk-io.c
2422
if (nodesize != le32_to_cpu(sb->__unused_leafsize)) {
fs/btrfs/disk-io.c
2424
le32_to_cpu(sb->__unused_leafsize), nodesize);
fs/btrfs/disk-io.c
2429
if (!IS_ALIGNED(btrfs_super_root(sb), sectorsize)) {
fs/btrfs/disk-io.c
2431
btrfs_super_root(sb));
fs/btrfs/disk-io.c
2434
if (!IS_ALIGNED(btrfs_super_chunk_root(sb), sectorsize)) {
fs/btrfs/disk-io.c
2436
btrfs_super_chunk_root(sb));
fs/btrfs/disk-io.c
2439
if (!IS_ALIGNED(btrfs_super_log_root(sb), sectorsize)) {
fs/btrfs/disk-io.c
2441
btrfs_super_log_root(sb));
fs/btrfs/disk-io.c
2446
memcmp(fs_info->fs_devices->fsid, sb->fsid, BTRFS_FSID_SIZE) != 0) {
fs/btrfs/disk-io.c
2449
sb->fsid, fs_info->fs_devices->fsid);
fs/btrfs/disk-io.c
2453
if (memcmp(fs_info->fs_devices->metadata_uuid, btrfs_sb_fsid_ptr(sb),
fs/btrfs/disk-io.c
2457
btrfs_sb_fsid_ptr(sb), fs_info->fs_devices->metadata_uuid);
fs/btrfs/disk-io.c
2461
if (memcmp(fs_info->fs_devices->metadata_uuid, sb->dev_item.fsid,
fs/btrfs/disk-io.c
2465
fs_info->fs_devices->metadata_uuid, sb->dev_item.fsid);
fs/btrfs/disk-io.c
2514
if (btrfs_super_bytes_used(sb) < 6 * btrfs_super_nodesize(sb)) {
fs/btrfs/disk-io.c
2516
btrfs_super_bytes_used(sb));
fs/btrfs/disk-io.c
2519
if (!is_power_of_2(btrfs_super_stripesize(sb))) {
fs/btrfs/disk-io.c
2521
btrfs_super_stripesize(sb));
fs/btrfs/disk-io.c
2524
if (btrfs_super_num_devices(sb) > (1UL << 31))
fs/btrfs/disk-io.c
2526
btrfs_super_num_devices(sb));
fs/btrfs/disk-io.c
2527
if (btrfs_super_num_devices(sb) == 0) {
fs/btrfs/disk-io.c
2533
btrfs_super_bytenr(sb) != btrfs_sb_offset(mirror_num)) {
fs/btrfs/disk-io.c
2535
btrfs_super_bytenr(sb), btrfs_sb_offset(mirror_num));
fs/btrfs/disk-io.c
2542
ret = validate_sys_chunk_array(fs_info, sb);
fs/btrfs/disk-io.c
2548
if (btrfs_super_sys_array_size(sb) > BTRFS_SYSTEM_CHUNK_ARRAY_SIZE) {
fs/btrfs/disk-io.c
2550
btrfs_super_sys_array_size(sb),
fs/btrfs/disk-io.c
2554
if (btrfs_super_sys_array_size(sb) < sizeof(struct btrfs_disk_key)
fs/btrfs/disk-io.c
2557
btrfs_super_sys_array_size(sb),
fs/btrfs/disk-io.c
2567
if (btrfs_super_generation(sb) < btrfs_super_chunk_root_generation(sb))
fs/btrfs/disk-io.c
2570
btrfs_super_generation(sb),
fs/btrfs/disk-io.c
2571
btrfs_super_chunk_root_generation(sb));
fs/btrfs/disk-io.c
2572
if (btrfs_super_generation(sb) < btrfs_super_cache_generation(sb)
fs/btrfs/disk-io.c
2573
&& btrfs_super_cache_generation(sb) != (u64)-1)
fs/btrfs/disk-io.c
2576
btrfs_super_generation(sb),
fs/btrfs/disk-io.c
2577
btrfs_super_cache_generation(sb));
fs/btrfs/disk-io.c
2599
struct btrfs_super_block *sb)
fs/btrfs/disk-io.c
2603
ret = btrfs_validate_super(fs_info, sb, -1);
fs/btrfs/disk-io.c
2606
if (unlikely(!btrfs_supported_super_csum(btrfs_super_csum_type(sb)))) {
fs/btrfs/disk-io.c
2609
btrfs_super_csum_type(sb), BTRFS_CSUM_TYPE_CRC32);
fs/btrfs/disk-io.c
2612
if (unlikely(btrfs_super_incompat_flags(sb) & ~BTRFS_FEATURE_INCOMPAT_SUPP)) {
fs/btrfs/disk-io.c
2616
btrfs_super_incompat_flags(sb),
fs/btrfs/disk-io.c
2656
struct btrfs_super_block *sb = fs_info->super_copy;
fs/btrfs/disk-io.c
2660
bytenr = btrfs_super_root(sb);
fs/btrfs/disk-io.c
2661
gen = btrfs_super_generation(sb);
fs/btrfs/disk-io.c
2662
level = btrfs_super_root_level(sb);
fs/btrfs/disk-io.c
2670
bytenr = btrfs_super_remap_root(sb);
fs/btrfs/disk-io.c
2671
gen = btrfs_super_remap_root_generation(sb);
fs/btrfs/disk-io.c
2672
level = btrfs_super_remap_root_level(sb);
fs/btrfs/disk-io.c
2686
struct btrfs_super_block *sb = fs_info->super_copy;
fs/btrfs/disk-io.c
2707
btrfs_set_super_log_root(sb, 0);
fs/btrfs/disk-io.c
2910
static int init_mount_fs_info(struct btrfs_fs_info *fs_info, struct super_block *sb)
fs/btrfs/disk-io.c
2914
fs_info->sb = sb;
fs/btrfs/disk-io.c
2916
sb->s_blocksize = BTRFS_BDEV_BLOCKSIZE;
fs/btrfs/disk-io.c
2917
sb->s_blocksize_bits = blksize_bits(BTRFS_BDEV_BLOCKSIZE);
fs/btrfs/disk-io.c
2949
if (sb_rdonly(sb))
fs/btrfs/disk-io.c
3286
if (!sb_rdonly(fs_info->sb))
fs/btrfs/disk-io.c
3293
int __cold open_ctree(struct super_block *sb, struct btrfs_fs_devices *fs_devices)
fs/btrfs/disk-io.c
3301
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/disk-io.c
3308
ret = init_mount_fs_info(fs_info, sb);
fs/btrfs/disk-io.c
3324
ret = btrfs_init_btree_inode(sb);
fs/btrfs/disk-io.c
3430
if (!btrfs_check_options(fs_info, &fs_info->mount_opt, sb->s_flags)) {
fs/btrfs/disk-io.c
3435
ret = btrfs_check_features(fs_info, !sb_rdonly(sb));
fs/btrfs/disk-io.c
3462
sb->s_bdi->ra_pages *= btrfs_super_num_devices(disk_super);
fs/btrfs/disk-io.c
3463
sb->s_bdi->ra_pages = max(sb->s_bdi->ra_pages, SZ_4M / PAGE_SIZE);
fs/btrfs/disk-io.c
3466
sb->s_blocksize = sectorsize;
fs/btrfs/disk-io.c
3467
sb->s_blocksize_bits = blksize_bits(sectorsize);
fs/btrfs/disk-io.c
3468
memcpy(&sb->s_uuid, fs_info->fs_devices->fsid, BTRFS_FSID_SIZE);
fs/btrfs/disk-io.c
3612
if (!sb_rdonly(sb) && fs_info->fs_devices->missing_devices &&
fs/btrfs/disk-io.c
3665
if (sb_rdonly(sb))
fs/btrfs/disk-io.c
3776
struct btrfs_super_block *sb, int max_mirrors)
fs/btrfs/disk-io.c
3810
btrfs_set_super_bytenr(sb, bytenr_orig);
fs/btrfs/disk-io.c
3812
btrfs_csum(fs_info->csum_type, (const u8 *)sb + BTRFS_CSUM_SIZE,
fs/btrfs/disk-io.c
3813
BTRFS_SUPER_INFO_SIZE - BTRFS_CSUM_SIZE, sb->csum);
fs/btrfs/disk-io.c
3828
memcpy(disk_super, sb, BTRFS_SUPER_INFO_SIZE);
fs/btrfs/disk-io.c
4050
struct btrfs_super_block *sb;
fs/btrfs/disk-io.c
4071
sb = fs_info->super_for_commit;
fs/btrfs/disk-io.c
4072
dev_item = &sb->dev_item;
fs/btrfs/disk-io.c
4112
flags = btrfs_super_flags(sb);
fs/btrfs/disk-io.c
4113
btrfs_set_super_flags(sb, flags | BTRFS_HEADER_FLAG_WRITTEN);
fs/btrfs/disk-io.c
4115
ret = btrfs_validate_write_super(fs_info, sb);
fs/btrfs/disk-io.c
4123
ret = write_dev_supers(dev, sb, max_mirrors);
fs/btrfs/disk-io.c
4399
if (!sb_rdonly(fs_info->sb)) {
fs/btrfs/disk-io.h
56
int __cold open_ctree(struct super_block *sb, struct btrfs_fs_devices *fs_devices);
fs/btrfs/disk-io.h
59
const struct btrfs_super_block *sb, int mirror_num);
fs/btrfs/export.c
106
static struct dentry *btrfs_fh_to_parent(struct super_block *sb, struct fid *fh,
fs/btrfs/export.c
127
return btrfs_get_dentry(sb, objectid, root_objectid, generation);
fs/btrfs/export.c
130
static struct dentry *btrfs_fh_to_dentry(struct super_block *sb, struct fid *fh,
fs/btrfs/export.c
149
return btrfs_get_dentry(sb, objectid, root_objectid, generation);
fs/btrfs/export.c
216
return btrfs_get_dentry(fs_info->sb, key.objectid,
fs/btrfs/export.c
79
struct dentry *btrfs_get_dentry(struct super_block *sb, u64 objectid,
fs/btrfs/export.c
82
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/export.h
25
struct dentry *btrfs_get_dentry(struct super_block *sb, u64 objectid,
fs/btrfs/fs.h
604
struct super_block *sb;
fs/btrfs/inode.c
5942
inode = iget5_locked_rcu(root->fs_info->sb, hashval, btrfs_find_actor,
fs/btrfs/inode.c
8048
struct inode *btrfs_alloc_inode(struct super_block *sb)
fs/btrfs/inode.c
8050
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/inode.c
8054
ei = alloc_inode_sb(sb, btrfs_inode_cachep, GFP_KERNEL);
fs/btrfs/inode.c
8083
if (sb->s_magic != BTRFS_TEST_MAGIC)
fs/btrfs/ioctl.c
2219
dentry = btrfs_get_dentry(fs_info->sb,
fs/btrfs/ioctl.c
3139
if (sb_rdonly(fs_info->sb)) {
fs/btrfs/ioctl.c
5125
ret = freeze_super(fs_info->sb, FREEZE_HOLDER_KERNEL, NULL);
fs/btrfs/ioctl.c
5129
ret = thaw_super(fs_info->sb, FREEZE_HOLDER_KERNEL, NULL);
fs/btrfs/messages.c
121
struct super_block *sb = fs_info->sb;
fs/btrfs/messages.c
136
if (error == -EROFS && sb_rdonly(sb))
fs/btrfs/messages.c
151
sb->s_id, statestr, function, line, error, errstr, &vaf);
fs/btrfs/messages.c
155
sb->s_id, statestr, function, line, error, errstr);
fs/btrfs/messages.c
166
if (!(sb->s_flags & SB_BORN))
fs/btrfs/messages.c
169
if (sb_rdonly(sb))
fs/btrfs/messages.c
175
btrfs_set_sb_rdonly(sb);
fs/btrfs/messages.c
237
fs_info->sb->s_id, statestr, &vaf);
fs/btrfs/messages.c
287
s_id = fs_info->sb->s_id;
fs/btrfs/qgroup.c
4598
if (!sb_rdonly(fs_info->sb))
fs/btrfs/qgroup.c
4628
if (!sb_rdonly(fs_info->sb))
fs/btrfs/relocation.c
5347
ASSERT(sb_write_started(fs_info->sb));
fs/btrfs/scrub.c
2110
if (fs_info->sb->s_writers.frozen > SB_UNFROZEN ||
fs/btrfs/scrub.c
2956
struct btrfs_super_block *sb = page_address(page);
fs/btrfs/scrub.c
2959
ret = bdev_rw_virt(dev->bdev, physical >> SECTOR_SHIFT, sb,
fs/btrfs/scrub.c
2963
ret = btrfs_check_super_csum(fs_info, sb);
fs/btrfs/scrub.c
2970
if (unlikely(btrfs_super_generation(sb) != generation)) {
fs/btrfs/scrub.c
2974
btrfs_super_generation(sb), generation);
fs/btrfs/scrub.c
2978
return btrfs_validate_super(fs_info, sb, -1);
fs/btrfs/super.c
1002
sb->s_flags |= SB_ACTIVE;
fs/btrfs/super.c
1010
int btrfs_sync_fs(struct super_block *sb, int wait)
fs/btrfs/super.c
1013
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
1042
if (sb_start_write_trylock(sb))
fs/btrfs/super.c
1043
sb_end_write(sb);
fs/btrfs/super.c
1115
if (!(info->sb->s_flags & SB_POSIXACL))
fs/btrfs/super.c
1267
sync_filesystem(fs_info->sb);
fs/btrfs/super.c
1281
(!btrfs_raw_test_opt(fs_info->mount_opt, AUTO_DEFRAG) || sb_rdonly(fs_info->sb))) {
fs/btrfs/super.c
1332
btrfs_clear_sb_rdonly(fs_info->sb);
fs/btrfs/super.c
1361
btrfs_set_sb_rdonly(fs_info->sb);
fs/btrfs/super.c
1501
struct super_block *sb = fc->root->d_sb;
fs/btrfs/super.c
1502
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
1518
sync_filesystem(sb);
fs/btrfs/super.c
1535
(!sb_rdonly(sb) || (fc->sb_flags & SB_RDONLY))) {
fs/btrfs/super.c
1550
if (!sb_rdonly(sb) && (fc->sb_flags & SB_RDONLY))
fs/btrfs/super.c
1552
else if (sb_rdonly(sb) && !(fc->sb_flags & SB_RDONLY))
fs/btrfs/super.c
1562
if ((fc->sb_flags & SB_POSIXACL) != (sb->s_flags & SB_POSIXACL))
fs/btrfs/super.c
1833
static int btrfs_fc_test_super(struct super_block *sb, struct fs_context *fc)
fs/btrfs/super.c
1836
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
1847
struct super_block *sb;
fs/btrfs/super.c
1877
sb = sget_fc(fc, btrfs_fc_test_super, set_anon_super_fc);
fs/btrfs/super.c
1878
if (IS_ERR(sb)) {
fs/btrfs/super.c
1893
return PTR_ERR(sb);
fs/btrfs/super.c
1896
if (sb->s_root) {
fs/btrfs/super.c
1928
ret = btrfs_open_devices(fs_devices, mode, sb);
fs/btrfs/super.c
1933
deactivate_locked_super(sb);
fs/btrfs/super.c
1937
deactivate_locked_super(sb);
fs/btrfs/super.c
1942
snprintf(sb->s_id, sizeof(sb->s_id), "%pg", bdev);
fs/btrfs/super.c
1943
shrinker_debugfs_rename(sb->s_shrink, "sb-btrfs:%s", sb->s_id);
fs/btrfs/super.c
1944
ret = btrfs_fill_super(sb, fs_devices);
fs/btrfs/super.c
1946
deactivate_locked_super(sb);
fs/btrfs/super.c
1953
fc->root = dget(sb->s_root);
fs/btrfs/super.c
2124
static void btrfs_kill_super(struct super_block *sb)
fs/btrfs/super.c
2126
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2127
kill_anon_super(sb);
fs/btrfs/super.c
2290
static int btrfs_freeze(struct super_block *sb)
fs/btrfs/super.c
2292
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2307
struct btrfs_super_block *sb;
fs/btrfs/super.c
2320
sb = btrfs_read_disk_super(dev->bdev, 0, true);
fs/btrfs/super.c
2321
if (IS_ERR(sb))
fs/btrfs/super.c
2322
return PTR_ERR(sb);
fs/btrfs/super.c
2325
csum_type = btrfs_super_csum_type(sb);
fs/btrfs/super.c
2333
if (unlikely(btrfs_check_super_csum(fs_info, sb))) {
fs/btrfs/super.c
2340
ret = btrfs_validate_super(fs_info, sb, 0);
fs/btrfs/super.c
2345
if (unlikely(btrfs_super_generation(sb) != last_trans)) {
fs/btrfs/super.c
2347
btrfs_super_generation(sb), last_trans);
fs/btrfs/super.c
2352
btrfs_release_disk_super(sb);
fs/btrfs/super.c
2356
static int btrfs_unfreeze(struct super_block *sb)
fs/btrfs/super.c
2358
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2405
static long btrfs_nr_cached_objects(struct super_block *sb, struct shrink_control *sc)
fs/btrfs/super.c
2407
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2415
static long btrfs_free_cached_objects(struct super_block *sb, struct shrink_control *sc)
fs/btrfs/super.c
2418
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2427
static int btrfs_remove_bdev(struct super_block *sb, struct block_device *bdev)
fs/btrfs/super.c
2429
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
2478
static void btrfs_shutdown(struct super_block *sb)
fs/btrfs/super.c
2480
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
69
static void btrfs_put_super(struct super_block *sb)
fs/btrfs/super.c
71
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
956
static int btrfs_fill_super(struct super_block *sb,
fs/btrfs/super.c
960
struct btrfs_fs_info *fs_info = btrfs_sb(sb);
fs/btrfs/super.c
963
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/btrfs/super.c
964
sb->s_magic = BTRFS_SUPER_MAGIC;
fs/btrfs/super.c
965
sb->s_op = &btrfs_super_ops;
fs/btrfs/super.c
966
set_default_d_op(sb, &btrfs_dentry_operations);
fs/btrfs/super.c
967
sb->s_export_op = &btrfs_export_ops;
fs/btrfs/super.c
969
sb->s_vop = &btrfs_verityops;
fs/btrfs/super.c
971
sb->s_xattr = btrfs_xattr_handlers;
fs/btrfs/super.c
972
sb->s_time_gran = 1;
fs/btrfs/super.c
973
sb->s_iflags |= SB_I_CGROUPWB | SB_I_ALLOW_HSM;
fs/btrfs/super.c
975
ret = super_setup_bdi(sb);
fs/btrfs/super.c
981
ret = open_ctree(sb, fs_devices);
fs/btrfs/super.c
996
sb->s_root = d_make_root(&inode->vfs_inode);
fs/btrfs/super.c
997
if (!sb->s_root) {
fs/btrfs/super.h
16
int btrfs_sync_fs(struct super_block *sb, int wait);
fs/btrfs/super.h
21
static inline struct btrfs_fs_info *btrfs_sb(struct super_block *sb)
fs/btrfs/super.h
23
return sb->s_fs_info;
fs/btrfs/super.h
26
static inline void btrfs_set_sb_rdonly(struct super_block *sb)
fs/btrfs/super.h
28
sb->s_flags |= SB_RDONLY;
fs/btrfs/super.h
29
set_bit(BTRFS_FS_STATE_RO, &btrfs_sb(sb)->fs_state);
fs/btrfs/super.h
32
static inline void btrfs_clear_sb_rdonly(struct super_block *sb)
fs/btrfs/super.h
34
sb->s_flags &= ~SB_RDONLY;
fs/btrfs/super.h
35
clear_bit(BTRFS_FS_STATE_RO, &btrfs_sb(sb)->fs_state);
fs/btrfs/sysfs.c
1096
if (sb_rdonly(fs_info->sb))
fs/btrfs/sysfs.c
205
if (sb_rdonly(fs_info->sb))
fs/btrfs/sysfs.c
2334
ret = sysfs_create_link(fsid_kobj, &fs_info->sb->s_bdi->dev->kobj, "bdi");
fs/btrfs/sysfs.c
867
if (sb_rdonly(fs_info->sb))
fs/btrfs/transaction.c
1085
sb_end_intwrite(info->sb);
fs/btrfs/transaction.c
2107
sb_end_intwrite(fs_info->sb);
fs/btrfs/transaction.c
2175
try_to_writeback_inodes_sb(fs_info->sb, WB_REASON_SYNC);
fs/btrfs/transaction.c
2618
sb_end_intwrite(fs_info->sb);
fs/btrfs/transaction.c
702
sb_start_intwrite(fs_info->sb);
fs/btrfs/transaction.c
804
sb_end_intwrite(fs_info->sb);
fs/btrfs/tree-checker.c
1205
if (unlikely(!sb_rdonly(fs_info->sb) &&
fs/btrfs/tree-log.c
7423
struct super_block *sb = inode->vfs_inode.i_sb;
fs/btrfs/tree-log.c
7428
sb != parent->d_sb)
fs/btrfs/volumes.c
2218
if (fs_info->sb->s_bdev &&
fs/btrfs/volumes.c
2219
(fs_info->sb->s_bdev == device->bdev))
fs/btrfs/volumes.c
2220
fs_info->sb->s_bdev = next_device->bdev;
fs/btrfs/volumes.c
2808
struct super_block *sb = fs_info->sb;
fs/btrfs/volumes.c
2817
if (sb_rdonly(sb) && !fs_devices->seeding)
fs/btrfs/volumes.c
2821
fs_info->sb, &fs_holder_ops);
fs/btrfs/volumes.c
2837
down_write(&sb->s_umount);
fs/btrfs/volumes.c
2981
up_write(&sb->s_umount);
fs/btrfs/volumes.c
3045
up_write(&sb->s_umount);
fs/btrfs/volumes.c
4947
guard(super_write)(fs_info->sb);
fs/btrfs/volumes.c
5100
if (sb_rdonly(fs_info->sb)) {
fs/btrfs/volumes.c
734
const u8 *btrfs_sb_fsid_ptr(const struct btrfs_super_block *sb)
fs/btrfs/volumes.c
736
bool has_metadata_uuid = (btrfs_super_incompat_flags(sb) &
fs/btrfs/volumes.c
739
return has_metadata_uuid ? sb->metadata_uuid : sb->fsid;
fs/btrfs/volumes.c
7504
ret = open_fs_devices(fs_devices, BLK_OPEN_READ, fs_info->sb);
fs/btrfs/volumes.c
7635
struct extent_buffer *sb;
fs/btrfs/volumes.c
7650
sb = alloc_dummy_extent_buffer(fs_info, BTRFS_SUPER_INFO_OFFSET);
fs/btrfs/volumes.c
7651
if (!sb)
fs/btrfs/volumes.c
7653
set_extent_buffer_uptodate(sb);
fs/btrfs/volumes.c
7655
write_extent_buffer(sb, super_copy, 0, BTRFS_SUPER_INFO_SIZE);
fs/btrfs/volumes.c
7682
ASSERT(btrfs_chunk_type(sb, chunk) & BTRFS_BLOCK_GROUP_SYSTEM);
fs/btrfs/volumes.c
7684
len = btrfs_chunk_item_size(btrfs_chunk_num_stripes(sb, chunk));
fs/btrfs/volumes.c
7688
ret = read_one_chunk(&key, sb, chunk);
fs/btrfs/volumes.c
7696
clear_extent_buffer_uptodate(sb);
fs/btrfs/volumes.c
7697
free_extent_buffer_stale(sb);
fs/btrfs/volumes.c
8518
guard(super_write)(fs_info->sb);
fs/btrfs/volumes.h
893
const u8 *btrfs_sb_fsid_ptr(const struct btrfs_super_block *sb);
fs/btrfs/zoned.c
2776
if (sb_rdonly(fs_info->sb))
fs/ceph/caps.c
4451
inode = ceph_find_inode(mdsc->fsc->sb, vino);
fs/ceph/crypto.c
77
static const union fscrypt_policy *ceph_get_dummy_policy(struct super_block *sb)
fs/ceph/crypto.c
79
return ceph_sb_to_fs_client(sb)->fsc_dummy_enc_policy.policy;
fs/ceph/crypto.c
92
void ceph_fscrypt_set_ops(struct super_block *sb)
fs/ceph/crypto.c
94
fscrypt_set_ops(sb, &ceph_fscrypt_ops);
fs/ceph/crypto.h
167
static inline void ceph_fscrypt_set_ops(struct super_block *sb)
fs/ceph/crypto.h
93
void ceph_fscrypt_set_ops(struct super_block *sb);
fs/ceph/debugfs.c
424
bdi_dev_name(fsc->sb->s_bdi));
fs/ceph/export.c
131
static struct inode *__lookup_inode(struct super_block *sb, u64 ino)
fs/ceph/export.c
133
struct ceph_mds_client *mdsc = ceph_sb_to_fs_client(sb)->mdsc;
fs/ceph/export.c
144
inode = ceph_find_inode(sb, vino);
fs/ceph/export.c
155
if (ceph_security_xattr_wanted(d_inode(sb->s_root)))
fs/ceph/export.c
177
struct inode *ceph_lookup_inode(struct super_block *sb, u64 ino)
fs/ceph/export.c
179
struct inode *inode = __lookup_inode(sb, ino);
fs/ceph/export.c
189
static struct dentry *__fh_to_dentry(struct super_block *sb, u64 ino)
fs/ceph/export.c
191
struct inode *inode = __lookup_inode(sb, ino);
fs/ceph/export.c
211
static struct dentry *__snapfh_to_dentry(struct super_block *sb,
fs/ceph/export.c
215
struct ceph_mds_client *mdsc = ceph_sb_to_fs_client(sb)->mdsc;
fs/ceph/export.c
240
inode = ceph_find_inode(sb, vino);
fs/ceph/export.c
255
if (ceph_security_xattr_wanted(d_inode(sb->s_root)))
fs/ceph/export.c
301
static struct dentry *ceph_fh_to_dentry(struct super_block *sb,
fs/ceph/export.c
305
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/export.c
310
return __snapfh_to_dentry(sb, sfh, false);
fs/ceph/export.c
320
return __fh_to_dentry(sb, fh->ino);
fs/ceph/export.c
323
static struct dentry *__get_parent(struct super_block *sb,
fs/ceph/export.c
326
struct ceph_mds_client *mdsc = ceph_sb_to_fs_client(sb)->mdsc;
fs/ceph/export.c
348
if (ceph_security_xattr_wanted(d_inode(sb->s_root)))
fs/ceph/export.c
420
static struct dentry *ceph_fh_to_parent(struct super_block *sb,
fs/ceph/export.c
424
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/export.c
430
return __snapfh_to_dentry(sb, sfh, true);
fs/ceph/export.c
439
dentry = __get_parent(sb, NULL, cfh->ino);
fs/ceph/export.c
441
dentry = __fh_to_dentry(sb, cfh->parent_ino);
fs/ceph/file.c
180
prepare_open_request(struct super_block *sb, int flags, int create_mode)
fs/ceph/file.c
182
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(sb);
fs/ceph/inode.c
101
return ceph_get_inode(sb, vino, NULL);
fs/ceph/inode.c
1554
int ceph_fill_trace(struct super_block *sb, struct ceph_mds_request *req)
fs/ceph/inode.c
1560
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/inode.c
1580
parent_dir = ceph_get_reply_dir(sb, req->r_parent, rinfo);
fs/ceph/inode.c
177
struct inode *ceph_get_inode(struct super_block *sb, struct ceph_vino vino,
fs/ceph/inode.c
180
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(sb);
fs/ceph/inode.c
193
inode = iget5_locked(sb, (unsigned long)vino.ino,
fs/ceph/inode.c
3109
struct super_block *sb = inode->i_sb;
fs/ceph/inode.c
3144
stat->dev = sb->s_dev;
fs/ceph/inode.c
3149
if (ceph_test_mount_opt(ceph_sb_to_fs_client(sb), RBYTES)) {
fs/ceph/inode.c
3156
parent = ceph_lookup_inode(sb, ceph_ino(inode));
fs/ceph/inode.c
612
struct inode *ceph_alloc_inode(struct super_block *sb)
fs/ceph/inode.c
614
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/inode.c
618
ci = alloc_inode_sb(sb, ceph_inode_cachep, GFP_NOFS);
fs/ceph/inode.c
78
static struct inode *ceph_get_reply_dir(struct super_block *sb,
fs/ceph/mds_client.c
1336
if (!dir || dir->i_sb != mdsc->fsc->sb) {
fs/ceph/mds_client.c
1919
struct super_block *sb = fsc->sb;
fs/ceph/mds_client.c
1948
inode = ceph_find_inode(sb, vino);
fs/ceph/mds_client.c
3035
struct inode *correct_dir = ceph_get_inode(mdsc->fsc->sb, path_info1.vino, NULL);
fs/ceph/mds_client.c
3965
in = ceph_get_inode(mdsc->fsc->sb, tvino, in);
fs/ceph/mds_client.c
4005
err = ceph_fill_trace(mdsc->fsc->sb, req);
fs/ceph/mds_client.c
4941
if (mdsc->fsc->sb->s_root)
fs/ceph/mds_client.c
4942
shrink_dcache_parent(mdsc->fsc->sb->s_root);
fs/ceph/mds_client.c
5247
struct super_block *sb = mdsc->fsc->sb;
fs/ceph/mds_client.c
5275
inode = ceph_find_inode(sb, vino);
fs/ceph/mds_client.c
5409
ceph_force_reconnect(fsc->sb);
fs/ceph/mds_client.c
6186
ceph_umount_begin(mdsc->fsc->sb);
fs/ceph/mds_client.c
6256
ceph_umount_begin(mdsc->fsc->sb);
fs/ceph/quota.c
131
struct super_block *sb,
fs/ceph/quota.c
163
in = ceph_lookup_inode(sb, realm->ino);
fs/ceph/quota.c
24
struct super_block *sb = inode->i_sb;
fs/ceph/quota.c
25
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(sb);
fs/ceph/quota.c
26
struct inode *root = d_inode(sb->s_root);
fs/ceph/quota.c
44
struct super_block *sb = mdsc->fsc->sb;
fs/ceph/quota.c
500
get_quota_realm(mdsc, d_inode(fsc->sb->s_root), QUOTA_GET_MAX_BYTES,
fs/ceph/quota.c
64
inode = ceph_find_inode(sb, vino);
fs/ceph/snap.c
1021
struct super_block *sb = mdsc->fsc->sb;
fs/ceph/snap.c
1089
struct inode *inode = ceph_find_inode(sb, vino);
fs/ceph/super.c
1020
void ceph_umount_begin(struct super_block *sb)
fs/ceph/super.c
1022
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1094
static int ceph_apply_test_dummy_encryption(struct super_block *sb,
fs/ceph/super.c
1098
struct ceph_fs_client *fsc = sb->s_fs_info;
fs/ceph/super.c
1129
static int ceph_apply_test_dummy_encryption(struct super_block *sb,
fs/ceph/super.c
1151
if (!fsc->sb->s_root) {
fs/ceph/super.c
1166
err = ceph_apply_test_dummy_encryption(fsc->sb, fc,
fs/ceph/super.c
1180
fsc->sb->s_root = dget(root);
fs/ceph/super.c
1182
root = dget(fsc->sb->s_root);
fs/ceph/super.c
1207
fsc->sb = s;
fs/ceph/super.c
1224
fsc->sb = NULL;
fs/ceph/super.c
1231
static int ceph_compare_super(struct super_block *sb, struct fs_context *fc)
fs/ceph/super.c
1236
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1239
doutc(cl, "%p\n", sb);
fs/ceph/super.c
124
static int ceph_sync_fs(struct super_block *sb, int wait)
fs/ceph/super.c
1250
if (fc->sb_flags != (sb->s_flags & ~SB_BORN)) {
fs/ceph/super.c
126
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1273
static int ceph_setup_bdi(struct super_block *sb, struct ceph_fs_client *fsc)
fs/ceph/super.c
1277
err = super_setup_bdi_name(sb, "ceph-%ld",
fs/ceph/super.c
1283
sb->s_bdi->ra_pages = fsc->mount_options->rasize >> PAGE_SHIFT;
fs/ceph/super.c
1286
sb->s_bdi->io_pages = fsc->mount_options->rsize >> PAGE_SHIFT;
fs/ceph/super.c
1295
struct super_block *sb;
fs/ceph/super.c
1326
sb = sget_fc(fc, compare_super, ceph_set_super);
fs/ceph/super.c
1328
if (IS_ERR(sb)) {
fs/ceph/super.c
1329
err = PTR_ERR(sb);
fs/ceph/super.c
1333
if (ceph_sb_to_fs_client(sb) != fsc) {
fs/ceph/super.c
1335
fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1339
err = ceph_setup_bdi(sb, fsc);
fs/ceph/super.c
1352
fc->root = fsc->sb->s_root;
fs/ceph/super.c
1362
deactivate_locked_super(sb);
fs/ceph/super.c
1388
struct super_block *sb = fc->root->d_sb;
fs/ceph/super.c
1389
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1391
err = ceph_apply_test_dummy_encryption(sb, fc, fsopt);
fs/ceph/super.c
1413
sync_filesystem(sb);
fs/ceph/super.c
1601
int ceph_force_reconnect(struct super_block *sb)
fs/ceph/super.c
1603
struct ceph_fs_client *fsc = ceph_sb_to_fs_client(sb);
fs/ceph/super.c
1622
if (sb->s_root) {
fs/ceph/super.c
1623
err = __ceph_do_getattr(d_inode(sb->s_root), NULL,
fs/ceph/super.c
838
fsc->sb = NULL;
fs/ceph/super.h
1011
extern int ceph_force_reconnect(struct super_block *sb);
fs/ceph/super.h
1036
void ceph_umount_begin(struct super_block *sb);
fs/ceph/super.h
1057
extern struct inode *ceph_alloc_inode(struct super_block *sb);
fs/ceph/super.h
1066
extern struct inode *ceph_get_inode(struct super_block *sb,
fs/ceph/super.h
1081
extern int ceph_fill_trace(struct super_block *sb,
fs/ceph/super.h
1371
struct inode *ceph_lookup_inode(struct super_block *sb, u64 ino);
fs/ceph/super.h
147
struct super_block *sb;
fs/ceph/super.h
526
ceph_sb_to_fs_client(const struct super_block *sb)
fs/ceph/super.h
528
return (struct ceph_fs_client *)sb->s_fs_info;
fs/ceph/super.h
532
ceph_sb_to_mdsc(const struct super_block *sb)
fs/ceph/super.h
534
return (struct ceph_mds_client *)ceph_sb_to_fs_client(sb)->mdsc;
fs/ceph/super.h
594
static inline u64 ceph_present_ino(struct super_block *sb, u64 ino)
fs/ceph/super.h
596
if (unlikely(ceph_test_mount_opt(ceph_sb_to_fs_client(sb), INO32)))
fs/ceph/super.h
640
static inline struct inode *ceph_find_inode(struct super_block *sb,
fs/ceph/super.h
651
return ilookup5(sb, (unsigned long)vino.ino, ceph_ino_compare, &vino);
fs/coda/cache.c
55
void coda_cache_clear_all(struct super_block *sb)
fs/coda/cnode.c
104
error = venus_getattr(sb, fid, &attr);
fs/coda/cnode.c
108
inode = coda_iget(sb, fid, &attr);
fs/coda/cnode.c
140
struct inode *coda_fid_to_inode(struct CodaFid *fid, struct super_block *sb)
fs/coda/cnode.c
145
inode = ilookup5(sb, hash, coda_test_inode, fid);
fs/coda/cnode.c
167
struct inode *coda_cnode_makectl(struct super_block *sb)
fs/coda/cnode.c
169
struct inode *inode = new_inode(sb);
fs/coda/cnode.c
60
struct inode * coda_iget(struct super_block * sb, struct CodaFid * fid,
fs/coda/cnode.c
69
inode = iget5_locked(sb, hash, coda_test_inode, coda_set_inode, fid);
fs/coda/cnode.c
97
struct inode *coda_cnode_make(struct CodaFid *fid, struct super_block *sb)
fs/coda/coda_cache.h
17
void coda_cache_clear_all(struct super_block *sb);
fs/coda/coda_fs_i.h
53
struct inode *coda_iget(struct super_block *sb, struct CodaFid *fid, struct coda_vattr *attr);
fs/coda/coda_fs_i.h
54
struct inode *coda_cnode_makectl(struct super_block *sb);
fs/coda/coda_fs_i.h
55
struct inode *coda_fid_to_inode(struct CodaFid *fid, struct super_block *sb);
fs/coda/coda_psdev.h
42
static inline struct venus_comm *coda_vcp(struct super_block *sb)
fs/coda/coda_psdev.h
44
return (struct venus_comm *)((sb)->s_fs_info);
fs/coda/coda_psdev.h
48
int venus_rootfid(struct super_block *sb, struct CodaFid *fidp);
fs/coda/coda_psdev.h
49
int venus_getattr(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
52
int venus_lookup(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
55
int venus_close(struct super_block *sb, struct CodaFid *fid, int flags,
fs/coda/coda_psdev.h
57
int venus_open(struct super_block *sb, struct CodaFid *fid, int flags,
fs/coda/coda_psdev.h
59
int venus_mkdir(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/coda_psdev.h
62
int venus_create(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/coda_psdev.h
65
int venus_rmdir(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/coda_psdev.h
67
int venus_remove(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/coda_psdev.h
69
int venus_readlink(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
71
int venus_rename(struct super_block *sb, struct CodaFid *new_fid,
fs/coda/coda_psdev.h
75
int venus_link(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
77
int venus_symlink(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
79
int venus_access(struct super_block *sb, struct CodaFid *fid, int mask);
fs/coda/coda_psdev.h
80
int venus_pioctl(struct super_block *sb, struct CodaFid *fid,
fs/coda/coda_psdev.h
84
int venus_fsync(struct super_block *sb, struct CodaFid *fid);
fs/coda/coda_psdev.h
86
int venus_access_intent(struct super_block *sb, struct CodaFid *fid,
fs/coda/dir.c
43
struct super_block *sb = dir->i_sb;
fs/coda/dir.c
57
inode = coda_cnode_makectl(sb);
fs/coda/dir.c
61
int error = venus_lookup(sb, coda_i2f(dir), name, length,
fs/coda/dir.c
63
inode = !error ? coda_cnode_make(&fid, sb) : ERR_PTR(error);
fs/coda/inode.c
199
static int coda_fill_super(struct super_block *sb, struct fs_context *fc)
fs/coda/inode.c
224
vc->vc_sb = sb;
fs/coda/inode.c
227
sb->s_fs_info = vc;
fs/coda/inode.c
228
sb->s_flags |= SB_NOATIME;
fs/coda/inode.c
229
sb->s_blocksize = 4096; /* XXXXX what do we put here?? */
fs/coda/inode.c
230
sb->s_blocksize_bits = 12;
fs/coda/inode.c
231
sb->s_magic = CODA_SUPER_MAGIC;
fs/coda/inode.c
232
sb->s_op = &coda_super_operations;
fs/coda/inode.c
233
set_default_d_op(sb, &coda_dentry_operations);
fs/coda/inode.c
234
sb->s_time_gran = 1;
fs/coda/inode.c
235
sb->s_time_min = S64_MIN;
fs/coda/inode.c
236
sb->s_time_max = S64_MAX;
fs/coda/inode.c
238
error = super_setup_bdi(sb);
fs/coda/inode.c
243
error = venus_rootfid(sb, &fid);
fs/coda/inode.c
252
root = coda_cnode_make(&fid, sb);
fs/coda/inode.c
262
sb->s_root = d_make_root(root);
fs/coda/inode.c
263
if (!sb->s_root) {
fs/coda/inode.c
272
sb->s_fs_info = NULL;
fs/coda/inode.c
278
static void coda_put_super(struct super_block *sb)
fs/coda/inode.c
280
struct venus_comm *vcp = coda_vcp(sb);
fs/coda/inode.c
283
sb->s_fs_info = NULL;
fs/coda/inode.c
45
static struct inode *coda_alloc_inode(struct super_block *sb)
fs/coda/inode.c
48
ei = alloc_inode_sb(sb, coda_inode_cachep, GFP_KERNEL);
fs/coda/upcall.c
103
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
111
int venus_setattr(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
124
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
130
int venus_lookup(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
150
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
160
int venus_close(struct super_block *sb, struct CodaFid *fid, int flags,
fs/coda/upcall.c
174
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
180
int venus_open(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
193
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
201
int venus_mkdir(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/upcall.c
221
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
232
int venus_rename(struct super_block *sb, struct CodaFid *old_fid,
fs/coda/upcall.c
263
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
269
int venus_create(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/upcall.c
292
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
302
int venus_rmdir(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/upcall.c
319
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
325
int venus_remove(struct super_block *sb, struct CodaFid *dirfid,
fs/coda/upcall.c
341
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
347
int venus_readlink(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
362
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
379
int venus_link(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
399
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
405
int venus_symlink(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
434
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
440
int venus_fsync(struct super_block *sb, struct CodaFid *fid)
fs/coda/upcall.c
450
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
456
int venus_access(struct super_block *sb, struct CodaFid *fid, int mask)
fs/coda/upcall.c
468
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
475
int venus_pioctl(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
517
error = coda_upcall(coda_vcp(sb), SIZE(ioctl) + data->vi.in_size,
fs/coda/upcall.c
572
int venus_access_intent(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
594
error = coda_upcall(coda_vcp(sb), insize,
fs/coda/upcall.c
75
int venus_rootfid(struct super_block *sb, struct CodaFid *fidp)
fs/coda/upcall.c
84
error = coda_upcall(coda_vcp(sb), insize, &outsize, inp);
fs/coda/upcall.c
860
struct super_block *sb;
fs/coda/upcall.c
895
sb = vcp->vc_sb;
fs/coda/upcall.c
896
if (!sb || !sb->s_root)
fs/coda/upcall.c
901
coda_cache_clear_all(sb);
fs/coda/upcall.c
902
shrink_dcache_sb(sb);
fs/coda/upcall.c
903
if (d_really_is_positive(sb->s_root))
fs/coda/upcall.c
904
coda_flag_inode(d_inode(sb->s_root), C_FLUSH);
fs/coda/upcall.c
908
coda_cache_clear_all(sb);
fs/coda/upcall.c
92
int venus_getattr(struct super_block *sb, struct CodaFid *fid,
fs/coda/upcall.c
928
inode = coda_fid_to_inode(fid, sb);
fs/configfs/dir.c
1636
struct super_block *sb = dentry->d_sb;
fs/configfs/dir.c
1677
ino = iunique(sb, 2);
fs/configfs/mount.c
63
static int configfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/configfs/mount.c
68
sb->s_blocksize = PAGE_SIZE;
fs/configfs/mount.c
69
sb->s_blocksize_bits = PAGE_SHIFT;
fs/configfs/mount.c
70
sb->s_magic = CONFIGFS_MAGIC;
fs/configfs/mount.c
71
sb->s_op = &configfs_ops;
fs/configfs/mount.c
72
sb->s_time_gran = 1;
fs/configfs/mount.c
75
&configfs_root, sb);
fs/configfs/mount.c
94
sb->s_root = root;
fs/configfs/mount.c
95
set_default_d_op(sb, &configfs_dentry_ops); /* the rest get that */
fs/configfs/mount.c
96
sb->s_d_flags |= DCACHE_DONTCACHE;
fs/configfs/symlink.c
118
struct super_block *sb)
fs/configfs/symlink.c
126
if (path.dentry->d_sb != sb)
fs/cramfs/inode.c
107
CRAMFS_SB(sb)->flags & CRAMFS_FLAG_EXT_BLOCK_POINTERS &&
fs/cramfs/inode.c
108
CRAMFS_SB(sb)->linear_phys_addr)
fs/cramfs/inode.c
192
static void *cramfs_blkdev_read(struct super_block *sb, unsigned int offset,
fs/cramfs/inode.c
195
struct address_space *mapping = sb->s_bdev->bd_mapping;
fs/cramfs/inode.c
211
if (buffer_dev[i] != sb)
fs/cramfs/inode.c
223
devsize = bdev_nr_bytes(sb->s_bdev) >> PAGE_SHIFT;
fs/cramfs/inode.c
244
buffer_dev[buffer] = sb;
fs/cramfs/inode.c
263
static void *cramfs_direct_read(struct super_block *sb, unsigned int offset,
fs/cramfs/inode.c
266
struct cramfs_sb_info *sbi = CRAMFS_SB(sb);
fs/cramfs/inode.c
279
static void *cramfs_read(struct super_block *sb, unsigned int offset,
fs/cramfs/inode.c
282
struct cramfs_sb_info *sbi = CRAMFS_SB(sb);
fs/cramfs/inode.c
285
return cramfs_direct_read(sb, offset, len);
fs/cramfs/inode.c
287
return cramfs_blkdev_read(sb, offset, len);
fs/cramfs/inode.c
457
struct super_block *sb = inode->i_sb;
fs/cramfs/inode.c
458
struct cramfs_sb_info *sbi = CRAMFS_SB(sb);
fs/cramfs/inode.c
494
static void cramfs_kill_sb(struct super_block *sb)
fs/cramfs/inode.c
496
struct cramfs_sb_info *sbi = CRAMFS_SB(sb);
fs/cramfs/inode.c
498
generic_shutdown_super(sb);
fs/cramfs/inode.c
50
static inline struct cramfs_sb_info *CRAMFS_SB(struct super_block *sb)
fs/cramfs/inode.c
500
if (IS_ENABLED(CONFIG_CRAMFS_MTD) && sb->s_mtd) {
fs/cramfs/inode.c
502
mtd_unpoint(sb->s_mtd, 0, sbi->mtd_point_size);
fs/cramfs/inode.c
503
put_mtd_device(sb->s_mtd);
fs/cramfs/inode.c
504
sb->s_mtd = NULL;
fs/cramfs/inode.c
505
} else if (IS_ENABLED(CONFIG_CRAMFS_BLOCKDEV) && sb->s_bdev) {
fs/cramfs/inode.c
506
sync_blockdev(sb->s_bdev);
fs/cramfs/inode.c
507
bdev_fput(sb->s_bdev_file);
fs/cramfs/inode.c
519
static int cramfs_read_super(struct super_block *sb, struct fs_context *fc,
fs/cramfs/inode.c
52
return sb->s_fs_info;
fs/cramfs/inode.c
522
struct cramfs_sb_info *sbi = CRAMFS_SB(sb);
fs/cramfs/inode.c
531
memcpy(super, cramfs_read(sb, 0, sizeof(*super)), sizeof(*super));
fs/cramfs/inode.c
546
cramfs_read(sb, 512, sizeof(*super)),
fs/cramfs/inode.c
597
static int cramfs_finalize_super(struct super_block *sb,
fs/cramfs/inode.c
603
sb->s_flags |= SB_RDONLY;
fs/cramfs/inode.c
604
sb->s_time_min = 0;
fs/cramfs/inode.c
605
sb->s_time_max = 0;
fs/cramfs/inode.c
606
sb->s_op = &cramfs_ops;
fs/cramfs/inode.c
607
root = get_cramfs_inode(sb, cramfs_root, 0);
fs/cramfs/inode.c
610
sb->s_root = d_make_root(root);
fs/cramfs/inode.c
611
if (!sb->s_root)
fs/cramfs/inode.c
616
static int cramfs_blkdev_fill_super(struct super_block *sb, struct fs_context *fc)
fs/cramfs/inode.c
625
sb->s_fs_info = sbi;
fs/cramfs/inode.c
631
err = cramfs_read_super(sb, fc, &super);
fs/cramfs/inode.c
634
return cramfs_finalize_super(sb, &super.root);
fs/cramfs/inode.c
637
static int cramfs_mtd_fill_super(struct super_block *sb, struct fs_context *fc)
fs/cramfs/inode.c
646
sb->s_fs_info = sbi;
fs/cramfs/inode.c
649
err = mtd_point(sb->s_mtd, 0, PAGE_SIZE, &sbi->mtd_point_size,
fs/cramfs/inode.c
653
sb->s_mtd->name);
fs/cramfs/inode.c
659
err = cramfs_read_super(sb, fc, &super);
fs/cramfs/inode.c
665
sb->s_mtd->name, sbi->size/1024);
fs/cramfs/inode.c
666
mtd_unpoint(sb->s_mtd, 0, PAGE_SIZE);
fs/cramfs/inode.c
667
err = mtd_point(sb->s_mtd, 0, sbi->size, &sbi->mtd_point_size,
fs/cramfs/inode.c
671
sb->s_mtd->name);
fs/cramfs/inode.c
675
return cramfs_finalize_super(sb, &super.root);
fs/cramfs/inode.c
680
struct super_block *sb = dentry->d_sb;
fs/cramfs/inode.c
683
if (sb->s_bdev)
fs/cramfs/inode.c
684
id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/cramfs/inode.c
685
else if (sb->s_dev)
fs/cramfs/inode.c
686
id = huge_encode_dev(sb->s_dev);
fs/cramfs/inode.c
690
buf->f_blocks = CRAMFS_SB(sb)->blocks;
fs/cramfs/inode.c
693
buf->f_files = CRAMFS_SB(sb)->files;
fs/cramfs/inode.c
706
struct super_block *sb = inode->i_sb;
fs/cramfs/inode.c
731
de = cramfs_read(sb, OFFSET(inode) + offset, sizeof(*de)+CRAMFS_MAXPATHLEN);
fs/cramfs/inode.c
834
struct super_block *sb = inode->i_sb;
fs/cramfs/inode.c
840
block_ptr = *(u32 *) cramfs_read(sb, blkptr_offset, 4);
fs/cramfs/inode.c
861
cramfs_read(sb, block_start, 2);
fs/cramfs/inode.c
875
cramfs_read(sb, blkptr_offset - 4, 4);
fs/cramfs/inode.c
886
cramfs_read(sb, block_start, 2);
fs/cramfs/inode.c
90
static struct inode *get_cramfs_inode(struct super_block *sb,
fs/cramfs/inode.c
903
cramfs_read(sb, block_start, block_len),
fs/cramfs/inode.c
909
cramfs_read(sb, block_start, block_len),
fs/cramfs/inode.c
96
inode = iget_locked(sb, cramino(cramfs_inode, offset));
fs/crypto/crypto.c
324
int fscrypt_initialize(struct super_block *sb)
fs/crypto/crypto.c
334
if (!sb->s_cop->needs_bounce_pages)
fs/crypto/fscrypt_private.h
336
int fscrypt_initialize(struct super_block *sb);
fs/crypto/fscrypt_private.h
373
fscrypt_max_file_dun_bits(const struct super_block *sb, int du_bits)
fs/crypto/fscrypt_private.h
375
return fls64(sb->s_maxbytes - 1) - du_bits;
fs/crypto/fscrypt_private.h
424
void fscrypt_destroy_inline_crypt_key(struct super_block *sb,
fs/crypto/fscrypt_private.h
427
int fscrypt_derive_sw_secret(struct super_block *sb,
fs/crypto/fscrypt_private.h
477
fscrypt_destroy_inline_crypt_key(struct super_block *sb,
fs/crypto/fscrypt_private.h
483
fscrypt_derive_sw_secret(struct super_block *sb,
fs/crypto/fscrypt_private.h
686
void fscrypt_put_master_key_activeref(struct super_block *sb,
fs/crypto/fscrypt_private.h
690
fscrypt_find_master_key(struct super_block *sb,
fs/crypto/fscrypt_private.h
696
int fscrypt_add_test_dummy_key(struct super_block *sb,
fs/crypto/fscrypt_private.h
699
int fscrypt_verify_key_added(struct super_block *sb,
fs/crypto/fscrypt_private.h
723
void fscrypt_destroy_prepared_key(struct super_block *sb,
fs/crypto/fscrypt_private.h
779
const union fscrypt_policy *fscrypt_get_dummy_policy(struct super_block *sb);
fs/crypto/inline_crypt.c
112
if (!(sb->s_flags & SB_INLINECRYPT))
fs/crypto/inline_crypt.c
125
sb->s_blocksize != PAGE_SIZE)
fs/crypto/inline_crypt.c
138
devs = fscrypt_get_devices(sb, &num_devs);
fs/crypto/inline_crypt.c
162
struct super_block *sb = inode->i_sb;
fs/crypto/inline_crypt.c
185
devs = fscrypt_get_devices(sb, &num_devs);
fs/crypto/inline_crypt.c
215
void fscrypt_destroy_inline_crypt_key(struct super_block *sb,
fs/crypto/inline_crypt.c
227
devs = fscrypt_get_devices(sb, &num_devs);
fs/crypto/inline_crypt.c
241
int fscrypt_derive_sw_secret(struct super_block *sb,
fs/crypto/inline_crypt.c
248
if (!(sb->s_flags & SB_INLINECRYPT)) {
fs/crypto/inline_crypt.c
25
static struct block_device **fscrypt_get_devices(struct super_block *sb,
fs/crypto/inline_crypt.c
251
sb->s_id);
fs/crypto/inline_crypt.c
255
err = blk_crypto_derive_sw_secret(sb->s_bdev, wrapped_key,
fs/crypto/inline_crypt.c
260
sb->s_id);
fs/crypto/inline_crypt.c
30
if (sb->s_cop->get_devices) {
fs/crypto/inline_crypt.c
31
devs = sb->s_cop->get_devices(sb, num_devs);
fs/crypto/inline_crypt.c
38
devs[0] = sb->s_bdev;
fs/crypto/inline_crypt.c
45
const struct super_block *sb = ci->ci_inode->i_sb;
fs/crypto/inline_crypt.c
59
dun_bits = fscrypt_max_file_dun_bits(sb, ci->ci_data_unit_bits);
fs/crypto/inline_crypt.c
97
struct super_block *sb = inode->i_sb;
fs/crypto/keyring.c
100
if (WARN_ON_ONCE(!sb->s_master_keys))
fs/crypto/keyring.c
1001
sb->s_id, busy_count, master_key_spec_type(&mk->mk_spec),
fs/crypto/keyring.c
1007
static int try_to_lock_encrypted_files(struct super_block *sb,
fs/crypto/keyring.c
102
spin_lock(&sb->s_master_keys->lock);
fs/crypto/keyring.c
1022
down_read(&sb->s_umount);
fs/crypto/keyring.c
1023
err1 = sync_filesystem(sb);
fs/crypto/keyring.c
1024
up_read(&sb->s_umount);
fs/crypto/keyring.c
104
spin_unlock(&sb->s_master_keys->lock);
fs/crypto/keyring.c
1042
err2 = check_for_busy_inodes(sb, mk);
fs/crypto/keyring.c
1069
struct super_block *sb = file_inode(filp)->i_sb;
fs/crypto/keyring.c
1095
mk = fscrypt_find_master_key(sb, &arg.key_spec);
fs/crypto/keyring.c
1127
fscrypt_initiate_key_removal(sb, mk);
fs/crypto/keyring.c
1135
err = try_to_lock_encrypted_files(sb, mk);
fs/crypto/keyring.c
115
sb, &mk->mk_direct_keys[i]);
fs/crypto/keyring.c
117
sb, &mk->mk_iv_ino_lblk_64_keys[i]);
fs/crypto/keyring.c
119
sb, &mk->mk_iv_ino_lblk_32_keys[i]);
fs/crypto/keyring.c
1193
struct super_block *sb = file_inode(filp)->i_sb;
fs/crypto/keyring.c
1211
mk = fscrypt_find_master_key(sb, &arg.key_spec);
fs/crypto/keyring.c
133
static void fscrypt_initiate_key_removal(struct super_block *sb,
fs/crypto/keyring.c
138
fscrypt_put_master_key_activeref(sb, mk);
fs/crypto/keyring.c
205
static int allocate_filesystem_keyring(struct super_block *sb)
fs/crypto/keyring.c
209
if (sb->s_master_keys)
fs/crypto/keyring.c
221
smp_store_release(&sb->s_master_keys, keyring);
fs/crypto/keyring.c
235
void fscrypt_destroy_keyring(struct super_block *sb)
fs/crypto/keyring.c
237
struct fscrypt_keyring *keyring = sb->s_master_keys;
fs/crypto/keyring.c
260
fscrypt_initiate_key_removal(sb, mk);
fs/crypto/keyring.c
264
sb->s_master_keys = NULL;
fs/crypto/keyring.c
289
fscrypt_find_master_key(struct super_block *sb,
fs/crypto/keyring.c
302
keyring = smp_load_acquire(&sb->s_master_keys);
fs/crypto/keyring.c
429
static int add_new_master_key(struct super_block *sb,
fs/crypto/keyring.c
433
struct fscrypt_keyring *keyring = sb->s_master_keys;
fs/crypto/keyring.c
515
static int do_add_master_key(struct super_block *sb,
fs/crypto/keyring.c
525
mk = fscrypt_find_master_key(sb, mk_spec);
fs/crypto/keyring.c
528
err = allocate_filesystem_keyring(sb);
fs/crypto/keyring.c
530
err = add_new_master_key(sb, secret, mk_spec);
fs/crypto/keyring.c
546
err = add_new_master_key(sb, secret, mk_spec);
fs/crypto/keyring.c
554
static int add_master_key(struct super_block *sb,
fs/crypto/keyring.c
573
err = fscrypt_derive_sw_secret(sb, secret->bytes,
fs/crypto/keyring.c
601
return do_add_master_key(sb, secret, key_spec);
fs/crypto/keyring.c
754
struct super_block *sb = file_inode(filp)->i_sb;
fs/crypto/keyring.c
804
err = add_master_key(sb, &secret, &arg.key_spec);
fs/crypto/keyring.c
858
int fscrypt_add_test_dummy_key(struct super_block *sb,
fs/crypto/keyring.c
865
err = add_master_key(sb, &secret, key_spec);
fs/crypto/keyring.c
87
void fscrypt_put_master_key_activeref(struct super_block *sb,
fs/crypto/keyring.c
886
int fscrypt_verify_key_added(struct super_block *sb,
fs/crypto/keyring.c
897
mk = fscrypt_find_master_key(sb, &mk_spec);
fs/crypto/keyring.c
967
static int check_for_busy_inodes(struct super_block *sb,
fs/crypto/keysetup.c
177
void fscrypt_destroy_prepared_key(struct super_block *sb,
fs/crypto/keysetup.c
181
fscrypt_destroy_inline_crypt_key(sb, prep_key);
fs/crypto/keysetup.c
199
const struct super_block *sb = inode->i_sb;
fs/crypto/keysetup.c
204
u8 hkdf_info[sizeof(mode_num) + sizeof(sb->s_uuid)];
fs/crypto/keysetup.c
215
if (sb->s_flags & SB_INLINECRYPT)
fs/crypto/keysetup.c
248
BUILD_BUG_ON(sizeof(sb->s_uuid) != 16);
fs/crypto/keysetup.c
252
memcpy(&hkdf_info[hkdf_infolen], &sb->s_uuid,
fs/crypto/keysetup.c
253
sizeof(sb->s_uuid));
fs/crypto/keysetup.c
254
hkdf_infolen += sizeof(sb->s_uuid);
fs/crypto/keysetup.c
456
struct super_block *sb = ci->ci_inode->i_sb;
fs/crypto/keysetup.c
465
mk = fscrypt_find_master_key(sb, &mk_spec);
fs/crypto/keysetup.c
468
fscrypt_get_dummy_policy(sb);
fs/crypto/keysetup.c
478
err = fscrypt_add_test_dummy_key(sb, &mk_spec);
fs/crypto/keysetup.c
481
mk = fscrypt_find_master_key(sb, &mk_spec);
fs/crypto/keysetup_v1.c
296
const struct super_block *sb = ci->ci_inode->i_sb;
fs/crypto/keysetup_v1.c
304
if (key == ERR_PTR(-ENOKEY) && sb->s_cop->legacy_key_prefix) {
fs/crypto/keysetup_v1.c
305
key = find_and_lock_process_key(sb->s_cop->legacy_key_prefix,
fs/crypto/policy.c
127
struct super_block *sb = inode->i_sb;
fs/crypto/policy.c
147
if (!sb->s_cop->has_stable_inodes ||
fs/crypto/policy.c
148
!sb->s_cop->has_stable_inodes(sb)) {
fs/crypto/policy.c
151
type, sb->s_id);
fs/crypto/policy.c
162
if (!sb->s_cop->has_32bit_inodes) {
fs/crypto/policy.c
165
type, sb->s_id);
fs/crypto/policy.c
173
if (fscrypt_max_file_dun_bits(sb,
fs/crypto/policy.c
177
type, sb->s_id);
fs/crypto/policy.c
58
const union fscrypt_policy *fscrypt_get_dummy_policy(struct super_block *sb)
fs/crypto/policy.c
60
if (!sb->s_cop->get_dummy_policy)
fs/crypto/policy.c
62
return sb->s_cop->get_dummy_policy(sb);
fs/crypto/policy.c
881
struct super_block *sb)
fs/crypto/policy.c
883
const union fscrypt_policy *policy = fscrypt_get_dummy_policy(sb);
fs/dcache.c
1249
long prune_dcache_sb(struct super_block *sb, struct shrink_control *sc)
fs/dcache.c
1254
freed = list_lru_shrink_walk(&sb->s_dentry_lru, sc,
fs/dcache.c
1288
void shrink_dcache_sb(struct super_block *sb)
fs/dcache.c
1293
list_lru_walk(&sb->s_dentry_lru,
fs/dcache.c
1296
} while (list_lru_count(&sb->s_dentry_lru) > 0);
fs/dcache.c
1663
void shrink_dcache_for_umount(struct super_block *sb)
fs/dcache.c
1667
rwsem_assert_held_write(&sb->s_umount);
fs/dcache.c
1669
dentry = sb->s_root;
fs/dcache.c
1670
sb->s_root = NULL;
fs/dcache.c
1673
while (!hlist_bl_empty(&sb->s_roots)) {
fs/dcache.c
1674
dentry = dget(hlist_bl_entry(hlist_bl_first(&sb->s_roots), struct dentry, d_hash));
fs/dcache.c
1734
static struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name)
fs/dcache.c
1740
dentry = kmem_cache_alloc_lru(dentry_cache, &sb->s_dentry_lru,
fs/dcache.c
1783
dentry->d_sb = sb;
fs/dcache.c
1784
dentry->d_op = sb->__s_d_op;
fs/dcache.c
1785
dentry->d_flags = sb->s_d_flags;
fs/dcache.c
1835
struct dentry *d_alloc_anon(struct super_block *sb)
fs/dcache.c
1837
return __d_alloc(sb, NULL);
fs/dcache.c
1866
struct dentry *d_alloc_pseudo(struct super_block *sb, const struct qstr *name)
fs/dcache.c
1871
struct dentry *dentry = __d_alloc(sb, name);
fs/dcache.c
2059
struct super_block *sb;
fs/dcache.c
2067
sb = inode->i_sb;
fs/dcache.c
2073
new = d_alloc_anon(sb);
fs/dcache.c
2092
hlist_bl_lock(&sb->s_roots);
fs/dcache.c
2093
hlist_bl_add_head(&new->d_hash, &sb->s_roots);
fs/dcache.c
2094
hlist_bl_unlock(&sb->s_roots);
fs/debugfs/inode.c
148
static void _debugfs_apply_options(struct super_block *sb, bool remount)
fs/debugfs/inode.c
150
struct debugfs_fs_info *fsi = sb->s_fs_info;
fs/debugfs/inode.c
151
struct inode *inode = d_inode(sb->s_root);
fs/debugfs/inode.c
170
static void debugfs_apply_options(struct super_block *sb)
fs/debugfs/inode.c
172
_debugfs_apply_options(sb, false);
fs/debugfs/inode.c
175
static void debugfs_apply_options_remount(struct super_block *sb)
fs/debugfs/inode.c
177
_debugfs_apply_options(sb, true);
fs/debugfs/inode.c
182
struct super_block *sb = fc->root->d_sb;
fs/debugfs/inode.c
183
struct debugfs_fs_info *sb_opts = sb->s_fs_info;
fs/debugfs/inode.c
189
sync_filesystem(sb);
fs/debugfs/inode.c
193
debugfs_apply_options_remount(sb);
fs/debugfs/inode.c
222
static struct inode *debugfs_alloc_inode(struct super_block *sb)
fs/debugfs/inode.c
225
info = alloc_inode_sb(sb, debugfs_inode_cachep, GFP_KERNEL);
fs/debugfs/inode.c
268
static int debugfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/debugfs/inode.c
273
err = simple_fill_super(sb, DEBUGFS_MAGIC, debug_files);
fs/debugfs/inode.c
277
sb->s_op = &debugfs_super_operations;
fs/debugfs/inode.c
278
set_default_d_op(sb, &debugfs_dops);
fs/debugfs/inode.c
279
sb->s_d_flags |= DCACHE_DONTCACHE;
fs/debugfs/inode.c
281
debugfs_apply_options(sb);
fs/debugfs/inode.c
70
static struct inode *debugfs_get_inode(struct super_block *sb)
fs/debugfs/inode.c
72
struct inode *inode = new_inode(sb);
fs/devpts/inode.c
104
struct super_block *sb;
fs/devpts/inode.c
108
static inline struct pts_fs_info *DEVPTS_SB(struct super_block *sb)
fs/devpts/inode.c
110
return sb->s_fs_info;
fs/devpts/inode.c
115
struct super_block *sb;
fs/devpts/inode.c
124
sb = path->mnt->mnt_sb;
fs/devpts/inode.c
125
if ((sb->s_magic != DEVPTS_SUPER_MAGIC) ||
fs/devpts/inode.c
126
(path->mnt->mnt_root != sb->s_root))
fs/devpts/inode.c
185
struct super_block *sb;
fs/devpts/inode.c
204
sb = path.mnt->mnt_sb;
fs/devpts/inode.c
205
atomic_inc(&sb->s_active);
fs/devpts/inode.c
206
result = DEVPTS_SB(sb);
fs/devpts/inode.c
215
deactivate_super(fsi->sb);
fs/devpts/inode.c
259
static int mknod_ptmx(struct super_block *sb, struct fs_context *fc)
fs/devpts/inode.c
264
struct dentry *root = sb->s_root;
fs/devpts/inode.c
265
struct pts_fs_info *fsi = DEVPTS_SB(sb);
fs/devpts/inode.c
279
inode = new_inode(sb);
fs/devpts/inode.c
372
fsi->sb = s;
fs/devpts/inode.c
443
static void devpts_kill_sb(struct super_block *sb)
fs/devpts/inode.c
445
struct pts_fs_info *fsi = DEVPTS_SB(sb);
fs/devpts/inode.c
450
kill_anon_super(sb);
fs/devpts/inode.c
501
struct super_block *sb = fsi->sb;
fs/devpts/inode.c
507
root = sb->s_root;
fs/devpts/inode.c
510
inode = new_inode(sb);
fs/direct-io.c
570
struct super_block *sb = dio->inode->i_sb;
fs/direct-io.c
575
if (!sb->s_dio_done_wq)
fs/direct-io.c
576
return sb_init_dio_done_wq(sb);
fs/drop_caches.c
19
static void drop_pagecache_sb(struct super_block *sb, void *unused)
fs/drop_caches.c
23
spin_lock(&sb->s_inode_list_lock);
fs/drop_caches.c
24
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/drop_caches.c
38
spin_unlock(&sb->s_inode_list_lock);
fs/drop_caches.c
45
spin_lock(&sb->s_inode_list_lock);
fs/drop_caches.c
47
spin_unlock(&sb->s_inode_list_lock);
fs/ecryptfs/crypto.c
1883
struct super_block *sb,
fs/ecryptfs/crypto.c
1887
&ecryptfs_superblock_to_private(sb)->mount_crypt_stat;
fs/ecryptfs/ecryptfs_kernel.h
461
ecryptfs_superblock_to_private(struct super_block *sb)
fs/ecryptfs/ecryptfs_kernel.h
463
return (struct ecryptfs_sb_info *)sb->s_fs_info;
fs/ecryptfs/ecryptfs_kernel.h
467
ecryptfs_set_superblock_private(struct super_block *sb,
fs/ecryptfs/ecryptfs_kernel.h
470
sb->s_fs_info = sb_info;
fs/ecryptfs/ecryptfs_kernel.h
474
ecryptfs_superblock_to_lower(struct super_block *sb)
fs/ecryptfs/ecryptfs_kernel.h
476
return ((struct ecryptfs_sb_info *)sb->s_fs_info)->wsi_sb;
fs/ecryptfs/ecryptfs_kernel.h
480
ecryptfs_set_superblock_lower(struct super_block *sb,
fs/ecryptfs/ecryptfs_kernel.h
483
((struct ecryptfs_sb_info *)sb->s_fs_info)->wsi_sb = lower_sb;
fs/ecryptfs/ecryptfs_kernel.h
538
struct super_block *sb);
fs/ecryptfs/ecryptfs_kernel.h
544
struct super_block *sb,
fs/ecryptfs/file.c
132
.sb = inode->i_sb,
fs/ecryptfs/file.c
73
struct super_block *sb;
fs/ecryptfs/file.c
92
buf->sb, lower_name,
fs/ecryptfs/inode.c
113
struct super_block *sb)
fs/ecryptfs/inode.c
115
struct inode *inode = __ecryptfs_get_inode(lower_inode, sb);
fs/ecryptfs/inode.c
134
struct dentry *dentry, struct super_block *sb)
fs/ecryptfs/inode.c
136
struct inode *inode = ecryptfs_get_inode(d_inode(lower_dentry), sb);
fs/ecryptfs/inode.c
83
struct super_block *sb)
fs/ecryptfs/inode.c
87
if (lower_inode->i_sb != ecryptfs_superblock_to_lower(sb))
fs/ecryptfs/inode.c
99
inode = iget5_locked(sb, (unsigned long)lower_inode,
fs/ecryptfs/main.c
577
static void ecryptfs_kill_block_super(struct super_block *sb)
fs/ecryptfs/main.c
579
struct ecryptfs_sb_info *sb_info = ecryptfs_superblock_to_private(sb);
fs/ecryptfs/main.c
580
kill_anon_super(sb);
fs/ecryptfs/super.c
131
struct super_block *sb = root->d_sb;
fs/ecryptfs/super.c
133
&ecryptfs_superblock_to_private(sb)->mount_crypt_stat;
fs/ecryptfs/super.c
36
static struct inode *ecryptfs_alloc_inode(struct super_block *sb)
fs/ecryptfs/super.c
41
inode_info = alloc_inode_sb(sb, ecryptfs_inode_info_cache, GFP_KERNEL);
fs/efivarfs/inode.c
19
struct inode *efivarfs_get_inode(struct super_block *sb,
fs/efivarfs/inode.c
23
struct inode *inode = new_inode(sb);
fs/efivarfs/inode.c
24
struct efivarfs_fs_info *fsi = sb->s_fs_info;
fs/efivarfs/internal.h
18
struct super_block *sb;
fs/efivarfs/internal.h
63
extern struct inode *efivarfs_get_inode(struct super_block *sb,
fs/efivarfs/super.c
125
static int efivarfs_freeze_fs(struct super_block *sb);
fs/efivarfs/super.c
126
static int efivarfs_unfreeze_fs(struct super_block *sb);
fs/efivarfs/super.c
216
struct super_block *sb = data;
fs/efivarfs/super.c
230
dentry = try_lookup_noperm(&QSTR(name), sb->s_root);
fs/efivarfs/super.c
238
static int efivarfs_create_dentry(struct super_block *sb, efi_char16_t *name16,
fs/efivarfs/super.c
244
struct dentry *dentry, *root = sb->s_root;
fs/efivarfs/super.c
256
inode = efivarfs_get_inode(sb, d_inode(root), S_IFREG | 0644, 0,
fs/efivarfs/super.c
297
struct super_block *sb = (struct super_block *)data;
fs/efivarfs/super.c
307
return efivarfs_create_dentry(sb, name16, name_size, vendor, name);
fs/efivarfs/super.c
33
sfi->sb->s_flags |= SB_RDONLY;
fs/efivarfs/super.c
345
static int efivarfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/efivarfs/super.c
347
struct efivarfs_fs_info *sfi = sb->s_fs_info;
fs/efivarfs/super.c
352
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/efivarfs/super.c
353
sb->s_blocksize = PAGE_SIZE;
fs/efivarfs/super.c
354
sb->s_blocksize_bits = PAGE_SHIFT;
fs/efivarfs/super.c
355
sb->s_magic = EFIVARFS_MAGIC;
fs/efivarfs/super.c
356
sb->s_op = &efivarfs_ops;
fs/efivarfs/super.c
357
set_default_d_op(sb, &efivarfs_d_ops);
fs/efivarfs/super.c
358
sb->s_d_flags |= DCACHE_DONTCACHE;
fs/efivarfs/super.c
359
sb->s_time_gran = 1;
fs/efivarfs/super.c
36
sfi->sb->s_flags &= ~SB_RDONLY;
fs/efivarfs/super.c
362
sb->s_flags |= SB_RDONLY;
fs/efivarfs/super.c
364
inode = efivarfs_get_inode(sb, NULL, S_IFDIR | 0755, 0, true);
fs/efivarfs/super.c
370
sb->s_root = root;
fs/efivarfs/super.c
374
sfi->sb = sb;
fs/efivarfs/super.c
380
return efivar_init(efivarfs_callback, sb, true);
fs/efivarfs/super.c
414
struct super_block *sb = data;
fs/efivarfs/super.c
425
dentry = try_lookup_noperm(&QSTR(name), sb->s_root);
fs/efivarfs/super.c
434
return efivarfs_create_dentry(sb, name16, name_size, vendor, name);
fs/efivarfs/super.c
448
static int efivarfs_freeze_fs(struct super_block *sb)
fs/efivarfs/super.c
45
static struct inode *efivarfs_alloc_inode(struct super_block *sb)
fs/efivarfs/super.c
454
static int efivarfs_unfreeze_fs(struct super_block *sb)
fs/efivarfs/super.c
472
child = find_next_child(sb->s_root, child);
fs/efivarfs/super.c
496
efivar_init(efivarfs_check_missing, sb, false);
fs/efivarfs/super.c
521
static void efivarfs_kill_sb(struct super_block *sb)
fs/efivarfs/super.c
523
struct efivarfs_fs_info *sfi = sb->s_fs_info;
fs/efivarfs/super.c
526
kill_anon_super(sb);
fs/efivarfs/super.c
67
struct super_block *sb = root->d_sb;
fs/efivarfs/super.c
68
struct efivarfs_fs_info *sbi = sb->s_fs_info;
fs/efs/efs.h
120
static inline struct efs_sb_info *SUPER_INFO(struct super_block *sb)
fs/efs/efs.h
122
return sb->s_fs_info;
fs/efs/efs.h
137
extern struct dentry *efs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/efs/efs.h
139
extern struct dentry *efs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/efs/inode.c
180
efs_extent_check(efs_extent *ptr, efs_block_t block, struct efs_sb_info *sb) {
fs/efs/inode.c
194
return(sb->fs_start + start + block - offset);
fs/efs/inode.c
201
struct efs_sb_info *sb = SUPER_INFO(inode->i_sb);
fs/efs/inode.c
214
if ((result = efs_extent_check(&in->extents[last], block, sb)))
fs/efs/inode.c
231
if ((result = efs_extent_check(&in->extents[cur], block, sb))) {
fs/efs/inode.c
271
iblock = sb->fs_start + in->extents[dirext].cooked.ex_bn +
fs/efs/inode.c
303
if ((result = efs_extent_check(&ext, block, sb))) {
fs/efs/inode.c
56
struct efs_sb_info *sb = SUPER_INFO(super);
fs/efs/inode.c
85
block = sb->fs_start + sb->first_block +
fs/efs/inode.c
86
(sb->group_size * (inode_index / sb->inode_blocks)) +
fs/efs/inode.c
87
(inode_index % sb->inode_blocks);
fs/efs/namei.c
103
struct dentry *efs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/efs/namei.c
106
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/efs/namei.c
77
static struct inode *efs_nfs_get_inode(struct super_block *sb, u64 ino,
fs/efs/namei.c
84
inode = efs_iget(sb, ino);
fs/efs/namei.c
96
struct dentry *efs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/efs/namei.c
99
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/efs/super.c
223
static int efs_validate_super(struct efs_sb_info *sb, struct efs_super *super) {
fs/efs/super.c
228
sb->fs_magic = be32_to_cpu(super->fs_magic);
fs/efs/super.c
229
sb->total_blocks = be32_to_cpu(super->fs_size);
fs/efs/super.c
230
sb->first_block = be32_to_cpu(super->fs_firstcg);
fs/efs/super.c
231
sb->group_size = be32_to_cpu(super->fs_cgfsize);
fs/efs/super.c
232
sb->data_free = be32_to_cpu(super->fs_tfree);
fs/efs/super.c
233
sb->inode_free = be32_to_cpu(super->fs_tinode);
fs/efs/super.c
234
sb->inode_blocks = be16_to_cpu(super->fs_cgisize);
fs/efs/super.c
235
sb->total_groups = be16_to_cpu(super->fs_ncg);
fs/efs/super.c
242
struct efs_sb_info *sb;
fs/efs/super.c
246
sb = kzalloc_obj(struct efs_sb_info);
fs/efs/super.c
247
if (!sb)
fs/efs/super.c
249
s->s_fs_info = sb;
fs/efs/super.c
274
sb->fs_start = efs_validate_vh((struct volume_header *) bh->b_data);
fs/efs/super.c
277
if (sb->fs_start == -1) {
fs/efs/super.c
281
bh = sb_bread(s, sb->fs_start + EFS_SUPER);
fs/efs/super.c
287
if (efs_validate_super(sb, (struct efs_super *) bh->b_data)) {
fs/efs/super.c
290
sb->fs_start + EFS_SUPER);
fs/efs/super.c
349
struct super_block *sb = dentry->d_sb;
fs/efs/super.c
350
struct efs_sb_info *sbi = SUPER_INFO(sb);
fs/efs/super.c
351
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/efs/super.c
65
static struct inode *efs_alloc_inode(struct super_block *sb)
fs/efs/super.c
68
ei = alloc_inode_sb(sb, efs_inode_cachep, GFP_KERNEL);
fs/erofs/compress.h
12
struct super_block *sb;
fs/erofs/compress.h
24
int (*config)(struct super_block *sb, struct erofs_super_block *dsb,
fs/erofs/data.c
110
pos = erofs_pos(sb, erofs_iblks(inode) - tailpacking);
fs/erofs/data.c
114
map->m_pa = erofs_pos(sb, vi->startblk) + map->m_la;
fs/erofs/data.c
118
vi->xattr_isize + erofs_blkoff(sb, map->m_la);
fs/erofs/data.c
134
idx = erofs_read_metabuf(&buf, sb, pos, erofs_inode_in_metabox(inode));
fs/erofs/data.c
149
EROFS_SB(sb)->device_id_mask;
fs/erofs/data.c
150
map->m_pa = erofs_pos(sb, startblk);
fs/erofs/data.c
156
map->m_pa = erofs_pos(sb, startblk);
fs/erofs/data.c
166
erofs_blkoff(sb, map->m_pa) + map->m_plen > blksz) {
fs/erofs/data.c
167
erofs_err(sb, "inline data across blocks @ nid %llu", vi->nid);
fs/erofs/data.c
177
struct super_block *sb, struct erofs_device_info *dif)
fs/erofs/data.c
179
map->m_sb = sb;
fs/erofs/data.c
186
int erofs_map_dev(struct super_block *sb, struct erofs_map_dev *map)
fs/erofs/data.c
188
struct erofs_dev_context *devs = EROFS_SB(sb)->devs;
fs/erofs/data.c
193
erofs_fill_from_devinfo(map, sb, &EROFS_SB(sb)->dif0);
fs/erofs/data.c
194
map->m_bdev = sb->s_bdev; /* use s_bdev for the primary device */
fs/erofs/data.c
203
map->m_pa += erofs_pos(sb, dif->uniaddr);
fs/erofs/data.c
207
erofs_fill_from_devinfo(map, sb, dif);
fs/erofs/data.c
215
startoff = erofs_pos(sb, dif->uniaddr);
fs/erofs/data.c
217
map->m_pa < startoff + erofs_pos(sb, dif->blocks)) {
fs/erofs/data.c
219
erofs_fill_from_devinfo(map, sb, dif);
fs/erofs/data.c
282
struct super_block *sb = realinode->i_sb;
fs/erofs/data.c
307
ret = erofs_map_dev(sb, &mdev);
fs/erofs/data.c
327
ptr = erofs_read_metabuf(&buf, sb, map.m_pa,
fs/erofs/data.c
53
int erofs_init_metabuf(struct erofs_buf *buf, struct super_block *sb,
fs/erofs/data.c
56
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/data.c
69
} else if (erofs_is_fscache_mode(sb))
fs/erofs/data.c
72
buf->mapping = sb->s_bdev->bd_mapping;
fs/erofs/data.c
76
void *erofs_read_metabuf(struct erofs_buf *buf, struct super_block *sb,
fs/erofs/data.c
81
err = erofs_init_metabuf(buf, sb, in_metabox);
fs/erofs/data.c
90
struct super_block *sb = inode->i_sb;
fs/erofs/data.c
91
unsigned int unit, blksz = sb->s_blocksize;
fs/erofs/decompressor.c
12
static int z_erofs_load_lz4_config(struct super_block *sb,
fs/erofs/decompressor.c
15
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/decompressor.c
21
erofs_err(sb, "invalid lz4 cfgs, size=%u", size);
fs/erofs/decompressor.c
213
rq->sb->s_blocksize - rq->pageofs_in));
fs/erofs/decompressor.c
219
(rq->sb->s_blocksize - 1));
fs/erofs/decompressor.c
294
const unsigned int bs = rq->sb->s_blocksize;
fs/erofs/decompressor.c
30
erofs_blknr(sb, Z_EROFS_PCLUSTER_MAX_SIZE)) {
fs/erofs/decompressor.c
31
erofs_err(sb, "too large lz4 pclusterblks %u",
fs/erofs/decompressor.c
447
int z_erofs_parse_cfgs(struct super_block *sb, struct erofs_super_block *dsb)
fs/erofs/decompressor.c
449
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/decompressor.c
456
return z_erofs_load_lz4_config(sb, dsb, NULL, 0);
fs/erofs/decompressor.c
461
erofs_err(sb, "unidentified algorithms %lx, please upgrade kernel",
fs/erofs/decompressor.c
466
(void)erofs_init_metabuf(&buf, sb, false);
fs/erofs/decompressor.c
472
data = erofs_read_metadata(sb, &buf, &offset, &size);
fs/erofs/decompressor.c
479
ret = dec->config(sb, dsb, data, size);
fs/erofs/decompressor.c
481
erofs_err(sb, "algorithm %ld isn't enabled on this kernel",
fs/erofs/decompressor.c
60
EROFS_SB(rq->sb)->lz4.max_distance_pages;
fs/erofs/decompressor_crypto.c
19
rq->sb->s_blocksize - rq->pageofs_in));
fs/erofs/decompressor_crypto.c
47
erofs_err(rq->sb, "failed to decompress %d in[%u, %u] out[%u]",
fs/erofs/decompressor_deflate.c
102
struct super_block *sb = rq->sb;
fs/erofs/decompressor_deflate.c
111
min(rq->inputsize, sb->s_blocksize - rq->pageofs_in));
fs/erofs/decompressor_deflate.c
52
static int z_erofs_load_deflate_config(struct super_block *sb,
fs/erofs/decompressor_deflate.c
60
erofs_err(sb, "invalid deflate cfgs, size=%u", size);
fs/erofs/decompressor_deflate.c
65
erofs_err(sb, "unsupported windowbits %u", dfl->windowbits);
fs/erofs/decompressor_lzma.c
152
struct super_block *sb = rq->sb;
fs/erofs/decompressor_lzma.c
162
min(rq->inputsize, sb->s_blocksize - rq->pageofs_in));
fs/erofs/decompressor_lzma.c
72
static int z_erofs_load_lzma_config(struct super_block *sb,
fs/erofs/decompressor_lzma.c
82
erofs_err(sb, "invalid lzma cfgs, size=%u", size);
fs/erofs/decompressor_lzma.c
86
erofs_err(sb, "unidentified lzma format %x, please check kernel version",
fs/erofs/decompressor_lzma.c
92
erofs_err(sb, "unsupported lzma dictionary size %u",
fs/erofs/decompressor_zstd.c
141
struct super_block *sb = rq->sb;
fs/erofs/decompressor_zstd.c
153
min(rq->inputsize, sb->s_blocksize - rq->pageofs_in));
fs/erofs/decompressor_zstd.c
75
static int z_erofs_load_zstd_config(struct super_block *sb,
fs/erofs/decompressor_zstd.c
85
erofs_err(sb, "unsupported zstd format, size=%u", size);
fs/erofs/decompressor_zstd.c
90
erofs_err(sb, "unsupported zstd window log %u", zstd->windowlog);
fs/erofs/dir.c
51
struct super_block *sb = dir->i_sb;
fs/erofs/dir.c
53
unsigned long bsz = sb->s_blocksize;
fs/erofs/dir.c
54
unsigned int ofs = erofs_blkoff(sb, ctx->pos);
fs/erofs/dir.c
84
erofs_err(sb, "failed to readdir of logical block %llu of nid %llu",
fs/erofs/dir.c
85
erofs_blknr(sb, dbstart), EROFS_I(dir)->nid);
fs/erofs/dir.c
92
erofs_err(sb, "invalid de[0].nameoff %u @ nid %llu",
fs/erofs/fileio.c
12
struct super_block *sb;
fs/erofs/fileio.c
54
if (test_opt(&EROFS_SB(rq->sb)->opt, DIRECT_IO) &&
fs/erofs/fileio.c
73
rq->sb = mdev->m_sb;
fs/erofs/fscache.c
245
struct super_block *sb = inode->i_sb;
fs/erofs/fscache.c
264
src = erofs_read_metabuf(&buf, sb, map.m_pa,
fs/erofs/fscache.c
297
ret = erofs_map_dev(sb, &mdev);
fs/erofs/fscache.c
389
static int erofs_fscache_register_volume(struct super_block *sb)
fs/erofs/fscache.c
391
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/fscache.c
404
erofs_err(sb, "failed to register volume for %s", name);
fs/erofs/fscache.c
414
static int erofs_fscache_init_domain(struct super_block *sb)
fs/erofs/fscache.c
418
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/fscache.c
430
err = erofs_fscache_register_volume(sb);
fs/erofs/fscache.c
454
static int erofs_fscache_register_domain(struct super_block *sb)
fs/erofs/fscache.c
458
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/fscache.c
470
err = erofs_fscache_init_domain(sb);
fs/erofs/fscache.c
475
static struct erofs_fscache *erofs_fscache_acquire_cookie(struct super_block *sb,
fs/erofs/fscache.c
478
struct fscache_volume *volume = EROFS_SB(sb)->volume;
fs/erofs/fscache.c
494
erofs_err(sb, "failed to get cookie for %s", name);
fs/erofs/fscache.c
504
isb = flags & EROFS_REG_COOKIE_SHARE ? erofs_pseudo_mnt->mnt_sb : sb;
fs/erofs/fscache.c
507
erofs_err(sb, "failed to get anon inode for %s", name);
fs/erofs/fscache.c
515
inode->i_blkbits = EROFS_SB(sb)->blkszbits;
fs/erofs/fscache.c
539
static struct erofs_fscache *erofs_domain_init_cookie(struct super_block *sb,
fs/erofs/fscache.c
543
struct erofs_domain *domain = EROFS_SB(sb)->domain;
fs/erofs/fscache.c
545
ctx = erofs_fscache_acquire_cookie(sb, name, flags);
fs/erofs/fscache.c
561
static struct erofs_fscache *erofs_domain_register_cookie(struct super_block *sb,
fs/erofs/fscache.c
565
struct erofs_domain *domain = EROFS_SB(sb)->domain;
fs/erofs/fscache.c
575
erofs_err(sb, "%s already exists in domain %s", name,
fs/erofs/fscache.c
582
ctx = erofs_domain_init_cookie(sb, name, flags);
fs/erofs/fscache.c
587
struct erofs_fscache *erofs_fscache_register_cookie(struct super_block *sb,
fs/erofs/fscache.c
591
if (EROFS_SB(sb)->domain_id)
fs/erofs/fscache.c
592
return erofs_domain_register_cookie(sb, name, flags);
fs/erofs/fscache.c
593
return erofs_fscache_acquire_cookie(sb, name, flags);
fs/erofs/fscache.c
616
int erofs_fscache_register_fs(struct super_block *sb)
fs/erofs/fscache.c
619
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/fscache.c
624
ret = erofs_fscache_register_domain(sb);
fs/erofs/fscache.c
626
ret = erofs_fscache_register_volume(sb);
fs/erofs/fscache.c
642
fscache = erofs_fscache_register_cookie(sb, sbi->fsid, flags);
fs/erofs/fscache.c
650
void erofs_fscache_unregister_fs(struct super_block *sb)
fs/erofs/fscache.c
652
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/inode.c
137
erofs_err(sb, "unsupported on-disk inode version %u of nid %llu",
fs/erofs/inode.c
144
erofs_err(sb, "negative i_size @ nid %llu", vi->nid);
fs/erofs/inode.c
180
erofs_err(sb, "bogus i_mode (%o) @ nid %llu", inode->i_mode,
fs/erofs/inode.c
187
inode->i_blocks = round_up(inode->i_size, sb->s_blocksize) >> 9;
fs/erofs/inode.c
189
erofs_err(sb, "compressed inode (nid %llu) is invalid in a plain filesystem",
fs/erofs/inode.c
195
(sb->s_blocksize_bits - 9);
fs/erofs/inode.c
202
erofs_err(sb, "unsupported chunk format %x of nid %llu",
fs/erofs/inode.c
207
vi->chunkbits = sb->s_blocksize_bits +
fs/erofs/inode.c
269
static ino_t erofs_squash_ino(struct super_block *sb, erofs_nid_t nid)
fs/erofs/inode.c
271
u64 ino64 = erofs_nid_to_ino64(EROFS_SB(sb), nid);
fs/erofs/inode.c
292
struct inode *erofs_iget(struct super_block *sb, erofs_nid_t nid)
fs/erofs/inode.c
296
inode = iget5_locked(sb, erofs_squash_ino(sb, nid), erofs_iget5_eq,
fs/erofs/inode.c
38
struct super_block *sb = inode->i_sb;
fs/erofs/inode.c
39
erofs_blk_t blkaddr = erofs_blknr(sb, erofs_iloc(inode));
fs/erofs/inode.c
40
unsigned int ofs = erofs_blkoff(sb, erofs_iloc(inode));
fs/erofs/inode.c
43
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/inode.c
52
ptr = erofs_read_metabuf(&buf, sb, erofs_pos(sb, blkaddr), in_mbox);
fs/erofs/inode.c
55
erofs_err(sb, "failed to read inode meta block (nid: %llu): %d",
fs/erofs/inode.c
63
erofs_err(sb, "unsupported i_format %u of nid %llu",
fs/erofs/inode.c
71
erofs_err(sb, "unsupported datalayout %u of nid %llu",
fs/erofs/inode.c
81
if (ofs + vi->inode_isize <= sb->s_blocksize) {
fs/erofs/inode.c
87
const unsigned int gotten = sb->s_blocksize - ofs;
fs/erofs/inode.c
90
ptr = erofs_read_metabuf(&buf, sb,
fs/erofs/inode.c
91
erofs_pos(sb, blkaddr + 1), in_mbox);
fs/erofs/inode.c
94
erofs_err(sb, "failed to read inode payload block (nid: %llu): %d",
fs/erofs/internal.h
170
#define EROFS_SB(sb) ((struct erofs_sb_info *)(sb)->s_fs_info)
fs/erofs/internal.h
192
static inline bool erofs_is_fscache_mode(struct super_block *sb)
fs/erofs/internal.h
195
!erofs_is_fileio_mode(EROFS_SB(sb)) && !sb->s_bdev;
fs/erofs/internal.h
213
#define erofs_blknr(sb, pos) ((erofs_blk_t)((pos) >> (sb)->s_blocksize_bits))
fs/erofs/internal.h
214
#define erofs_blkoff(sb, pos) ((pos) & ((sb)->s_blocksize - 1))
fs/erofs/internal.h
215
#define erofs_pos(sb, blk) ((erofs_off_t)(blk) << (sb)->s_blocksize_bits)
fs/erofs/internal.h
23
__printf(2, 3) void _erofs_printk(struct super_block *sb, const char *fmt, ...);
fs/erofs/internal.h
24
#define erofs_err(sb, fmt, ...) \
fs/erofs/internal.h
25
_erofs_printk(sb, KERN_ERR fmt "\n", ##__VA_ARGS__)
fs/erofs/internal.h
26
#define erofs_info(sb, fmt, ...) \
fs/erofs/internal.h
27
_erofs_printk(sb, KERN_INFO fmt "\n", ##__VA_ARGS__)
fs/erofs/internal.h
436
void *erofs_read_metadata(struct super_block *sb, struct erofs_buf *buf,
fs/erofs/internal.h
441
int erofs_init_metabuf(struct erofs_buf *buf, struct super_block *sb,
fs/erofs/internal.h
443
void *erofs_read_metabuf(struct erofs_buf *buf, struct super_block *sb,
fs/erofs/internal.h
445
int erofs_map_dev(struct super_block *sb, struct erofs_map_dev *dev);
fs/erofs/internal.h
452
struct inode *erofs_iget(struct super_block *sb, erofs_nid_t nid);
fs/erofs/internal.h
494
int erofs_register_sysfs(struct super_block *sb);
fs/erofs/internal.h
495
void erofs_unregister_sysfs(struct super_block *sb);
fs/erofs/internal.h
515
void erofs_shrinker_register(struct super_block *sb);
fs/erofs/internal.h
516
void erofs_shrinker_unregister(struct super_block *sb);
fs/erofs/internal.h
521
int z_erofs_init_super(struct super_block *sb);
fs/erofs/internal.h
532
static inline void erofs_shrinker_register(struct super_block *sb) {}
fs/erofs/internal.h
533
static inline void erofs_shrinker_unregister(struct super_block *sb) {}
fs/erofs/internal.h
538
static inline int z_erofs_init_super(struct super_block *sb) { return 0; }
fs/erofs/internal.h
540
int z_erofs_parse_cfgs(struct super_block *sb, struct erofs_super_block *dsb);
fs/erofs/internal.h
551
int erofs_fscache_register_fs(struct super_block *sb);
fs/erofs/internal.h
552
void erofs_fscache_unregister_fs(struct super_block *sb);
fs/erofs/internal.h
554
struct erofs_fscache *erofs_fscache_register_cookie(struct super_block *sb,
fs/erofs/internal.h
560
static inline int erofs_fscache_register_fs(struct super_block *sb)
fs/erofs/internal.h
564
static inline void erofs_fscache_unregister_fs(struct super_block *sb) {}
fs/erofs/internal.h
567
struct erofs_fscache *erofs_fscache_register_cookie(struct super_block *sb,
fs/erofs/super.c
1060
struct super_block *sb = dentry->d_sb;
fs/erofs/super.c
1061
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
1063
buf->f_type = sb->s_magic;
fs/erofs/super.c
1064
buf->f_bsize = sb->s_blocksize;
fs/erofs/super.c
1071
if (uuid_is_null(&sb->s_uuid))
fs/erofs/super.c
1072
buf->f_fsid = u64_to_fsid(!sb->s_bdev ? 0 :
fs/erofs/super.c
1073
huge_encode_dev(sb->s_bdev->bd_dev));
fs/erofs/super.c
1075
buf->f_fsid = uuid_to_fsid(sb->s_uuid.b);
fs/erofs/super.c
112
cnt = min_t(int, sb->s_blocksize - erofs_blkoff(sb, *offset),
fs/erofs/super.c
125
static int erofs_init_device(struct erofs_buf *buf, struct super_block *sb,
fs/erofs/super.c
128
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
133
dis = erofs_read_metabuf(buf, sb, *pos, false);
fs/erofs/super.c
139
erofs_err(sb, "empty device tag @ pos %llu", *pos);
fs/erofs/super.c
147
if (erofs_is_fscache_mode(sb)) {
fs/erofs/super.c
148
fscache = erofs_fscache_register_cookie(sb, dif->path, 0);
fs/erofs/super.c
156
BLK_OPEN_READ, sb->s_type, NULL);
fs/erofs/super.c
171
erofs_info(sb, "DAX unsupported by %s. Turning off DAX.",
fs/erofs/super.c
185
static int erofs_scan_devices(struct super_block *sb,
fs/erofs/super.c
188
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
203
erofs_err(sb, "extra devices don't match (ondisk %u, given %u)",
fs/erofs/super.c
209
erofs_info(sb, "DAX unsupported by block device. Turning off DAX.");
fs/erofs/super.c
215
if (!sbi->devs->extra_devices && !erofs_is_fscache_mode(sb))
fs/erofs/super.c
22
void _erofs_printk(struct super_block *sb, const char *fmt, ...)
fs/erofs/super.c
223
err = erofs_init_device(&buf, sb, dif, &pos);
fs/erofs/super.c
242
err = erofs_init_device(&buf, sb, dif, &pos);
fs/erofs/super.c
252
static int erofs_read_superblock(struct super_block *sb)
fs/erofs/super.c
254
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
260
data = erofs_read_metabuf(&buf, sb, 0, false);
fs/erofs/super.c
262
erofs_err(sb, "cannot read erofs superblock");
fs/erofs/super.c
269
erofs_err(sb, "cannot find valid erofs superblock");
fs/erofs/super.c
275
erofs_err(sb, "blkszbits %u isn't supported", sbi->blkszbits);
fs/erofs/super.c
279
erofs_err(sb, "dirblkbits %u isn't supported", dsb->dirblkbits);
fs/erofs/super.c
285
ret = erofs_superblock_csum_verify(sb, data);
fs/erofs/super.c
293
erofs_err(sb, "unidentified incompatible feature %x, please upgrade kernel",
fs/erofs/super.c
300
erofs_err(sb, "invalid sb_extslots %u (more than a fs block)",
fs/erofs/super.c
313
erofs_err(sb, "invalid ishare xattr prefix id %u",
fs/erofs/super.c
33
if (sb)
fs/erofs/super.c
343
super_set_uuid(sb, (void *)dsb->uuid, sizeof(dsb->uuid));
fs/erofs/super.c
35
KERN_SOH_ASCII, level, sb->s_id, &vaf);
fs/erofs/super.c
355
ret = z_erofs_parse_cfgs(sb, dsb);
fs/erofs/super.c
360
erofs_err(sb, "compression disabled, unable to mount compressed EROFS");
fs/erofs/super.c
365
ret = erofs_scan_devices(sb, dsb);
fs/erofs/super.c
368
erofs_info(sb, "EXPERIMENTAL 48-bit layout support in use. Use at your own risk!");
fs/erofs/super.c
370
erofs_info(sb, "EXPERIMENTAL metadata compression support in use. Use at your own risk!");
fs/erofs/super.c
371
if (erofs_is_fscache_mode(sb))
fs/erofs/super.c
372
erofs_info(sb, "[deprecated] fscache-based on-demand read feature in use. Use at your own risk!");
fs/erofs/super.c
41
static int erofs_superblock_csum_verify(struct super_block *sb, void *sbdata)
fs/erofs/super.c
44
u32 len = 1 << EROFS_SB(sb)->blkszbits, crc;
fs/erofs/super.c
55
erofs_err(sb, "invalid checksum 0x%08x, 0x%08x expected",
fs/erofs/super.c
575
static struct dentry *erofs_fh_to_dentry(struct super_block *sb,
fs/erofs/super.c
582
return d_obtain_alias(erofs_iget(sb,
fs/erofs/super.c
586
static struct dentry *erofs_fh_to_parent(struct super_block *sb,
fs/erofs/super.c
592
return d_obtain_alias(erofs_iget(sb,
fs/erofs/super.c
615
static void erofs_set_sysfs_name(struct super_block *sb)
fs/erofs/super.c
617
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
620
super_set_sysfs_name_generic(sb, "%s,%s", sbi->domain_id,
fs/erofs/super.c
623
super_set_sysfs_name_generic(sb, "%s", sbi->fsid);
fs/erofs/super.c
625
super_set_sysfs_name_generic(sb, "%s",
fs/erofs/super.c
626
bdi_dev_name(sb->s_bdi));
fs/erofs/super.c
628
super_set_sysfs_name_id(sb);
fs/erofs/super.c
631
static int erofs_fc_fill_super(struct super_block *sb, struct fs_context *fc)
fs/erofs/super.c
634
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
637
sb->s_magic = EROFS_SUPER_MAGIC;
fs/erofs/super.c
638
sb->s_flags |= SB_RDONLY | SB_NOATIME;
fs/erofs/super.c
639
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/erofs/super.c
640
sb->s_op = &erofs_sops;
fs/erofs/super.c
652
if (!sb->s_bdev) {
fs/erofs/super.c
67
static struct inode *erofs_alloc_inode(struct super_block *sb)
fs/erofs/super.c
672
erofs_err(sb, "file-backed mounts cannot be applied to stacked fses");
fs/erofs/super.c
676
sb->s_blocksize = PAGE_SIZE;
fs/erofs/super.c
677
sb->s_blocksize_bits = PAGE_SHIFT;
fs/erofs/super.c
679
if (erofs_is_fscache_mode(sb)) {
fs/erofs/super.c
680
err = erofs_fscache_register_fs(sb);
fs/erofs/super.c
684
err = super_setup_bdi(sb);
fs/erofs/super.c
688
if (!sb_set_blocksize(sb, PAGE_SIZE)) {
fs/erofs/super.c
693
sbi->dif0.dax_dev = fs_dax_get_by_bdev(sb->s_bdev,
fs/erofs/super.c
697
err = erofs_read_superblock(sb);
fs/erofs/super.c
70
alloc_inode_sb(sb, erofs_inode_cachep, GFP_KERNEL);
fs/erofs/super.c
701
if (sb->s_blocksize_bits != sbi->blkszbits) {
fs/erofs/super.c
702
if (erofs_is_fscache_mode(sb)) {
fs/erofs/super.c
708
sb->s_blocksize = 1 << sbi->blkszbits;
fs/erofs/super.c
709
sb->s_blocksize_bits = sbi->blkszbits;
fs/erofs/super.c
710
} else if (!sb_set_blocksize(sb, 1 << sbi->blkszbits)) {
fs/erofs/super.c
717
if (sbi->dif0.fsoff & (sb->s_blocksize - 1))
fs/erofs/super.c
719
sbi->dif0.fsoff, sb->s_blocksize);
fs/erofs/super.c
720
if (erofs_is_fscache_mode(sb))
fs/erofs/super.c
725
erofs_info(sb, "unsupported blocksize for DAX");
fs/erofs/super.c
729
erofs_info(sb, "on-disk ishare xattrs not found. Turning off inode_share.");
fs/erofs/super.c
733
erofs_info(sb, "EXPERIMENTAL EROFS page cache share support in use. Use at your own risk!");
fs/erofs/super.c
735
sb->s_time_gran = 1;
fs/erofs/super.c
736
sb->s_xattr = erofs_xattr_handlers;
fs/erofs/super.c
737
sb->s_export_op = &erofs_export_ops;
fs/erofs/super.c
740
sb->s_flags |= SB_POSIXACL;
fs/erofs/super.c
742
sb->s_flags &= ~SB_POSIXACL;
fs/erofs/super.c
744
err = z_erofs_init_super(sb);
fs/erofs/super.c
749
inode = erofs_iget(sb, sbi->packed_nid);
fs/erofs/super.c
755
inode = erofs_iget(sb, sbi->metabox_nid);
fs/erofs/super.c
761
inode = erofs_iget(sb, sbi->root_nid);
fs/erofs/super.c
766
erofs_err(sb, "rootino(nid %llu) is not a directory(i_mode %o)",
fs/erofs/super.c
771
sb->s_root = d_make_root(inode);
fs/erofs/super.c
772
if (!sb->s_root)
fs/erofs/super.c
775
erofs_shrinker_register(sb);
fs/erofs/super.c
776
err = erofs_xattr_prefixes_init(sb);
fs/erofs/super.c
780
erofs_set_sysfs_name(sb);
fs/erofs/super.c
781
err = erofs_register_sysfs(sb);
fs/erofs/super.c
786
erofs_info(sb, "mounted with root inode @ nid %llu.", sbi->root_nid);
fs/erofs/super.c
820
struct super_block *sb = fc->root->d_sb;
fs/erofs/super.c
821
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
824
DBG_BUGON(!sb_rdonly(sb));
fs/erofs/super.c
827
erofs_info(sb, "ignoring reconfiguration for fsid|domain_id.");
fs/erofs/super.c
91
void *erofs_read_metadata(struct super_block *sb, struct erofs_buf *buf,
fs/erofs/super.c
923
static void erofs_kill_sb(struct super_block *sb)
fs/erofs/super.c
925
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/super.c
929
kill_anon_super(sb);
fs/erofs/super.c
931
kill_block_super(sb);
fs/erofs/super.c
934
erofs_fscache_unregister_fs(sb);
fs/erofs/super.c
936
sb->s_fs_info = NULL;
fs/erofs/super.c
939
static void erofs_put_super(struct super_block *sb)
fs/erofs/super.c
941
struct erofs_sb_info *const sbi = EROFS_SB(sb);
fs/erofs/super.c
943
erofs_unregister_sysfs(sb);
fs/erofs/super.c
944
erofs_shrinker_unregister(sb);
fs/erofs/super.c
945
erofs_xattr_prefixes_cleanup(sb);
fs/erofs/super.c
949
erofs_fscache_unregister_fs(sb);
fs/erofs/sysfs.c
254
int erofs_register_sysfs(struct super_block *sb)
fs/erofs/sysfs.c
256
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/sysfs.c
262
sb->s_sysfs_name);
fs/erofs/sysfs.c
270
void erofs_unregister_sysfs(struct super_block *sb)
fs/erofs/sysfs.c
272
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/xattr.c
12
struct super_block *sb;
fs/erofs/xattr.c
121
struct super_block *sb = it->sb;
fs/erofs/xattr.c
128
slice = min_t(unsigned int, sb->s_blocksize -
fs/erofs/xattr.c
129
erofs_blkoff(sb, it->pos), len - processed);
fs/erofs/xattr.c
150
struct erofs_sb_info *sbi = EROFS_SB(it->sb);
fs/erofs/xattr.c
190
struct super_block *sb = it->sb;
fs/erofs/xattr.c
201
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/xattr.c
231
sb->s_blocksize - erofs_blkoff(sb, it->pos),
fs/erofs/xattr.c
266
ret = erofs_init_metabuf(&it->buf, it->sb, erofs_inode_in_metabox(inode));
fs/erofs/xattr.c
302
struct super_block *const sb = it->sb;
fs/erofs/xattr.c
303
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/xattr.c
307
ret = erofs_init_metabuf(&it->buf, sb,
fs/erofs/xattr.c
313
it->pos = erofs_pos(sb, sbi->xattr_blkaddr) +
fs/erofs/xattr.c
34
struct super_block *sb = inode->i_sb;
fs/erofs/xattr.c
359
it.sb = inode->i_sb;
fs/erofs/xattr.c
384
it.sb = dentry->d_sb;
fs/erofs/xattr.c
475
void erofs_xattr_prefixes_cleanup(struct super_block *sb)
fs/erofs/xattr.c
477
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/xattr.c
488
int erofs_xattr_prefixes_init(struct super_block *sb)
fs/erofs/xattr.c
490
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/xattr.c
506
(void)erofs_init_metabuf(&buf, sb, true);
fs/erofs/xattr.c
513
(void)erofs_init_metabuf(&buf, sb, false);
fs/erofs/xattr.c
516
void *ptr = erofs_read_metadata(sb, &buf, &pos, &len);
fs/erofs/xattr.c
547
erofs_xattr_prefixes_cleanup(sb);
fs/erofs/xattr.c
69
erofs_err(sb, "xattr_isize %d of nid %llu is not supported yet",
fs/erofs/xattr.c
74
erofs_err(sb, "bogus xattr ibody @ nid %llu", vi->nid);
fs/erofs/xattr.c
81
ih = erofs_read_metabuf(&buf, sb, pos, erofs_inode_in_metabox(inode));
fs/erofs/xattr.h
16
int erofs_xattr_prefixes_init(struct super_block *sb);
fs/erofs/xattr.h
17
void erofs_xattr_prefixes_cleanup(struct super_block *sb);
fs/erofs/xattr.h
20
static inline int erofs_xattr_prefixes_init(struct super_block *sb) { return 0; }
fs/erofs/xattr.h
21
static inline void erofs_xattr_prefixes_cleanup(struct super_block *sb) {}
fs/erofs/zdata.c
1108
struct super_block *sb;
fs/erofs/zdata.c
1238
erofs_folio_is_managed(EROFS_SB(be->sb), page_folio(page))) {
fs/erofs/zdata.c
1255
struct erofs_sb_info *const sbi = EROFS_SB(be->sb);
fs/erofs/zdata.c
1298
.sb = be->sb,
fs/erofs/zdata.c
1316
erofs_err(be->sb, "failed to decompress (%s) %pe @ pa %llu size %u => %u",
fs/erofs/zdata.c
1321
erofs_err(be->sb, "failed to decompress (%s) %s @ pa %llu size %u => %u",
fs/erofs/zdata.c
1398
.sb = io->sb,
fs/erofs/zdata.c
1447
struct erofs_sb_info *const sbi = EROFS_SB(io->sb);
fs/erofs/zdata.c
1590
static struct z_erofs_decompressqueue *jobqueue_init(struct super_block *sb,
fs/erofs/zdata.c
1615
q->sb = sb;
fs/erofs/zdata.c
1648
if (!erofs_folio_is_managed(EROFS_SB(q->sb), folio))
fs/erofs/zdata.c
1666
struct super_block *sb = f->inode->i_sb;
fs/erofs/zdata.c
1667
struct address_space *mc = MNGD_MAPPING(EROFS_SB(sb));
fs/erofs/zdata.c
1679
q[JQ_BYPASS] = jobqueue_init(sb, fgq + JQ_BYPASS, NULL);
fs/erofs/zdata.c
1680
q[JQ_SUBMIT] = jobqueue_init(sb, fgq + JQ_SUBMIT, force_fg);
fs/erofs/zdata.c
1704
.m_pa = round_down(pcl->pos, sb->s_blocksize),
fs/erofs/zdata.c
1706
(void)erofs_map_dev(sb, &mdev);
fs/erofs/zdata.c
1710
sb->s_blocksize);
fs/erofs/zdata.c
1716
if (erofs_is_fileio_mode(EROFS_SB(sb)))
fs/erofs/zdata.c
1718
else if (erofs_is_fscache_mode(sb))
fs/erofs/zdata.c
1736
DBG_BUGON(bvec.bv_len < sb->s_blocksize);
fs/erofs/zdata.c
1746
if (erofs_is_fileio_mode(EROFS_SB(sb)))
fs/erofs/zdata.c
1748
else if (erofs_is_fscache_mode(sb))
fs/erofs/zdata.c
1776
if (erofs_is_fileio_mode(EROFS_SB(sb)))
fs/erofs/zdata.c
1778
else if (erofs_is_fscache_mode(sb))
fs/erofs/zdata.c
400
static int z_erofs_init_pcpu_workers(struct super_block *sb)
fs/erofs/zdata.c
409
erofs_err(sb, "per-cpu workers: failed to allocate.");
fs/erofs/zdata.c
415
erofs_err(sb, "per-cpu workers: failed CPU hotplug init.");
fs/erofs/zdata.c
418
erofs_info(sb, "initialized per-cpu workers successfully.");
fs/erofs/zdata.c
436
static inline int z_erofs_init_pcpu_workers(struct super_block *sb) { return 0; }
fs/erofs/zdata.c
668
int z_erofs_init_super(struct super_block *sb)
fs/erofs/zdata.c
673
err = z_erofs_init_pcpu_workers(sb);
fs/erofs/zdata.c
677
inode = new_inode(sb);
fs/erofs/zdata.c
684
EROFS_SB(sb)->managed_cache = inode;
fs/erofs/zdata.c
685
xa_init(&EROFS_SB(sb)->managed_pslots);
fs/erofs/zdata.c
743
struct super_block *sb = fe->inode->i_sb;
fs/erofs/zdata.c
744
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/zdata.c
749
pageofs_in = erofs_blkoff(sb, map->m_pa);
fs/erofs/zdata.c
807
struct super_block *sb = fe->inode->i_sb;
fs/erofs/zdata.c
817
ret = erofs_init_metabuf(&map->buf, sb,
fs/erofs/zdata.c
823
erofs_err(sb, "failed to read inline data %pe @ pa %llu of nid %llu",
fs/erofs/zdata.c
831
pcl = xa_load(&EROFS_SB(sb)->managed_pslots, map->m_pa);
fs/erofs/zdata.c
94
struct super_block *sb;
fs/erofs/zdata.c
983
static int z_erofs_read_fragment(struct super_block *sb, struct folio *folio,
fs/erofs/zdata.c
986
struct inode *packed_inode = EROFS_SB(sb)->packed_inode;
fs/erofs/zdata.c
996
cnt = min(end - cur, sb->s_blocksize - erofs_blkoff(sb, pos));
fs/erofs/zmap.c
267
struct super_block *sb = m->inode->i_sb;
fs/erofs/zmap.c
289
erofs_err(sb, "bogus lookback distance %u @ lcn %lu of nid %llu",
fs/erofs/zmap.c
299
struct super_block *sb = inode->i_sb;
fs/erofs/zmap.c
334
erofs_err(sb, "bogus CBLKCNT @ lcn %lu of nid %llu", lcn, vi->nid);
fs/erofs/zmap.c
346
m->map->m_plen = erofs_pos(sb, m->compressedblks);
fs/erofs/zmap.c
391
struct super_block *sb = inode->i_sb;
fs/erofs/zmap.c
460
if (erofs_blkoff(sb, map->m_pa) + map->m_plen > sb->s_blocksize) {
fs/erofs/zmap.c
461
erofs_err(sb, "ztailpacking inline data across blocks @ nid %llu",
fs/erofs/zmap.c
469
map->m_pa = erofs_pos(sb, m.pblk);
fs/erofs/zmap.c
511
struct super_block *sb = inode->i_sb;
fs/erofs/zmap.c
516
unsigned int bmask = sb->s_blocksize - 1;
fs/erofs/zmap.c
527
ext = erofs_read_metabuf(&map->buf, sb, pos, in_mbox);
fs/erofs/zmap.c
540
ext = erofs_read_metabuf(&map->buf, sb, pos, in_mbox);
fs/erofs/zmap.c
559
ext = erofs_read_metabuf(&map->buf, sb,
fs/erofs/zmap.c
620
struct super_block *const sb = inode->i_sb;
fs/erofs/zmap.c
641
h = erofs_read_metabuf(&map->buf, sb, pos, erofs_inode_in_metabox(inode));
fs/erofs/zmap.c
658
vi->z_lclusterbits = sb->s_blocksize_bits + (h->h_clusterbits & 15);
fs/erofs/zmap.c
673
if (!erofs_sb_has_big_pcluster(EROFS_SB(sb)) &&
fs/erofs/zmap.c
676
erofs_err(sb, "per-inode big pcluster without sb feature for nid %llu",
fs/erofs/zmap.c
684
erofs_err(sb, "big pcluster head1/2 of compact indexes should be consistent for nid %llu",
fs/erofs/zutil.c
216
void erofs_shrinker_register(struct super_block *sb)
fs/erofs/zutil.c
218
struct erofs_sb_info *sbi = EROFS_SB(sb);
fs/erofs/zutil.c
227
void erofs_shrinker_unregister(struct super_block *sb)
fs/erofs/zutil.c
229
struct erofs_sb_info *const sbi = EROFS_SB(sb);
fs/exfat/balloc.c
104
max_ra_count = min(sb->s_bdi->ra_pages, sb->s_bdi->io_pages) <<
fs/exfat/balloc.c
105
(PAGE_SHIFT - sb->s_blocksize_bits);
fs/exfat/balloc.c
111
sb_breadahead(sb, sector + j);
fs/exfat/balloc.c
115
sbi->vol_amap[i] = sb_bread(sb, sector + i);
fs/exfat/balloc.c
120
if (exfat_test_bitmap_range(sb, sbi->map_clu,
fs/exfat/balloc.c
137
int exfat_load_bitmap(struct super_block *sb)
fs/exfat/balloc.c
141
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
149
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/balloc.c
158
err = exfat_allocate_bitmap(sb, ep);
fs/exfat/balloc.c
168
if (exfat_get_next_cluster(sb, &clu.dir))
fs/exfat/balloc.c
185
int exfat_set_bitmap(struct super_block *sb, unsigned int clu, bool sync)
fs/exfat/balloc.c
189
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
195
i = BITMAP_OFFSET_SECTOR_INDEX(sb, ent_idx);
fs/exfat/balloc.c
196
b = BITMAP_OFFSET_BIT_IN_SECTOR(sb, ent_idx);
fs/exfat/balloc.c
203
int exfat_clear_bitmap(struct super_block *sb, unsigned int clu, bool sync)
fs/exfat/balloc.c
207
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
213
i = BITMAP_OFFSET_SECTOR_INDEX(sb, ent_idx);
fs/exfat/balloc.c
214
b = BITMAP_OFFSET_BIT_IN_SECTOR(sb, ent_idx);
fs/exfat/balloc.c
226
bool exfat_test_bitmap(struct super_block *sb, unsigned int clu)
fs/exfat/balloc.c
230
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
239
i = BITMAP_OFFSET_SECTOR_INDEX(sb, ent_idx);
fs/exfat/balloc.c
240
b = BITMAP_OFFSET_BIT_IN_SECTOR(sb, ent_idx);
fs/exfat/balloc.c
252
unsigned int exfat_find_free_bitmap(struct super_block *sb, unsigned int clu)
fs/exfat/balloc.c
257
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
265
map_i = BITMAP_OFFSET_SECTOR_INDEX(sb, ent_idx);
fs/exfat/balloc.c
266
map_b = BITMAP_OFFSET_BYTE_IN_SECTOR(sb, ent_idx);
fs/exfat/balloc.c
284
if (map_b >= sb->s_blocksize ||
fs/exfat/balloc.c
297
int exfat_count_used_clusters(struct super_block *sb, unsigned int *ret_count)
fs/exfat/balloc.c
299
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
30
static bool exfat_test_bitmap_range(struct super_block *sb, unsigned int clu,
fs/exfat/balloc.c
311
if (map_b >= (unsigned int)sb->s_blocksize) {
fs/exfat/balloc.c
33
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
331
struct super_block *sb = inode->i_sb;
fs/exfat/balloc.c
332
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
348
trim_begin = trim_end = exfat_find_free_bitmap(sb, clu_start);
fs/exfat/balloc.c
352
next_free_clu = exfat_find_free_bitmap(sb, trim_end + 1);
fs/exfat/balloc.c
364
err = sb_issue_discard(sb,
fs/exfat/balloc.c
385
next_free_clu = exfat_find_free_bitmap(sb, next_free_clu + 1);
fs/exfat/balloc.c
392
err = sb_issue_discard(sb, exfat_cluster_to_sector(sbi, trim_begin),
fs/exfat/balloc.c
46
i = BITMAP_OFFSET_SECTOR_INDEX(sb, ent_idx);
fs/exfat/balloc.c
47
b = BITMAP_OFFSET_BIT_IN_SECTOR(sb, ent_idx);
fs/exfat/balloc.c
73
static int exfat_allocate_bitmap(struct super_block *sb,
fs/exfat/balloc.c
76
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/balloc.c
88
exfat_err(sb, "bogus allocation bitmap size(need : %u, cur : %lld)",
fs/exfat/balloc.c
98
(sb->s_blocksize_bits)) + 1;
fs/exfat/cache.c
265
struct super_block *sb = inode->i_sb;
fs/exfat/cache.c
273
exfat_fs_error(sb,
fs/exfat/cache.c
312
if (exfat_ent_get(sb, *dclus, &content, &bh))
fs/exfat/cache.c
334
if (exfat_ent_get(sb, clu, &content, &bh))
fs/exfat/dir.c
1004
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
1037
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/dir.c
108
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/dir.c
1112
if (exfat_uniname_ncmp(sb, uniname,
fs/exfat/dir.c
1142
if (exfat_get_next_cluster(sb, &clu.dir))
fs/exfat/dir.c
1189
ret = exfat_get_next_cluster(sb, &clu.dir);
fs/exfat/dir.c
119
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/dir.c
1205
int exfat_count_dir_entries(struct super_block *sb, struct exfat_chain *p_dir)
fs/exfat/dir.c
1213
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
1222
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/dir.c
1241
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/dir.c
1245
exfat_fs_error(sb, "FAT or bitmap is corrupted");
fs/exfat/dir.c
1255
static int exfat_get_volume_label_dentry(struct super_block *sb,
fs/exfat/dir.c
1261
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
1263
struct exfat_inode_info *ei = EXFAT_I(sb->s_root->d_inode);
fs/exfat/dir.c
1273
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/dir.c
1297
es->sb = sb;
fs/exfat/dir.c
1301
es->start_off = EXFAT_DEN_TO_B(i) % sb->s_blocksize;
fs/exfat/dir.c
1306
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/dir.c
1322
int exfat_read_volume_label(struct super_block *sb, struct exfat_uni_name *label_out)
fs/exfat/dir.c
1325
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
1332
ret = exfat_get_volume_label_dentry(sb, &es);
fs/exfat/dir.c
1361
int exfat_write_volume_label(struct super_block *sb,
fs/exfat/dir.c
1365
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
1366
struct inode *root_inode = sb->s_root->d_inode;
fs/exfat/dir.c
1376
ret = exfat_get_volume_label_dentry(sb, &es);
fs/exfat/dir.c
138
err = exfat_get_uniname_from_ext_entry(sb, &clu, i,
fs/exfat/dir.c
144
exfat_utf16_to_nls(sb, &uni_name,
fs/exfat/dir.c
149
ep = exfat_get_dentry(sb, &clu, i + 1, &bh);
fs/exfat/dir.c
169
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/dir.c
212
struct super_block *sb = inode->i_sb;
fs/exfat/dir.c
239
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/dir.c
251
cpos += 1 << (sb->s_blocksize_bits);
fs/exfat/dir.c
252
cpos &= ~(sb->s_blocksize - 1);
fs/exfat/dir.c
263
tmp = exfat_iget(sb, i_pos);
fs/exfat/dir.c
268
inum = iunique(sb, EXFAT_ROOT_INO);
fs/exfat/dir.c
271
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/dir.c
282
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/dir.c
33
static int exfat_get_uniname_from_ext_entry(struct super_block *sb,
fs/exfat/dir.c
40
err = exfat_get_dentry_set(&es, sb, p_dir, entry, ES_ALL_ENTRIES);
fs/exfat/dir.c
440
struct super_block *sb = es->sb;
fs/exfat/dir.c
441
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
470
struct super_block *sb = inode->i_sb;
fs/exfat/dir.c
481
EXFAT_B_TO_CLU_ROUND_UP(size, EXFAT_SB(sb)),
fs/exfat/dir.c
564
static int exfat_walk_fat_chain(struct super_block *sb,
fs/exfat/dir.c
568
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
579
if (exfat_get_next_cluster(sb, &cur_clu))
fs/exfat/dir.c
582
exfat_fs_error(sb,
fs/exfat/dir.c
596
static int exfat_find_location(struct super_block *sb, struct exfat_chain *p_dir,
fs/exfat/dir.c
601
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
605
ret = exfat_walk_fat_chain(sb, p_dir, off, &clu);
fs/exfat/dir.c
609
if (!exfat_test_bitmap(sb, clu)) {
fs/exfat/dir.c
610
exfat_err(sb, "failed to test cluster bit(%u)", clu);
fs/exfat/dir.c
618
*offset = EXFAT_BLK_OFFSET(off, sb);
fs/exfat/dir.c
621
*sector = EXFAT_B_TO_BLK(off, sb);
fs/exfat/dir.c
627
static int exfat_dir_readahead(struct super_block *sb, sector_t sec)
fs/exfat/dir.c
629
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
631
unsigned int max_ra_count = EXFAT_MAX_RA_SIZE >> sb->s_blocksize_bits;
fs/exfat/dir.c
632
unsigned int page_ra_count = PAGE_SIZE >> sb->s_blocksize_bits;
fs/exfat/dir.c
641
exfat_err(sb, "requested sector is invalid(sect:%llu, root:%llu)",
fs/exfat/dir.c
650
bh = sb_find_get_block(sb, sec);
fs/exfat/dir.c
655
sb_breadahead(sb, (sector_t)(sec + i));
fs/exfat/dir.c
661
struct exfat_dentry *exfat_get_dentry(struct super_block *sb,
fs/exfat/dir.c
669
exfat_err(sb, "abnormal access to deleted dentry");
fs/exfat/dir.c
673
if (exfat_find_location(sb, p_dir, entry, &sec, &off))
fs/exfat/dir.c
678
exfat_dir_readahead(sb, sec);
fs/exfat/dir.c
680
*bh = sb_bread(sb, sec);
fs/exfat/dir.c
734
struct buffer_head *bh = es->bh[EXFAT_B_TO_BLK(off, es->sb)];
fs/exfat/dir.c
735
char *p = bh->b_data + EXFAT_BLK_OFFSET(off, es->sb);
fs/exfat/dir.c
76
struct super_block *sb = inode->i_sb;
fs/exfat/dir.c
760
struct super_block *sb, struct exfat_chain *p_dir, int entry,
fs/exfat/dir.c
766
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
77
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/dir.c
770
exfat_err(sb, "access to deleted dentry");
fs/exfat/dir.c
774
ret = exfat_find_location(sb, p_dir, entry, &sec, &off);
fs/exfat/dir.c
779
es->sb = sb;
fs/exfat/dir.c
784
bh = sb_bread(sb, sec);
fs/exfat/dir.c
803
num_bh = EXFAT_B_TO_BLK_ROUND_UP(off + num_entries * DENTRY_SIZE, sb);
fs/exfat/dir.c
820
else if (exfat_get_next_cluster(sb, &clu))
fs/exfat/dir.c
827
bh = sb_bread(sb, sec);
fs/exfat/dir.c
841
struct super_block *sb, struct exfat_chain *p_dir,
fs/exfat/dir.c
848
ret = __exfat_get_dentry_set(es, sb, p_dir, entry, num_entries);
fs/exfat/dir.c
894
bh = es->bh[EXFAT_B_TO_BLK(off, es->sb)];
fs/exfat/dir.c
896
exfat_fs_error(es->sb,
fs/exfat/dir.c
903
es->num_entries = EXFAT_B_TO_DEN(EXFAT_BLK_TO_B(es->num_bh, es->sb) - es->start_off);
fs/exfat/dir.c
928
struct super_block *sb, struct exfat_chain *p_dir,
fs/exfat/dir.c
933
ret = __exfat_get_dentry_set(es, sb, p_dir, entry, num_entries);
fs/exfat/dir.c
992
int exfat_find_dir_entry(struct super_block *sb, struct exfat_inode_info *ei,
fs/exfat/exfat_fs.h
119
#define FAT_ENT_OFFSET_SECTOR(sb, loc) (EXFAT_SB(sb)->FAT1_start_sector + \
fs/exfat/exfat_fs.h
120
(((u64)loc << FAT_ENT_SIZE_BITS) >> sb->s_blocksize_bits))
fs/exfat/exfat_fs.h
121
#define FAT_ENT_OFFSET_BYTE_IN_SECTOR(sb, loc) \
fs/exfat/exfat_fs.h
122
((loc << FAT_ENT_SIZE_BITS) & (sb->s_blocksize - 1))
fs/exfat/exfat_fs.h
129
#define BITS_PER_SECTOR(sb) ((sb)->s_blocksize * BITS_PER_BYTE)
fs/exfat/exfat_fs.h
130
#define BITS_PER_SECTOR_MASK(sb) (BITS_PER_SECTOR(sb) - 1)
fs/exfat/exfat_fs.h
131
#define BITMAP_OFFSET_SECTOR_INDEX(sb, ent) \
fs/exfat/exfat_fs.h
132
((ent / BITS_PER_BYTE) >> (sb)->s_blocksize_bits)
fs/exfat/exfat_fs.h
133
#define BITMAP_OFFSET_BIT_IN_SECTOR(sb, ent) (ent & BITS_PER_SECTOR_MASK(sb))
fs/exfat/exfat_fs.h
134
#define BITMAP_OFFSET_BYTE_IN_SECTOR(sb, ent) \
fs/exfat/exfat_fs.h
135
((ent / BITS_PER_BYTE) & ((sb)->s_blocksize - 1))
fs/exfat/exfat_fs.h
192
struct super_block *sb;
fs/exfat/exfat_fs.h
331
static inline struct exfat_sb_info *EXFAT_SB(struct super_block *sb)
fs/exfat/exfat_fs.h
333
return sb->s_fs_info;
fs/exfat/exfat_fs.h
341
static inline int exfat_forced_shutdown(struct super_block *sb)
fs/exfat/exfat_fs.h
343
return test_bit(EXFAT_FLAGS_SHUTDOWN, &EXFAT_SB(sb)->s_exfat_flags);
fs/exfat/exfat_fs.h
431
int exfat_set_volume_dirty(struct super_block *sb);
fs/exfat/exfat_fs.h
432
int exfat_clear_volume_dirty(struct super_block *sb);
fs/exfat/exfat_fs.h
435
#define exfat_get_next_cluster(sb, pclu) exfat_ent_get(sb, *(pclu), pclu, NULL)
fs/exfat/exfat_fs.h
440
int exfat_ent_get(struct super_block *sb, unsigned int loc,
fs/exfat/exfat_fs.h
442
int exfat_ent_set(struct super_block *sb, unsigned int loc,
fs/exfat/exfat_fs.h
444
int exfat_chain_cont_cluster(struct super_block *sb, unsigned int chain,
fs/exfat/exfat_fs.h
447
int exfat_find_last_cluster(struct super_block *sb, struct exfat_chain *p_chain,
fs/exfat/exfat_fs.h
449
int exfat_count_num_clusters(struct super_block *sb,
fs/exfat/exfat_fs.h
453
int exfat_load_bitmap(struct super_block *sb);
fs/exfat/exfat_fs.h
455
int exfat_set_bitmap(struct super_block *sb, unsigned int clu, bool sync);
fs/exfat/exfat_fs.h
456
int exfat_clear_bitmap(struct super_block *sb, unsigned int clu, bool sync);
fs/exfat/exfat_fs.h
457
bool exfat_test_bitmap(struct super_block *sb, unsigned int clu);
fs/exfat/exfat_fs.h
458
unsigned int exfat_find_free_bitmap(struct super_block *sb, unsigned int clu);
fs/exfat/exfat_fs.h
459
int exfat_count_used_clusters(struct super_block *sb, unsigned int *ret_count);
fs/exfat/exfat_fs.h
475
int exfat_force_shutdown(struct super_block *sb, u32 flags);
fs/exfat/exfat_fs.h
504
int exfat_find_dir_entry(struct super_block *sb, struct exfat_inode_info *ei,
fs/exfat/exfat_fs.h
508
struct exfat_dentry *exfat_get_dentry(struct super_block *sb,
fs/exfat/exfat_fs.h
513
struct super_block *sb, struct exfat_chain *p_dir, int entry,
fs/exfat/exfat_fs.h
515
#define exfat_get_dentry_set_by_ei(es, sb, ei) \
fs/exfat/exfat_fs.h
516
exfat_get_dentry_set(es, sb, &(ei)->dir, (ei)->entry, ES_ALL_ENTRIES)
fs/exfat/exfat_fs.h
518
struct super_block *sb, struct exfat_chain *p_dir, int entry,
fs/exfat/exfat_fs.h
521
int exfat_count_dir_entries(struct super_block *sb, struct exfat_chain *p_dir);
fs/exfat/exfat_fs.h
522
int exfat_read_volume_label(struct super_block *sb,
fs/exfat/exfat_fs.h
524
int exfat_write_volume_label(struct super_block *sb,
fs/exfat/exfat_fs.h
530
struct inode *exfat_build_inode(struct super_block *sb,
fs/exfat/exfat_fs.h
534
struct inode *exfat_iget(struct super_block *sb, loff_t i_pos);
fs/exfat/exfat_fs.h
541
unsigned short exfat_toupper(struct super_block *sb, unsigned short a);
fs/exfat/exfat_fs.h
542
int exfat_uniname_ncmp(struct super_block *sb, unsigned short *a,
fs/exfat/exfat_fs.h
544
int exfat_utf16_to_nls(struct super_block *sb,
fs/exfat/exfat_fs.h
547
int exfat_nls_to_utf16(struct super_block *sb,
fs/exfat/exfat_fs.h
550
int exfat_create_upcase_table(struct super_block *sb);
fs/exfat/exfat_fs.h
554
void __exfat_fs_error(struct super_block *sb, int report, const char *fmt, ...)
fs/exfat/exfat_fs.h
556
#define exfat_fs_error(sb, fmt, args...) \
fs/exfat/exfat_fs.h
557
__exfat_fs_error(sb, 1, fmt, ## args)
fs/exfat/exfat_fs.h
558
#define exfat_fs_error_ratelimit(sb, fmt, args...) \
fs/exfat/exfat_fs.h
559
__exfat_fs_error(sb, __ratelimit(&EXFAT_SB(sb)->ratelimit), \
fs/exfat/exfat_fs.h
563
#define exfat_err(sb, fmt, ...) \
fs/exfat/exfat_fs.h
564
pr_err("exFAT-fs (%s): " fmt "\n", (sb)->s_id, ##__VA_ARGS__)
fs/exfat/exfat_fs.h
565
#define exfat_warn(sb, fmt, ...) \
fs/exfat/exfat_fs.h
566
pr_warn("exFAT-fs (%s): " fmt "\n", (sb)->s_id, ##__VA_ARGS__)
fs/exfat/exfat_fs.h
567
#define exfat_info(sb, fmt, ...) \
fs/exfat/exfat_fs.h
568
pr_info("exFAT-fs (%s): " fmt "\n", (sb)->s_id, ##__VA_ARGS__)
fs/exfat/exfat_fs.h
569
#define exfat_debug(sb, fmt, ...) \
fs/exfat/exfat_fs.h
570
pr_debug("exFAT-fs (%s): " fmt "\n", (sb)->s_id, ##__VA_ARGS__)
fs/exfat/exfat_fs.h
94
#define EXFAT_BLK_TO_B(b, sb) ((b) << (sb)->s_blocksize_bits)
fs/exfat/exfat_fs.h
95
#define EXFAT_B_TO_BLK(b, sb) ((b) >> (sb)->s_blocksize_bits)
fs/exfat/exfat_fs.h
96
#define EXFAT_B_TO_BLK_ROUND_UP(b, sb) \
fs/exfat/exfat_fs.h
97
(((b - 1) >> (sb)->s_blocksize_bits) + 1)
fs/exfat/exfat_fs.h
98
#define EXFAT_BLK_OFFSET(off, sb) ((off) & ((sb)->s_blocksize - 1))
fs/exfat/fatent.c
100
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
106
if (unlikely(__exfat_ent_get(sb, loc, content, last))) {
fs/exfat/fatent.c
107
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
114
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
121
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
128
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
14
static int exfat_mirror_bh(struct super_block *sb, sector_t sec,
fs/exfat/fatent.c
145
int exfat_chain_cont_cluster(struct super_block *sb, unsigned int chain,
fs/exfat/fatent.c
152
if (exfat_ent_set(sb, chain, chain + 1))
fs/exfat/fatent.c
158
if (exfat_ent_set(sb, chain, EXFAT_EOF_CLUSTER))
fs/exfat/fatent.c
163
static inline void exfat_discard_cluster(struct super_block *sb,
fs/exfat/fatent.c
167
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
169
ret = sb_issue_discard(sb, exfat_cluster_to_sector(sbi, clu),
fs/exfat/fatent.c
172
exfat_err(sb, "discard not supported by device, disabling");
fs/exfat/fatent.c
18
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
180
struct super_block *sb = inode->i_sb;
fs/exfat/fatent.c
181
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
198
exfat_err(sb, "invalid start cluster (%u)", p_chain->dir);
fs/exfat/fatent.c
205
BITMAP_OFFSET_SECTOR_INDEX(sb, CLUSTER_TO_BITMAP_ENT(clu));
fs/exfat/fatent.c
216
BITMAP_OFFSET_SECTOR_INDEX(sb, CLUSTER_TO_BITMAP_ENT(clu+1));
fs/exfat/fatent.c
224
err = exfat_clear_bitmap(sb, clu, (sync && IS_DIRSYNC(inode)));
fs/exfat/fatent.c
232
exfat_discard_cluster(sb, p_chain->dir, p_chain->size);
fs/exfat/fatent.c
239
int err = exfat_get_next_cluster(sb, &n_clu);
fs/exfat/fatent.c
24
c_bh = sb_getblk(sb, sec2);
fs/exfat/fatent.c
245
BITMAP_OFFSET_SECTOR_INDEX(sb, CLUSTER_TO_BITMAP_ENT(n_clu));
fs/exfat/fatent.c
252
if (exfat_clear_bitmap(sb, clu, (sync && IS_DIRSYNC(inode))))
fs/exfat/fatent.c
259
exfat_discard_cluster(sb, clu - nr_clu + 1, nr_clu);
fs/exfat/fatent.c
27
memcpy(c_bh->b_data, bh->b_data, sb->s_blocksize);
fs/exfat/fatent.c
275
exfat_count_used_clusters(sb, &sbi->used_clusters);
fs/exfat/fatent.c
297
int exfat_find_last_cluster(struct super_block *sb, struct exfat_chain *p_chain,
fs/exfat/fatent.c
30
if (sb->s_flags & SB_SYNCHRONOUS)
fs/exfat/fatent.c
313
if (exfat_ent_get(sb, clu, &next, &bh))
fs/exfat/fatent.c
319
exfat_fs_error(sb,
fs/exfat/fatent.c
331
struct super_block *sb = dir->i_sb;
fs/exfat/fatent.c
332
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
340
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
349
bh = sb_getblk(sb, i);
fs/exfat/fatent.c
353
memset(bh->b_data, 0, sb->s_blocksize);
fs/exfat/fatent.c
360
return sync_blockdev_range(sb->s_bdev,
fs/exfat/fatent.c
361
EXFAT_BLK_TO_B(blknr, sb),
fs/exfat/fatent.c
362
EXFAT_BLK_TO_B(last_blknr, sb) - 1);
fs/exfat/fatent.c
373
struct super_block *sb = inode->i_sb;
fs/exfat/fatent.c
374
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
379
exfat_fs_error_ratelimit(sb,
fs/exfat/fatent.c
38
static int __exfat_ent_get(struct super_block *sb, unsigned int loc,
fs/exfat/fatent.c
394
exfat_err(sb, "sbi->clu_srch_ptr is invalid (%u)",
fs/exfat/fatent.c
399
hint_clu = exfat_find_free_bitmap(sb, sbi->clu_srch_ptr);
fs/exfat/fatent.c
409
exfat_err(sb, "hint_cluster is invalid (%u), rewind to the first cluster",
fs/exfat/fatent.c
417
while ((new_clu = exfat_find_free_bitmap(sb, hint_clu)) !=
fs/exfat/fatent.c
421
if (exfat_chain_cont_cluster(sb, p_chain->dir,
fs/exfat/fatent.c
430
if (exfat_set_bitmap(sb, new_clu, sync_bmap)) {
fs/exfat/fatent.c
437
if (exfat_ent_set(sb, new_clu, EXFAT_EOF_CLUSTER)) {
fs/exfat/fatent.c
446
if (exfat_ent_set(sb, last_clu, new_clu)) {
fs/exfat/fatent.c
45
sec = FAT_ENT_OFFSET_SECTOR(sb, loc);
fs/exfat/fatent.c
46
off = FAT_ENT_OFFSET_BYTE_IN_SECTOR(sb, loc);
fs/exfat/fatent.c
468
if (exfat_chain_cont_cluster(sb, p_chain->dir,
fs/exfat/fatent.c
484
int exfat_count_num_clusters(struct super_block *sb,
fs/exfat/fatent.c
489
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/fatent.c
50
bh = sb_bread(sb, sec);
fs/exfat/fatent.c
506
if (exfat_ent_get(sb, clu, &clu, &bh))
fs/exfat/fatent.c
520
exfat_fs_error(sb, "The cluster chain has a loop");
fs/exfat/fatent.c
68
int exfat_ent_set(struct super_block *sb, unsigned int loc,
fs/exfat/fatent.c
76
sec = FAT_ENT_OFFSET_SECTOR(sb, loc);
fs/exfat/fatent.c
77
off = FAT_ENT_OFFSET_BYTE_IN_SECTOR(sb, loc);
fs/exfat/fatent.c
79
bh = sb_bread(sb, sec);
fs/exfat/fatent.c
85
exfat_update_bh(bh, sb->s_flags & SB_SYNCHRONOUS);
fs/exfat/fatent.c
86
exfat_mirror_bh(sb, sec, bh);
fs/exfat/fatent.c
94
int exfat_ent_get(struct super_block *sb, unsigned int loc,
fs/exfat/fatent.c
97
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/file.c
152
struct super_block *sb = inode->i_sb;
fs/exfat/file.c
153
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/file.c
160
exfat_set_volume_dirty(sb);
fs/exfat/file.c
185
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/file.c
220
if (exfat_ent_set(sb, last_clu, EXFAT_EOF_CLUSTER))
fs/exfat/file.c
246
struct super_block *sb = inode->i_sb;
fs/exfat/file.c
247
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/file.c
25
struct super_block *sb = inode->i_sb;
fs/exfat/file.c
256
exfat_fs_error(sb, "tried to truncate zeroed cluster.");
fs/exfat/file.c
26
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/file.c
43
ret = exfat_find_last_cluster(sb, &clu, &last_clu);
fs/exfat/file.c
479
static int exfat_ioctl_shutdown(struct super_block *sb, unsigned long arg)
fs/exfat/file.c
489
return exfat_force_shutdown(sb, flags);
fs/exfat/file.c
492
static int exfat_ioctl_get_volume_label(struct super_block *sb, unsigned long arg)
fs/exfat/file.c
498
ret = exfat_read_volume_label(sb, &uniname);
fs/exfat/file.c
502
ret = exfat_utf16_to_nls(sb, &uniname, label, uniname.name_len);
fs/exfat/file.c
512
static int exfat_ioctl_set_volume_label(struct super_block *sb,
fs/exfat/file.c
528
ret = exfat_nls_to_utf16(sb, label, label_len,
fs/exfat/file.c
538
return exfat_write_volume_label(sb, &uniname);
fs/exfat/file.c
64
if (exfat_chain_cont_cluster(sb, ei->start_clu, num_clusters))
fs/exfat/file.c
68
if (exfat_ent_set(sb, last_clu, clu.dir))
fs/exfat/inode.c
132
struct super_block *sb = inode->i_sb;
fs/exfat/inode.c
133
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/inode.c
169
exfat_set_volume_dirty(sb);
fs/exfat/inode.c
179
exfat_fs_error(sb, "broken FAT chain.");
fs/exfat/inode.c
190
exfat_fs_error(sb,
fs/exfat/inode.c
207
exfat_chain_cont_cluster(sb, ei->start_clu,
fs/exfat/inode.c
212
if (exfat_ent_set(sb, last_clu, new_clu.dir))
fs/exfat/inode.c
232
if (exfat_get_next_cluster(sb, clu))
fs/exfat/inode.c
25
struct super_block *sb = inode->i_sb;
fs/exfat/inode.c
251
struct super_block *sb = inode->i_sb;
fs/exfat/inode.c
252
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/inode.c
26
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/inode.c
264
last_block = EXFAT_B_TO_BLK_ROUND_UP(i_size, sb);
fs/exfat/inode.c
274
exfat_fs_error_ratelimit(sb,
fs/exfat/inode.c
290
map_bh(bh_result, sb, phys);
fs/exfat/inode.c
306
valid_blks = EXFAT_B_TO_BLK_ROUND_UP(ei->valid_size, sb);
fs/exfat/inode.c
308
valid_blks = EXFAT_B_TO_BLK(ei->valid_size, sb);
fs/exfat/inode.c
323
ei->valid_size = EXFAT_BLK_TO_B(iblock + max_blocks, sb);
fs/exfat/inode.c
334
(ei->valid_size & (sb->s_blocksize - 1))) {
fs/exfat/inode.c
353
pos = EXFAT_BLK_TO_B(iblock, sb);
fs/exfat/inode.c
360
exfat_fs_error_ratelimit(sb,
fs/exfat/inode.c
384
bh_result->b_size = EXFAT_BLK_TO_B(max_blocks, sb);
fs/exfat/inode.c
43
exfat_set_volume_dirty(sb);
fs/exfat/inode.c
46
if (exfat_get_dentry_set_by_ei(&es, sb, ei))
fs/exfat/inode.c
583
struct inode *exfat_iget(struct super_block *sb, loff_t i_pos)
fs/exfat/inode.c
585
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/inode.c
592
WARN_ON(info->vfs_inode.i_sb != sb);
fs/exfat/inode.c
659
struct inode *exfat_build_inode(struct super_block *sb,
fs/exfat/inode.c
665
inode = exfat_iget(sb, i_pos);
fs/exfat/inode.c
668
inode = new_inode(sb);
fs/exfat/inode.c
673
inode->i_ino = iunique(sb, EXFAT_ROOT_INO);
fs/exfat/misc.c
26
void __exfat_fs_error(struct super_block *sb, int report, const char *fmt, ...)
fs/exfat/misc.c
28
struct exfat_mount_options *opts = &EXFAT_SB(sb)->options;
fs/exfat/misc.c
36
exfat_err(sb, "error, %pV", &vaf);
fs/exfat/misc.c
42
sb->s_id);
fs/exfat/misc.c
43
} else if (opts->errors == EXFAT_ERRORS_RO && !sb_rdonly(sb)) {
fs/exfat/misc.c
44
sb->s_flags |= SB_RDONLY;
fs/exfat/misc.c
45
exfat_err(sb, "Filesystem has been set read-only");
fs/exfat/namei.c
1016
struct super_block *sb = parent_inode->i_sb;
fs/exfat/namei.c
1020
if (unlikely(exfat_forced_shutdown(sb)))
fs/exfat/namei.c
1027
ret = exfat_get_dentry_set_by_ei(&old_es, sb, ei);
fs/exfat/namei.c
104
struct super_block *sb = dentry->d_sb;
fs/exfat/namei.c
105
struct nls_table *t = EXFAT_SB(sb)->nls_io;
fs/exfat/namei.c
107
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
109
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
1142
struct super_block *sb = old_parent_inode->i_sb;
fs/exfat/namei.c
1143
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
1153
exfat_err(sb, "abnormal access to deleted source dentry");
fs/exfat/namei.c
1163
exfat_err(sb, "abnormal access to deleted target dentry");
fs/exfat/namei.c
1177
ret = exfat_check_dir_empty(sb, &new_clu);
fs/exfat/namei.c
1188
exfat_set_volume_dirty(sb);
fs/exfat/namei.c
1199
ret = exfat_get_dentry_set_by_ei(&es, sb, new_ei);
fs/exfat/namei.c
123
if (exfat_toupper(sb, c1) != exfat_toupper(sb, c2))
fs/exfat/namei.c
1248
struct super_block *sb = old_dir->i_sb;
fs/exfat/namei.c
1261
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
1306
exfat_warn(sb, "abnormal access to an inode dropped");
fs/exfat/namei.c
1313
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
138
struct super_block *sb = dentry->d_sb;
fs/exfat/namei.c
141
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
154
hash = partial_name_hash(u <= 0xFFFF ? exfat_toupper(sb, u) : u,
fs/exfat/namei.c
165
struct super_block *sb = dentry->d_sb;
fs/exfat/namei.c
167
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
169
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
185
if (exfat_toupper(sb, u_a) != exfat_toupper(sb, u_b))
fs/exfat/namei.c
203
static int exfat_search_empty_slot(struct super_block *sb,
fs/exfat/namei.c
210
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
239
ret = exfat_get_empty_dentry_set(es, sb, &clu, i, num_entries);
fs/exfat/namei.c
255
if (exfat_get_next_cluster(sb, &clu.dir))
fs/exfat/namei.c
311
struct super_block *sb = inode->i_sb;
fs/exfat/namei.c
312
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
326
while ((dentry = exfat_search_empty_slot(sb, &hint_femp, p_dir,
fs/exfat/namei.c
339
if (exfat_find_last_cluster(sb, p_dir, &last_clu))
fs/exfat/namei.c
368
exfat_chain_cont_cluster(sb, p_dir->dir, p_dir->size);
fs/exfat/namei.c
374
if (exfat_ent_set(sb, last_clu, clu.dir))
fs/exfat/namei.c
408
struct super_block *sb = inode->i_sb;
fs/exfat/namei.c
416
if (EXFAT_SB(sb)->options.keep_last_dots) {
fs/exfat/namei.c
439
namelen = exfat_nls_to_utf16(sb, path, namelen, p_uniname,
fs/exfat/namei.c
471
struct super_block *sb = inode->i_sb;
fs/exfat/namei.c
472
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
548
struct super_block *sb = dir->i_sb;
fs/exfat/namei.c
555
if (unlikely(exfat_forced_shutdown(sb)))
fs/exfat/namei.c
558
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
559
exfat_set_volume_dirty(sb);
fs/exfat/namei.c
572
inode = exfat_build_inode(sb, &info, i_pos);
fs/exfat/namei.c
585
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
596
struct super_block *sb = dir->i_sb;
fs/exfat/namei.c
597
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
624
dentry = exfat_find_dir_entry(sb, ei, &cdir, &uni_name, &hint_opt);
fs/exfat/namei.c
638
if (exfat_get_dentry_set(&es, sb, &cdir, dentry, ES_2_ENTRIES))
fs/exfat/namei.c
650
exfat_warn(sb, "start_clu is invalid cluster(0x%x)",
fs/exfat/namei.c
657
exfat_warn(sb, "valid_size(%lld) is greater than size(%lld)",
fs/exfat/namei.c
686
exfat_fs_error(sb, "data valid size is invalid(%lld)", info->valid_size);
fs/exfat/namei.c
691
exfat_fs_error(sb, "data size is invalid(%lld)", info->size);
fs/exfat/namei.c
696
exfat_fs_error(sb,
fs/exfat/namei.c
705
count = exfat_count_dir_entries(sb, &cdir);
fs/exfat/namei.c
722
struct super_block *sb = dir->i_sb;
fs/exfat/namei.c
730
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
741
inode = exfat_build_inode(sb, &info, i_pos);
fs/exfat/namei.c
765
exfat_info(sb, "rehashed a dentry(%p) in read lookup",
fs/exfat/namei.c
780
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
785
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
791
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
798
struct super_block *sb = dir->i_sb;
fs/exfat/namei.c
804
if (unlikely(exfat_forced_shutdown(sb)))
fs/exfat/namei.c
807
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
809
exfat_err(sb, "abnormal access to deleted dentry");
fs/exfat/namei.c
81
struct super_block *sb = dentry->d_sb;
fs/exfat/namei.c
814
err = exfat_get_dentry_set_by_ei(&es, sb, ei);
fs/exfat/namei.c
82
struct nls_table *t = EXFAT_SB(sb)->nls_io;
fs/exfat/namei.c
820
exfat_set_volume_dirty(sb);
fs/exfat/namei.c
843
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
85
EXFAT_SB(sb)->options.keep_last_dots);
fs/exfat/namei.c
850
struct super_block *sb = dir->i_sb;
fs/exfat/namei.c
857
if (unlikely(exfat_forced_shutdown(sb)))
fs/exfat/namei.c
860
mutex_lock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
861
exfat_set_volume_dirty(sb);
fs/exfat/namei.c
875
inode = exfat_build_inode(sb, &info, i_pos);
fs/exfat/namei.c
888
mutex_unlock(&EXFAT_SB(sb)->s_lock);
fs/exfat/namei.c
892
static int exfat_check_dir_empty(struct super_block *sb,
fs/exfat/namei.c
900
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
912
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/namei.c
932
if (exfat_get_next_cluster(sb, &(clu.dir)))
fs/exfat/namei.c
94
hash = partial_name_hash(exfat_toupper(sb, c), hash);
fs/exfat/namei.c
948
struct super_block *sb = inode->i_sb;
fs/exfat/namei.c
949
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/namei.c
954
if (unlikely(exfat_forced_shutdown(sb)))
fs/exfat/namei.c
960
exfat_err(sb, "abnormal access to deleted dentry");
fs/exfat/namei.c
968
err = exfat_check_dir_empty(sb, &clu_to_free);
fs/exfat/namei.c
971
exfat_err(sb, "failed to exfat_check_dir_empty : err(%d)",
fs/exfat/namei.c
976
err = exfat_get_dentry_set_by_ei(&es, sb, ei);
fs/exfat/namei.c
982
exfat_set_volume_dirty(sb);
fs/exfat/nls.c
452
unsigned short exfat_toupper(struct super_block *sb, unsigned short a)
fs/exfat/nls.c
454
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/nls.c
468
int exfat_uniname_ncmp(struct super_block *sb, unsigned short *a,
fs/exfat/nls.c
474
if (exfat_toupper(sb, *a) != exfat_toupper(sb, *b))
fs/exfat/nls.c
479
static int exfat_utf16_to_utf8(struct super_block *sb,
fs/exfat/nls.c
493
static int exfat_utf8_to_utf16(struct super_block *sb,
fs/exfat/nls.c
506
exfat_err(sb, "failed to %s (err : %d) nls len : %d",
fs/exfat/nls.c
512
exfat_debug(sb, "failed to %s (estr:ENAMETOOLONG) nls len : %d, unilen : %d > %d",
fs/exfat/nls.c
522
upname[i] = cpu_to_le16(exfat_toupper(sb, *uniname));
fs/exfat/nls.c
540
static int __exfat_utf16_to_nls(struct super_block *sb,
fs/exfat/nls.c
547
struct nls_table *nls = EXFAT_SB(sb)->nls_io;
fs/exfat/nls.c
595
static int exfat_nls_to_ucs2(struct super_block *sb,
fs/exfat/nls.c
602
struct nls_table *nls = EXFAT_SB(sb)->nls_io;
fs/exfat/nls.c
614
upname[unilen] = cpu_to_le16(exfat_toupper(sb, *uniname));
fs/exfat/nls.c
629
int exfat_utf16_to_nls(struct super_block *sb, struct exfat_uni_name *uniname,
fs/exfat/nls.c
632
if (EXFAT_SB(sb)->options.utf8)
fs/exfat/nls.c
633
return exfat_utf16_to_utf8(sb, uniname, p_cstring,
fs/exfat/nls.c
635
return __exfat_utf16_to_nls(sb, uniname, p_cstring, buflen);
fs/exfat/nls.c
638
int exfat_nls_to_utf16(struct super_block *sb, const unsigned char *p_cstring,
fs/exfat/nls.c
641
if (EXFAT_SB(sb)->options.utf8)
fs/exfat/nls.c
642
return exfat_utf8_to_utf16(sb, p_cstring, len,
fs/exfat/nls.c
644
return exfat_nls_to_ucs2(sb, p_cstring, len, uniname, p_lossy);
fs/exfat/nls.c
647
static int exfat_load_upcase_table(struct super_block *sb,
fs/exfat/nls.c
651
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/nls.c
652
unsigned int sect_size = sb->s_blocksize;
fs/exfat/nls.c
668
bh = sb_bread(sb, sector);
fs/exfat/nls.c
670
exfat_err(sb, "failed to read sector(0x%llx)",
fs/exfat/nls.c
697
exfat_err(sb, "failed to load upcase table (idx : 0x%08x, chksum : 0x%08x, utbl_chksum : 0x%08x)",
fs/exfat/nls.c
702
static int exfat_load_default_upcase_table(struct super_block *sb)
fs/exfat/nls.c
705
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/nls.c
738
int exfat_create_upcase_table(struct super_block *sb)
fs/exfat/nls.c
744
unsigned char blksize_bits = sb->s_blocksize_bits;
fs/exfat/nls.c
747
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/nls.c
755
ep = exfat_get_dentry(sb, &clu, i, &bh);
fs/exfat/nls.c
775
ret = exfat_load_upcase_table(sb, sector, num_sectors,
fs/exfat/nls.c
789
if (exfat_get_next_cluster(sb, &clu.dir))
fs/exfat/nls.c
795
return exfat_load_default_upcase_table(sb);
fs/exfat/super.c
102
int exfat_set_volume_dirty(struct super_block *sb)
fs/exfat/super.c
104
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
106
return exfat_set_vol_flags(sb, sbi->vol_flags | VOLUME_DIRTY);
fs/exfat/super.c
109
int exfat_clear_volume_dirty(struct super_block *sb)
fs/exfat/super.c
111
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
113
return exfat_set_vol_flags(sb, sbi->vol_flags & ~VOLUME_DIRTY);
fs/exfat/super.c
118
struct super_block *sb = root->d_sb;
fs/exfat/super.c
119
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
155
int exfat_force_shutdown(struct super_block *sb, u32 flags)
fs/exfat/super.c
158
struct exfat_sb_info *sbi = sb->s_fs_info;
fs/exfat/super.c
161
if (exfat_forced_shutdown(sb))
fs/exfat/super.c
167
ret = bdev_freeze(sb->s_bdev);
fs/exfat/super.c
170
bdev_thaw(sb->s_bdev);
fs/exfat/super.c
185
static void exfat_shutdown(struct super_block *sb)
fs/exfat/super.c
187
exfat_force_shutdown(sb, EXFAT_GOING_DOWN_NOSYNC);
fs/exfat/super.c
190
static struct inode *exfat_alloc_inode(struct super_block *sb)
fs/exfat/super.c
194
ei = alloc_inode_sb(sb, exfat_inode_cachep, GFP_NOFS);
fs/exfat/super.c
344
static void exfat_hash_init(struct super_block *sb)
fs/exfat/super.c
346
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
356
struct super_block *sb = inode->i_sb;
fs/exfat/super.c
357
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
374
num_subdirs = exfat_count_dir_entries(sb, root_clu);
fs/exfat/super.c
396
static int exfat_calibrate_blocksize(struct super_block *sb, int logical_sect)
fs/exfat/super.c
398
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
401
exfat_err(sb, "bogus logical sector size %u", logical_sect);
fs/exfat/super.c
405
if (logical_sect < sb->s_blocksize) {
fs/exfat/super.c
406
exfat_err(sb, "logical sector size too small for device (logical sector size = %u)",
fs/exfat/super.c
411
if (logical_sect > sb->s_blocksize) {
fs/exfat/super.c
415
if (!sb_set_blocksize(sb, logical_sect)) {
fs/exfat/super.c
416
exfat_err(sb, "unable to set blocksize %u",
fs/exfat/super.c
420
sbi->boot_bh = sb_bread(sb, 0);
fs/exfat/super.c
422
exfat_err(sb, "unable to read boot sector (logical sector size = %lu)",
fs/exfat/super.c
423
sb->s_blocksize);
fs/exfat/super.c
430
static int exfat_read_boot_sector(struct super_block *sb)
fs/exfat/super.c
433
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
436
if (!sb_min_blocksize(sb, 512)) {
fs/exfat/super.c
437
exfat_err(sb, "unable to set blocksize");
fs/exfat/super.c
44
static void exfat_put_super(struct super_block *sb)
fs/exfat/super.c
442
sbi->boot_bh = sb_bread(sb, 0);
fs/exfat/super.c
444
exfat_err(sb, "unable to read boot sector");
fs/exfat/super.c
451
exfat_err(sb, "invalid boot record signature");
fs/exfat/super.c
456
exfat_err(sb, "invalid fs_name"); /* fs_name may unprintable */
fs/exfat/super.c
46
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
468
exfat_err(sb, "bogus number of FAT structure");
fs/exfat/super.c
477
exfat_err(sb, "bogus sector size bits : %u",
fs/exfat/super.c
486
exfat_err(sb, "bogus sectors bits per cluster : %u",
fs/exfat/super.c
49
exfat_clear_volume_dirty(sb);
fs/exfat/super.c
518
exfat_err(sb, "bogus fat length");
fs/exfat/super.c
525
exfat_err(sb, "bogus data start sector");
fs/exfat/super.c
530
exfat_warn(sb, "Volume was not properly unmounted. Some data may be corrupt. Please run fsck.");
fs/exfat/super.c
532
exfat_warn(sb, "Medium has reported failures. Some data may be lost.");
fs/exfat/super.c
535
sb->s_maxbytes = (u64)(sbi->num_clusters - EXFAT_RESERVED_CLUSTERS) <<
fs/exfat/super.c
539
if (exfat_calibrate_blocksize(sb, 1 << p_boot->sect_size_bits))
fs/exfat/super.c
545
static int exfat_verify_boot_region(struct super_block *sb)
fs/exfat/super.c
554
bh = sb_bread(sb, sn);
fs/exfat/super.c
560
p_sig = (__le32 *)&bh->b_data[sb->s_blocksize - 4];
fs/exfat/super.c
562
exfat_warn(sb, "Invalid exboot-signature(sector = %d): 0x%08x",
fs/exfat/super.c
566
chksum = exfat_calc_chksum32(bh->b_data, sb->s_blocksize,
fs/exfat/super.c
57
struct super_block *sb = dentry->d_sb;
fs/exfat/super.c
572
bh = sb_bread(sb, sn);
fs/exfat/super.c
576
for (i = 0; i < sb->s_blocksize; i += sizeof(u32)) {
fs/exfat/super.c
579
exfat_err(sb, "Invalid boot checksum (boot checksum : 0x%08x, checksum : 0x%08x)",
fs/exfat/super.c
58
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
59
unsigned long long id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/exfat/super.c
590
static int __exfat_fill_super(struct super_block *sb,
fs/exfat/super.c
594
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
596
ret = exfat_read_boot_sector(sb);
fs/exfat/super.c
598
exfat_err(sb, "failed to read boot sector");
fs/exfat/super.c
602
ret = exfat_verify_boot_region(sb);
fs/exfat/super.c
604
exfat_err(sb, "invalid boot region");
fs/exfat/super.c
61
buf->f_type = sb->s_magic;
fs/exfat/super.c
614
ret = exfat_count_num_clusters(sb, root_clu, &root_clu->size);
fs/exfat/super.c
616
exfat_err(sb, "failed to count the number of clusters in root");
fs/exfat/super.c
620
ret = exfat_create_upcase_table(sb);
fs/exfat/super.c
622
exfat_err(sb, "failed to load upcase table");
fs/exfat/super.c
626
ret = exfat_load_bitmap(sb);
fs/exfat/super.c
628
exfat_err(sb, "failed to load alloc-bitmap");
fs/exfat/super.c
632
if (!exfat_test_bitmap(sb, sbi->root_dir)) {
fs/exfat/super.c
633
exfat_warn(sb, "failed to test first cluster bit of root dir(%u)",
fs/exfat/super.c
640
exfat_set_bitmap(sb, sbi->root_dir, false);
fs/exfat/super.c
643
ret = exfat_count_used_clusters(sb, &sbi->used_clusters);
fs/exfat/super.c
645
exfat_err(sb, "failed to scan clusters");
fs/exfat/super.c
658
static int exfat_fill_super(struct super_block *sb, struct fs_context *fc)
fs/exfat/super.c
660
struct exfat_sb_info *sbi = sb->s_fs_info;
fs/exfat/super.c
669
if (opts->discard && !bdev_max_discard_sectors(sb->s_bdev)) {
fs/exfat/super.c
670
exfat_warn(sb, "mounting with \"discard\" option, but the device does not support discard");
fs/exfat/super.c
674
sb->s_flags |= SB_NODIRATIME;
fs/exfat/super.c
675
sb->s_magic = EXFAT_SUPER_MAGIC;
fs/exfat/super.c
676
sb->s_op = &exfat_sops;
fs/exfat/super.c
678
sb->s_time_gran = 10 * NSEC_PER_MSEC;
fs/exfat/super.c
679
sb->s_time_min = EXFAT_MIN_TIMESTAMP_SECS;
fs/exfat/super.c
680
sb->s_time_max = EXFAT_MAX_TIMESTAMP_SECS;
fs/exfat/super.c
682
err = __exfat_fill_super(sb, &root_clu);
fs/exfat/super.c
684
exfat_err(sb, "failed to recognize exfat type");
fs/exfat/super.c
689
exfat_hash_init(sb);
fs/exfat/super.c
692
set_default_d_op(sb, &exfat_utf8_dentry_ops);
fs/exfat/super.c
696
exfat_err(sb, "IO charset %s not found",
fs/exfat/super.c
701
set_default_d_op(sb, &exfat_dentry_ops);
fs/exfat/super.c
704
root_inode = new_inode(sb);
fs/exfat/super.c
706
exfat_err(sb, "failed to allocate root inode");
fs/exfat/super.c
715
exfat_err(sb, "failed to initialize root inode");
fs/exfat/super.c
72
static int exfat_set_vol_flags(struct super_block *sb, unsigned short new_flags)
fs/exfat/super.c
722
sb->s_root = d_make_root(root_inode);
fs/exfat/super.c
723
if (!sb->s_root) {
fs/exfat/super.c
724
exfat_err(sb, "failed to get the root dentry");
fs/exfat/super.c
733
sb->s_root = NULL;
fs/exfat/super.c
74
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
764
struct super_block *sb = fc->root->d_sb;
fs/exfat/super.c
766
struct exfat_sb_info *sbi = EXFAT_SB(sb);
fs/exfat/super.c
772
sync_filesystem(sb);
fs/exfat/super.c
774
exfat_clear_volume_dirty(sb);
fs/exfat/super.c
797
!bdev_max_discard_sectors(sb->s_bdev)) {
fs/exfat/super.c
798
exfat_warn(sb, "remounting with \"discard\" option, but the device does not support discard");
fs/exfat/super.c
828
struct super_block *sb = fc->root->d_sb;
fs/exfat/super.c
829
struct exfat_mount_options *cur_opts = &EXFAT_SB(sb)->options;
fs/exfat/super.c
860
static void exfat_kill_sb(struct super_block *sb)
fs/exfat/super.c
862
struct exfat_sb_info *sbi = sb->s_fs_info;
fs/exfat/super.c
864
kill_block_super(sb);
fs/exfat/super.c
89
if (sb_rdonly(sb))
fs/ext2/balloc.c
1013
struct super_block *sb, int size)
fs/ext2/balloc.c
1017
spinlock_t *rsv_lock = &EXT2_SB(sb)->s_rsv_window_lock;
fs/ext2/balloc.c
105
offset + EXT2_SB(sb)->s_itb_per_group - 1 > max_bit)
fs/ext2/balloc.c
1064
ext2_try_to_allocate_with_rsv(struct super_block *sb, unsigned int group,
fs/ext2/balloc.c
108
offset + EXT2_SB(sb)->s_itb_per_group,
fs/ext2/balloc.c
1080
return ext2_try_to_allocate(sb, group, bitmap_bh,
fs/ext2/balloc.c
1089
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
1090
group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
110
if (next_zero_bit >= offset + EXT2_SB(sb)->s_itb_per_group)
fs/ext2/balloc.c
1110
grp_goal, group, sb)) {
fs/ext2/balloc.c
1113
ret = alloc_new_reservation(my_rsv, grp_goal, sb,
fs/ext2/balloc.c
1119
grp_goal, group, sb))
fs/ext2/balloc.c
1126
try_to_extend_reservation(my_rsv, sb,
fs/ext2/balloc.c
1132
ext2_error(sb, __func__,
fs/ext2/balloc.c
1137
rsv_window_dump(&EXT2_SB(sb)->s_rsv_window_root, 1);
fs/ext2/balloc.c
1140
ret = ext2_try_to_allocate(sb, group, bitmap_bh, grp_goal,
fs/ext2/balloc.c
115
ext2_error(sb, __func__,
fs/ext2/balloc.c
1221
struct super_block *sb;
fs/ext2/balloc.c
1233
sb = inode->i_sb;
fs/ext2/balloc.c
1244
sbi = EXT2_SB(sb);
fs/ext2/balloc.c
1245
es = EXT2_SB(sb)->s_es;
fs/ext2/balloc.c
1274
EXT2_BLOCKS_PER_GROUP(sb);
fs/ext2/balloc.c
1277
gdp = ext2_get_group_desc(sb, group_no, &gdp_bh);
fs/ext2/balloc.c
129
read_block_bitmap(struct super_block *sb, unsigned int block_group)
fs/ext2/balloc.c
1293
EXT2_BLOCKS_PER_GROUP(sb));
fs/ext2/balloc.c
1301
bitmap_bh = read_block_bitmap(sb, group_no);
fs/ext2/balloc.c
1304
grp_alloc_blk = ext2_try_to_allocate_with_rsv(sb, group_no,
fs/ext2/balloc.c
1311
ngroups = EXT2_SB(sb)->s_groups_count;
fs/ext2/balloc.c
1322
gdp = ext2_get_group_desc(sb, group_no, &gdp_bh);
fs/ext2/balloc.c
1342
bitmap_bh = read_block_bitmap(sb, group_no);
fs/ext2/balloc.c
1348
grp_alloc_blk = ext2_try_to_allocate_with_rsv(sb, group_no,
fs/ext2/balloc.c
136
desc = ext2_get_group_desc(sb, block_group, NULL);
fs/ext2/balloc.c
1375
ret_block = grp_alloc_blk + ext2_group_first_block_no(sb, group_no);
fs/ext2/balloc.c
1380
EXT2_SB(sb)->s_itb_per_group) ||
fs/ext2/balloc.c
1382
EXT2_SB(sb)->s_itb_per_group)) {
fs/ext2/balloc.c
1383
ext2_error(sb, "ext2_new_blocks",
fs/ext2/balloc.c
1399
ext2_error(sb, "ext2_new_blocks",
fs/ext2/balloc.c
140
bh = sb_getblk(sb, bitmap_blk);
fs/ext2/balloc.c
1406
group_adjust_blocks(sb, group_no, gdp, gdp_bh, -num);
fs/ext2/balloc.c
1410
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ext2/balloc.c
142
ext2_error(sb, __func__,
fs/ext2/balloc.c
1445
unsigned long ext2_count_free_blocks (struct super_block * sb)
fs/ext2/balloc.c
1454
es = EXT2_SB(sb)->s_es;
fs/ext2/balloc.c
1458
for (i = 0; i < EXT2_SB(sb)->s_groups_count; i++) {
fs/ext2/balloc.c
1460
desc = ext2_get_group_desc (sb, i, NULL);
fs/ext2/balloc.c
1464
bitmap_bh = read_block_bitmap(sb, i);
fs/ext2/balloc.c
1468
x = ext2_count_free(bitmap_bh, sb->s_blocksize);
fs/ext2/balloc.c
1479
for (i = 0; i < EXT2_SB(sb)->s_groups_count; i++) {
fs/ext2/balloc.c
1480
desc = ext2_get_group_desc(sb, i, NULL);
fs/ext2/balloc.c
1514
int ext2_bg_has_super(struct super_block *sb, int group)
fs/ext2/balloc.c
1516
if (EXT2_HAS_RO_COMPAT_FEATURE(sb,EXT2_FEATURE_RO_COMPAT_SPARSE_SUPER)&&
fs/ext2/balloc.c
153
ext2_error(sb, __func__,
fs/ext2/balloc.c
1531
unsigned long ext2_bg_num_gdb(struct super_block *sb, int group)
fs/ext2/balloc.c
1533
return ext2_bg_has_super(sb, group) ? EXT2_SB(sb)->s_gdb_count : 0;
fs/ext2/balloc.c
160
ext2_valid_block_bitmap(sb, desc, block_group, bh);
fs/ext2/balloc.c
168
static void group_adjust_blocks(struct super_block *sb, int group_no,
fs/ext2/balloc.c
172
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/balloc.c
271
unsigned int group, struct super_block * sb)
fs/ext2/balloc.c
275
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
276
group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
335
void ext2_rsv_window_add(struct super_block *sb,
fs/ext2/balloc.c
338
struct rb_root *root = &EXT2_SB(sb)->s_rsv_window_root;
fs/ext2/balloc.c
374
static void rsv_window_remove(struct super_block *sb,
fs/ext2/balloc.c
380
rb_erase(&rsv->rsv_node, &EXT2_SB(sb)->s_rsv_window_root);
fs/ext2/balloc.c
39
struct ext2_group_desc * ext2_get_group_desc(struct super_block * sb,
fs/ext2/balloc.c
420
struct super_block *sb = inode->i_sb;
fs/ext2/balloc.c
434
if (!test_opt(sb, RESERVATION))
fs/ext2/balloc.c
46
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/balloc.c
491
struct super_block * sb = inode->i_sb;
fs/ext2/balloc.c
492
struct ext2_sb_info * sbi = EXT2_SB(sb);
fs/ext2/balloc.c
498
ext2_error (sb, "ext2_free_blocks",
fs/ext2/balloc.c
509
EXT2_BLOCKS_PER_GROUP(sb);
fs/ext2/balloc.c
511
EXT2_BLOCKS_PER_GROUP(sb);
fs/ext2/balloc.c
516
if (bit + count > EXT2_BLOCKS_PER_GROUP(sb)) {
fs/ext2/balloc.c
517
overflow = bit + count - EXT2_BLOCKS_PER_GROUP(sb);
fs/ext2/balloc.c
521
bitmap_bh = read_block_bitmap(sb, block_group);
fs/ext2/balloc.c
525
desc = ext2_get_group_desc (sb, block_group, &bh2);
fs/ext2/balloc.c
535
ext2_error (sb, "ext2_free_blocks",
fs/ext2/balloc.c
545
ext2_error(sb, __func__,
fs/ext2/balloc.c
553
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ext2/balloc.c
556
group_adjust_blocks(sb, block_group, desc, bh2, group_freed);
fs/ext2/balloc.c
56
group_desc = block_group >> EXT2_DESC_PER_BLOCK_BITS(sb);
fs/ext2/balloc.c
57
offset = block_group & (EXT2_DESC_PER_BLOCK(sb) - 1);
fs/ext2/balloc.c
668
ext2_try_to_allocate(struct super_block *sb, int group,
fs/ext2/balloc.c
673
ext2_fsblk_t group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
674
ext2_fsblk_t group_last_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
689
BUG_ON(start > EXT2_BLOCKS_PER_GROUP(sb));
fs/ext2/balloc.c
707
if (ext2_set_bit_atomic(sb_bgl_lock(EXT2_SB(sb), group),
fs/ext2/balloc.c
71
static int ext2_valid_block_bitmap(struct super_block *sb,
fs/ext2/balloc.c
758
struct super_block * sb,
fs/ext2/balloc.c
82
group_first_block = ext2_group_first_block_no(sb, block_group);
fs/ext2/balloc.c
821
rsv_window_remove(sb, my_rsv);
fs/ext2/balloc.c
83
max_bit = ext2_group_last_block_no(sb, block_group) - group_first_block;
fs/ext2/balloc.c
835
ext2_rsv_window_add(sb, my_rsv);
fs/ext2/balloc.c
871
ext2_grpblk_t grp_goal, struct super_block *sb,
fs/ext2/balloc.c
877
struct rb_root *fs_rsv_root = &EXT2_SB(sb)->s_rsv_window_root;
fs/ext2/balloc.c
880
spinlock_t *rsv_lock = &EXT2_SB(sb)->s_rsv_window_lock;
fs/ext2/balloc.c
882
group_first_block = ext2_group_first_block_no(sb, group);
fs/ext2/balloc.c
883
group_end_block = ext2_group_last_block_no(sb, group);
fs/ext2/balloc.c
941
ret = find_next_reservable_window(search_head, my_rsv, sb,
fs/ext2/balloc.c
946
rsv_window_remove(sb, my_rsv);
fs/ext2/balloc.c
972
rsv_window_remove(sb, my_rsv);
fs/ext2/dir.c
103
struct super_block *sb = dir->i_sb;
fs/ext2/dir.c
105
u32 max_inumber = le32_to_cpu(EXT2_SB(sb)->s_es->s_inodes_count);
fs/ext2/dir.c
143
ext2_error(sb, __func__,
fs/ext2/dir.c
163
ext2_error(sb, __func__, "bad entry in directory #%lu: : %s - "
fs/ext2/dir.c
172
ext2_error(sb, "ext2_check_folio",
fs/ext2/dir.c
262
struct super_block *sb = inode->i_sb;
fs/ext2/dir.c
274
EXT2_HAS_INCOMPAT_FEATURE(sb, EXT2_FEATURE_INCOMPAT_FILETYPE);
fs/ext2/dir.c
283
ext2_error(sb, __func__,
fs/ext2/dir.c
301
ext2_error(sb, __func__,
fs/ext2/ext2.h
169
static inline struct ext2_sb_info *EXT2_SB(struct super_block *sb)
fs/ext2/ext2.h
171
return sb->s_fs_info;
fs/ext2/ext2.h
385
#define test_opt(sb, opt) (EXT2_SB(sb)->s_mount_opt & \
fs/ext2/ext2.h
506
#define EXT2_HAS_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
507
( EXT2_SB(sb)->s_es->s_feature_compat & cpu_to_le32(mask) )
fs/ext2/ext2.h
508
#define EXT2_HAS_RO_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
509
( EXT2_SB(sb)->s_es->s_feature_ro_compat & cpu_to_le32(mask) )
fs/ext2/ext2.h
510
#define EXT2_HAS_INCOMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
511
( EXT2_SB(sb)->s_es->s_feature_incompat & cpu_to_le32(mask) )
fs/ext2/ext2.h
512
#define EXT2_SET_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
513
EXT2_SB(sb)->s_es->s_feature_compat |= cpu_to_le32(mask)
fs/ext2/ext2.h
514
#define EXT2_SET_RO_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
515
EXT2_SB(sb)->s_es->s_feature_ro_compat |= cpu_to_le32(mask)
fs/ext2/ext2.h
516
#define EXT2_SET_INCOMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
517
EXT2_SB(sb)->s_es->s_feature_incompat |= cpu_to_le32(mask)
fs/ext2/ext2.h
518
#define EXT2_CLEAR_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
519
EXT2_SB(sb)->s_es->s_feature_compat &= ~cpu_to_le32(mask)
fs/ext2/ext2.h
520
#define EXT2_CLEAR_RO_COMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
521
EXT2_SB(sb)->s_es->s_feature_ro_compat &= ~cpu_to_le32(mask)
fs/ext2/ext2.h
522
#define EXT2_CLEAR_INCOMPAT_FEATURE(sb,mask) \
fs/ext2/ext2.h
523
EXT2_SB(sb)->s_es->s_feature_incompat &= ~cpu_to_le32(mask)
fs/ext2/ext2.h
702
extern int ext2_bg_has_super(struct super_block *sb, int group);
fs/ext2/ext2.h
703
extern unsigned long ext2_bg_num_gdb(struct super_block *sb, int group);
fs/ext2/ext2.h
711
extern struct ext2_group_desc * ext2_get_group_desc(struct super_block * sb,
fs/ext2/ext2.h
715
extern int ext2_should_retry_alloc(struct super_block *sb, int *retries);
fs/ext2/ext2.h
717
extern void ext2_rsv_window_add(struct super_block *sb, struct ext2_reserve_window_node *rsv);
fs/ext2/ext2.h
766
extern void ext2_update_dynamic_rev (struct super_block *sb);
fs/ext2/ext2.h
767
extern void ext2_sync_super(struct super_block *sb, struct ext2_super_block *es,
fs/ext2/ext2.h
797
ext2_group_first_block_no(struct super_block *sb, unsigned long group_no)
fs/ext2/ext2.h
799
return group_no * (ext2_fsblk_t)EXT2_BLOCKS_PER_GROUP(sb) +
fs/ext2/ext2.h
800
le32_to_cpu(EXT2_SB(sb)->s_es->s_first_data_block);
fs/ext2/ext2.h
804
ext2_group_last_block_no(struct super_block *sb, unsigned long group_no)
fs/ext2/ext2.h
806
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/ext2.h
811
return ext2_group_first_block_no(sb, group_no) +
fs/ext2/ext2.h
812
EXT2_BLOCKS_PER_GROUP(sb) - 1;
fs/ext2/file.c
159
struct super_block *sb = file->f_mapping->host->i_sb;
fs/ext2/file.c
164
ext2_error(sb, __func__,
fs/ext2/ialloc.c
107
struct super_block * sb = inode->i_sb;
fs/ext2/ialloc.c
126
es = EXT2_SB(sb)->s_es;
fs/ext2/ialloc.c
129
if (ino < EXT2_FIRST_INO(sb) ||
fs/ext2/ialloc.c
131
ext2_error (sb, "ext2_free_inode",
fs/ext2/ialloc.c
135
block_group = (ino - 1) / EXT2_INODES_PER_GROUP(sb);
fs/ext2/ialloc.c
136
bit = (ino - 1) % EXT2_INODES_PER_GROUP(sb);
fs/ext2/ialloc.c
137
bitmap_bh = read_inode_bitmap(sb, block_group);
fs/ext2/ialloc.c
142
if (!ext2_clear_bit_atomic(sb_bgl_lock(EXT2_SB(sb), block_group),
fs/ext2/ialloc.c
144
ext2_error (sb, "ext2_free_inode",
fs/ext2/ialloc.c
147
ext2_release_inode(sb, block_group, is_directory);
fs/ext2/ialloc.c
149
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ext2/ialloc.c
199
static int find_group_dir(struct super_block *sb, struct inode *parent)
fs/ext2/ialloc.c
201
int ngroups = EXT2_SB(sb)->s_groups_count;
fs/ext2/ialloc.c
202
int avefreei = ext2_count_free_inodes(sb) / ngroups;
fs/ext2/ialloc.c
207
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
251
static int find_group_orlov(struct super_block *sb, struct inode *parent)
fs/ext2/ialloc.c
254
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/ialloc.c
257
int inodes_per_group = EXT2_INODES_PER_GROUP(sb);
fs/ext2/ialloc.c
274
if ((parent == d_inode(sb->s_root)) ||
fs/ext2/ialloc.c
282
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
308
min_blocks = avefreeb - EXT2_BLOCKS_PER_GROUP(sb) / 4;
fs/ext2/ialloc.c
310
max_debt = EXT2_BLOCKS_PER_GROUP(sb) / max(blocks_per_dir, BLOCK_COST);
fs/ext2/ialloc.c
320
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
337
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
359
static int find_group_other(struct super_block *sb, struct inode *parent)
fs/ext2/ialloc.c
362
int ngroups = EXT2_SB(sb)->s_groups_count;
fs/ext2/ialloc.c
370
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
394
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
408
desc = ext2_get_group_desc (sb, group, NULL);
fs/ext2/ialloc.c
422
struct super_block *sb;
fs/ext2/ialloc.c
434
sb = dir->i_sb;
fs/ext2/ialloc.c
435
inode = new_inode(sb);
fs/ext2/ialloc.c
440
sbi = EXT2_SB(sb);
fs/ext2/ialloc.c
443
if (test_opt(sb, OLDALLOC))
fs/ext2/ialloc.c
444
group = find_group_dir(sb, dir);
fs/ext2/ialloc.c
446
group = find_group_orlov(sb, dir);
fs/ext2/ialloc.c
448
group = find_group_other(sb, dir);
fs/ext2/ialloc.c
456
gdp = ext2_get_group_desc(sb, group, &bh2);
fs/ext2/ialloc.c
463
bitmap_bh = read_inode_bitmap(sb, group);
fs/ext2/ialloc.c
47
read_inode_bitmap(struct super_block * sb, unsigned long block_group)
fs/ext2/ialloc.c
472
EXT2_INODES_PER_GROUP(sb), ino);
fs/ext2/ialloc.c
473
if (ino >= EXT2_INODES_PER_GROUP(sb)) {
fs/ext2/ialloc.c
489
if (++ino >= EXT2_INODES_PER_GROUP(sb)) {
fs/ext2/ialloc.c
509
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ext2/ialloc.c
513
ino += group * EXT2_INODES_PER_GROUP(sb) + 1;
fs/ext2/ialloc.c
514
if (ino < EXT2_FIRST_INO(sb) || ino > le32_to_cpu(es->s_inodes_count)) {
fs/ext2/ialloc.c
515
ext2_error (sb, "ext2_new_inode",
fs/ext2/ialloc.c
52
desc = ext2_get_group_desc(sb, block_group, NULL);
fs/ext2/ialloc.c
540
if (test_opt(sb, GRPID)) {
fs/ext2/ialloc.c
56
bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap));
fs/ext2/ialloc.c
568
ext2_error(sb, "ext2_new_inode",
fs/ext2/ialloc.c
58
ext2_error(sb, "read_inode_bitmap",
fs/ext2/ialloc.c
612
unsigned long ext2_count_free_inodes (struct super_block * sb)
fs/ext2/ialloc.c
623
es = EXT2_SB(sb)->s_es;
fs/ext2/ialloc.c
624
for (i = 0; i < EXT2_SB(sb)->s_groups_count; i++) {
fs/ext2/ialloc.c
627
desc = ext2_get_group_desc (sb, i, NULL);
fs/ext2/ialloc.c
632
bitmap_bh = read_inode_bitmap(sb, i);
fs/ext2/ialloc.c
636
x = ext2_count_free(bitmap_bh, EXT2_INODES_PER_GROUP(sb) / 8);
fs/ext2/ialloc.c
644
percpu_counter_read(&EXT2_SB(sb)->s_freeinodes_counter),
fs/ext2/ialloc.c
648
for (i = 0; i < EXT2_SB(sb)->s_groups_count; i++) {
fs/ext2/ialloc.c
649
desc = ext2_get_group_desc (sb, i, NULL);
fs/ext2/ialloc.c
659
unsigned long ext2_count_dirs (struct super_block * sb)
fs/ext2/ialloc.c
66
static void ext2_release_inode(struct super_block *sb, int group, int dir)
fs/ext2/ialloc.c
664
for (i = 0; i < EXT2_SB(sb)->s_groups_count; i++) {
fs/ext2/ialloc.c
665
struct ext2_group_desc *gdp = ext2_get_group_desc (sb, i, NULL);
fs/ext2/ialloc.c
71
desc = ext2_get_group_desc(sb, group, &bh);
fs/ext2/ialloc.c
73
ext2_error(sb, "ext2_release_inode",
fs/ext2/ialloc.c
78
spin_lock(sb_bgl_lock(EXT2_SB(sb), group));
fs/ext2/ialloc.c
82
spin_unlock(sb_bgl_lock(EXT2_SB(sb), group));
fs/ext2/ialloc.c
83
percpu_counter_inc(&EXT2_SB(sb)->s_freeinodes_counter);
fs/ext2/ialloc.c
85
percpu_counter_dec(&EXT2_SB(sb)->s_dirs_counter);
fs/ext2/inode.c
1314
static struct ext2_inode *ext2_get_inode(struct super_block *sb, ino_t ino,
fs/ext2/inode.c
1324
if ((ino != EXT2_ROOT_INO && ino < EXT2_FIRST_INO(sb)) ||
fs/ext2/inode.c
1325
ino > le32_to_cpu(EXT2_SB(sb)->s_es->s_inodes_count))
fs/ext2/inode.c
1328
block_group = (ino - 1) / EXT2_INODES_PER_GROUP(sb);
fs/ext2/inode.c
1329
gdp = ext2_get_group_desc(sb, block_group, NULL);
fs/ext2/inode.c
1335
offset = ((ino - 1) % EXT2_INODES_PER_GROUP(sb)) * EXT2_INODE_SIZE(sb);
fs/ext2/inode.c
1337
(offset >> EXT2_BLOCK_SIZE_BITS(sb));
fs/ext2/inode.c
1338
if (!(bh = sb_bread(sb, block)))
fs/ext2/inode.c
1342
offset &= (EXT2_BLOCK_SIZE(sb) - 1);
fs/ext2/inode.c
1346
ext2_error(sb, "ext2_get_inode", "bad inode number: %lu",
fs/ext2/inode.c
1350
ext2_error(sb, "ext2_get_inode",
fs/ext2/inode.c
1387
struct inode *ext2_iget (struct super_block *sb, unsigned long ino)
fs/ext2/inode.c
1398
inode = iget_locked(sb, ino);
fs/ext2/inode.c
1448
!ext2_data_block_valid(EXT2_SB(sb), ei->i_file_acl, 1)) {
fs/ext2/inode.c
1449
ext2_error(sb, "ext2_iget", "bad extended attribute block %u",
fs/ext2/inode.c
1515
struct super_block *sb = inode->i_sb;
fs/ext2/inode.c
1520
struct ext2_inode * raw_inode = ext2_get_inode(sb, ino, &bh);
fs/ext2/inode.c
1530
memset(raw_inode, 0, EXT2_SB(sb)->s_inode_size);
fs/ext2/inode.c
1533
if (!(test_opt(sb, NO_UID32))) {
fs/ext2/inode.c
1571
if (!EXT2_HAS_RO_COMPAT_FEATURE(sb,
fs/ext2/inode.c
1573
EXT2_SB(sb)->s_es->s_rev_level ==
fs/ext2/inode.c
1578
spin_lock(&EXT2_SB(sb)->s_lock);
fs/ext2/inode.c
1579
ext2_update_dynamic_rev(sb);
fs/ext2/inode.c
1580
EXT2_SET_RO_COMPAT_FEATURE(sb,
fs/ext2/inode.c
1582
spin_unlock(&EXT2_SB(sb)->s_lock);
fs/ext2/inode.c
1583
ext2_sync_super(sb, EXT2_SB(sb)->s_es, 1);
fs/ext2/inode.c
1607
sb->s_id, (unsigned long) ino);
fs/ext2/inode.c
240
struct super_block *sb = inode->i_sb;
fs/ext2/inode.c
250
bh = sb_bread(sb, le32_to_cpu(p->key));
fs/ext2/namei.c
160
struct super_block * sb = dir->i_sb;
fs/ext2/namei.c
165
if (l > sb->s_blocksize)
fs/ext2/super.c
1000
ext2_msg(sb, KERN_ERR, "unsupported blocksize for DAX\n");
fs/ext2/super.c
1006
if (sb->s_blocksize != blocksize) {
fs/ext2/super.c
1009
if (!sb_set_blocksize(sb, blocksize)) {
fs/ext2/super.c
1010
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1017
bh = sb_bread(sb, logic_sb_block);
fs/ext2/super.c
1019
ext2_msg(sb, KERN_ERR, "error: couldn't read"
fs/ext2/super.c
1026
ext2_msg(sb, KERN_ERR, "error: magic mismatch");
fs/ext2/super.c
1031
sb->s_maxbytes = ext2_max_size(sb->s_blocksize_bits);
fs/ext2/super.c
1032
sb->s_max_links = EXT2_LINK_MAX;
fs/ext2/super.c
1033
sb->s_time_min = S32_MIN;
fs/ext2/super.c
1034
sb->s_time_max = S32_MAX;
fs/ext2/super.c
1045
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1055
sbi->s_inodes_per_block = sb->s_blocksize / EXT2_INODE_SIZE(sb);
fs/ext2/super.c
1060
sbi->s_desc_per_block = sb->s_blocksize /
fs/ext2/super.c
1065
ilog2 (EXT2_ADDR_PER_BLOCK(sb));
fs/ext2/super.c
1067
ilog2 (EXT2_DESC_PER_BLOCK(sb));
fs/ext2/super.c
1069
if (sb->s_magic != EXT2_SUPER_MAGIC)
fs/ext2/super.c
1072
if (sb->s_blocksize != bh->b_size) {
fs/ext2/super.c
1074
ext2_msg(sb, KERN_ERR, "error: unsupported blocksize");
fs/ext2/super.c
1079
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1081
le32_to_cpu(es->s_log_frag_size), sb->s_blocksize_bits);
fs/ext2/super.c
1085
if (sbi->s_blocks_per_group > sb->s_blocksize * 8) {
fs/ext2/super.c
1086
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1093
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1099
sbi->s_inodes_per_group > sb->s_blocksize * 8) {
fs/ext2/super.c
1100
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1105
if (sb_bdev_nr_blocks(sb) < le32_to_cpu(es->s_blocks_count)) {
fs/ext2/super.c
1106
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1109
(unsigned)sb_bdev_nr_blocks(sb));
fs/ext2/super.c
111
void ext2_msg(struct super_block *sb, const char *prefix,
fs/ext2/super.c
1115
/ EXT2_BLOCKS_PER_GROUP(sb)) + 1;
fs/ext2/super.c
1118
ext2_msg(sb, KERN_ERR, "error: invalid #inodes: %u vs computed %llu",
fs/ext2/super.c
1123
db_count = (sbi->s_groups_count + EXT2_DESC_PER_BLOCK(sb) - 1) /
fs/ext2/super.c
1124
EXT2_DESC_PER_BLOCK(sb);
fs/ext2/super.c
1128
ext2_msg(sb, KERN_ERR, "error: not enough memory");
fs/ext2/super.c
1135
ext2_msg(sb, KERN_ERR, "error: not enough memory");
fs/ext2/super.c
1139
block = descriptor_loc(sb, logic_sb_block, i);
fs/ext2/super.c
1140
sbi->s_group_desc[i] = sb_bread(sb, block);
fs/ext2/super.c
1144
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1149
if (!ext2_check_descriptors (sb)) {
fs/ext2/super.c
1150
ext2_msg(sb, KERN_ERR, "group descriptors corrupted");
fs/ext2/super.c
1170
ext2_rsv_window_add(sb, &sbi->s_rsv_window_head);
fs/ext2/super.c
1173
ext2_count_free_blocks(sb), GFP_KERNEL);
fs/ext2/super.c
1176
ext2_count_free_inodes(sb), GFP_KERNEL);
fs/ext2/super.c
1180
ext2_count_dirs(sb), GFP_KERNEL);
fs/ext2/super.c
1184
ext2_msg(sb, KERN_ERR, "error: insufficient memory");
fs/ext2/super.c
1192
ext2_msg(sb, KERN_ERR, "Failed to create ea_block_cache");
fs/ext2/super.c
1199
sb->s_op = &ext2_sops;
fs/ext2/super.c
1200
sb->s_export_op = &ext2_export_ops;
fs/ext2/super.c
1201
sb->s_xattr = ext2_xattr_handlers;
fs/ext2/super.c
1204
sb->dq_op = &dquot_operations;
fs/ext2/super.c
1205
sb->s_qcop = &ext2_quotactl_ops;
fs/ext2/super.c
1206
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
fs/ext2/super.c
1209
root = ext2_iget(sb, EXT2_ROOT_INO);
fs/ext2/super.c
1216
ext2_msg(sb, KERN_ERR, "error: corrupt root inode, run e2fsck");
fs/ext2/super.c
122
printk("%sEXT2-fs (%s): %pV\n", prefix, sb->s_id, &vaf);
fs/ext2/super.c
1220
sb->s_root = d_make_root(root);
fs/ext2/super.c
1221
if (!sb->s_root) {
fs/ext2/super.c
1222
ext2_msg(sb, KERN_ERR, "error: get root inode failed");
fs/ext2/super.c
1226
if (EXT2_HAS_COMPAT_FEATURE(sb, EXT3_FEATURE_COMPAT_HAS_JOURNAL))
fs/ext2/super.c
1227
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
1229
if (ext2_setup_super (sb, es, sb_rdonly(sb)))
fs/ext2/super.c
1230
sb->s_flags |= SB_RDONLY;
fs/ext2/super.c
1231
ext2_write_super(sb);
fs/ext2/super.c
1236
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1238
sb->s_id);
fs/ext2/super.c
1255
sb->s_fs_info = NULL;
fs/ext2/super.c
1261
static void ext2_clear_super_error(struct super_block *sb)
fs/ext2/super.c
1263
struct buffer_head *sbh = EXT2_SB(sb)->s_sbh;
fs/ext2/super.c
1274
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
1281
void ext2_sync_super(struct super_block *sb, struct ext2_super_block *es,
fs/ext2/super.c
1284
ext2_clear_super_error(sb);
fs/ext2/super.c
1285
spin_lock(&EXT2_SB(sb)->s_lock);
fs/ext2/super.c
1286
es->s_free_blocks_count = cpu_to_le32(ext2_count_free_blocks(sb));
fs/ext2/super.c
1287
es->s_free_inodes_count = cpu_to_le32(ext2_count_free_inodes(sb));
fs/ext2/super.c
1290
spin_unlock(&EXT2_SB(sb)->s_lock);
fs/ext2/super.c
1291
mark_buffer_dirty(EXT2_SB(sb)->s_sbh);
fs/ext2/super.c
1293
sync_dirty_buffer(EXT2_SB(sb)->s_sbh);
fs/ext2/super.c
130
void ext2_update_dynamic_rev(struct super_block *sb)
fs/ext2/super.c
1306
static int ext2_sync_fs(struct super_block *sb, int wait)
fs/ext2/super.c
1308
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
1309
struct ext2_super_block *es = EXT2_SB(sb)->s_es;
fs/ext2/super.c
1315
dquot_writeback_dquots(sb, -1);
fs/ext2/super.c
132
struct ext2_super_block *es = EXT2_SB(sb)->s_es;
fs/ext2/super.c
1323
ext2_sync_super(sb, es, wait);
fs/ext2/super.c
1327
static int ext2_freeze(struct super_block *sb)
fs/ext2/super.c
1329
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
1336
if (atomic_long_read(&sb->s_remove_count)) {
fs/ext2/super.c
1337
ext2_sync_fs(sb, 1);
fs/ext2/super.c
1344
ext2_sync_super(sb, sbi->s_es, 1);
fs/ext2/super.c
1349
static int ext2_unfreeze(struct super_block *sb)
fs/ext2/super.c
1352
ext2_write_super(sb);
fs/ext2/super.c
1357
static void ext2_write_super(struct super_block *sb)
fs/ext2/super.c
1359
if (!sb_rdonly(sb))
fs/ext2/super.c
1360
ext2_sync_fs(sb, 1);
fs/ext2/super.c
1366
struct super_block *sb = fc->root->d_sb;
fs/ext2/super.c
1367
struct ext2_sb_info * sbi = EXT2_SB(sb);
fs/ext2/super.c
137
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
1373
sync_filesystem(sb);
fs/ext2/super.c
1382
ext2_msg(sb, KERN_WARNING, "warning: refusing change of "
fs/ext2/super.c
1386
if ((bool)(flags & SB_RDONLY) == sb_rdonly(sb))
fs/ext2/super.c
1401
err = dquot_suspend(sb, -1);
fs/ext2/super.c
1405
ext2_sync_super(sb, es, 1);
fs/ext2/super.c
1407
__le32 ret = EXT2_HAS_RO_COMPAT_FEATURE(sb,
fs/ext2/super.c
1411
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
1423
if (!ext2_setup_super (sb, es, 0))
fs/ext2/super.c
1424
sb->s_flags &= ~SB_RDONLY;
fs/ext2/super.c
1427
ext2_write_super(sb);
fs/ext2/super.c
1429
dquot_resume(sb, -1);
fs/ext2/super.c
1437
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/ext2/super.c
1438
(test_opt(sb, POSIX_ACL) ? SB_POSIXACL : 0);
fs/ext2/super.c
1446
struct super_block *sb = dentry->d_sb;
fs/ext2/super.c
1447
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
1452
if (test_opt (sb, MINIX_DF))
fs/ext2/super.c
1476
overhead += ext2_bg_has_super(sb, i) +
fs/ext2/super.c
1477
ext2_bg_num_gdb(sb, i);
fs/ext2/super.c
1491
buf->f_bsize = sb->s_blocksize;
fs/ext2/super.c
1493
buf->f_bfree = ext2_count_free_blocks(sb);
fs/ext2/super.c
1499
buf->f_ffree = ext2_count_free_inodes(sb);
fs/ext2/super.c
1518
static ssize_t ext2_quota_read(struct super_block *sb, int type, char *data,
fs/ext2/super.c
1521
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext2/super.c
1522
sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb);
fs/ext2/super.c
1524
int offset = off & (sb->s_blocksize - 1);
fs/ext2/super.c
1537
tocopy = min_t(size_t, sb->s_blocksize - offset, toread);
fs/ext2/super.c
1540
tmp_bh.b_size = sb->s_blocksize;
fs/ext2/super.c
1547
bh = sb_bread(sb, tmp_bh.b_blocknr);
fs/ext2/super.c
156
static int ext2_quota_off(struct super_block *sb, int type);
fs/ext2/super.c
1562
static ssize_t ext2_quota_write(struct super_block *sb, int type,
fs/ext2/super.c
1565
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext2/super.c
1566
sector_t blk = off >> EXT2_BLOCK_SIZE_BITS(sb);
fs/ext2/super.c
1568
int offset = off & (sb->s_blocksize - 1);
fs/ext2/super.c
1575
tocopy = min_t(size_t, sb->s_blocksize - offset, towrite);
fs/ext2/super.c
1578
tmp_bh.b_size = sb->s_blocksize;
fs/ext2/super.c
158
static void ext2_quota_off_umount(struct super_block *sb)
fs/ext2/super.c
1582
if (offset || tocopy != EXT2_BLOCK_SIZE(sb))
fs/ext2/super.c
1583
bh = sb_bread(sb, tmp_bh.b_blocknr);
fs/ext2/super.c
1585
bh = sb_getblk(sb, tmp_bh.b_blocknr);
fs/ext2/super.c
1613
static int ext2_quota_on(struct super_block *sb, int type, int format_id,
fs/ext2/super.c
1619
err = dquot_quota_on(sb, type, format_id, path);
fs/ext2/super.c
163
ext2_quota_off(sb, type);
fs/ext2/super.c
1634
static int ext2_quota_off(struct super_block *sb, int type)
fs/ext2/super.c
1636
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext2/super.c
1642
err = dquot_quota_off(sb, type);
fs/ext2/super.c
1655
return dquot_quota_off(sb, type);
fs/ext2/super.c
166
static inline void ext2_quota_off_umount(struct super_block *sb)
fs/ext2/super.c
1676
struct super_block *sb = fc->root->d_sb;
fs/ext2/super.c
1677
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
1681
ctx->vals_s_flags = sb->s_flags;
fs/ext2/super.c
171
static void ext2_put_super (struct super_block * sb)
fs/ext2/super.c
175
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
177
ext2_quota_off_umount(sb);
fs/ext2/super.c
182
if (!sb_rdonly(sb)) {
fs/ext2/super.c
188
ext2_sync_super(sb, es, 1);
fs/ext2/super.c
199
sb->s_fs_info = NULL;
fs/ext2/super.c
207
static struct inode *ext2_alloc_inode(struct super_block *sb)
fs/ext2/super.c
210
ei = alloc_inode_sb(sb, ext2_inode_cachep, GFP_KERNEL);
fs/ext2/super.c
264
struct super_block *sb = root->d_sb;
fs/ext2/super.c
265
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
274
if (test_opt(sb, MINIX_DF))
fs/ext2/super.c
276
if (test_opt(sb, GRPID))
fs/ext2/super.c
278
if (!test_opt(sb, GRPID) && (def_mount_opts & EXT2_DEFM_BSDGROUPS))
fs/ext2/super.c
290
if (test_opt(sb, ERRORS_RO)) {
fs/ext2/super.c
298
if (test_opt(sb, ERRORS_CONT))
fs/ext2/super.c
300
if (test_opt(sb, ERRORS_PANIC))
fs/ext2/super.c
302
if (test_opt(sb, NO_UID32))
fs/ext2/super.c
304
if (test_opt(sb, DEBUG))
fs/ext2/super.c
306
if (test_opt(sb, OLDALLOC))
fs/ext2/super.c
310
if (test_opt(sb, XATTR_USER))
fs/ext2/super.c
312
if (!test_opt(sb, XATTR_USER) &&
fs/ext2/super.c
319
if (test_opt(sb, POSIX_ACL))
fs/ext2/super.c
321
if (!test_opt(sb, POSIX_ACL) && (def_mount_opts & EXT2_DEFM_ACL))
fs/ext2/super.c
325
if (test_opt(sb, USRQUOTA))
fs/ext2/super.c
328
if (test_opt(sb, GRPQUOTA))
fs/ext2/super.c
331
if (test_opt(sb, XIP))
fs/ext2/super.c
334
if (test_opt(sb, DAX))
fs/ext2/super.c
337
if (!test_opt(sb, RESERVATION))
fs/ext2/super.c
345
static ssize_t ext2_quota_read(struct super_block *sb, int type, char *data, size_t len, loff_t off);
fs/ext2/super.c
346
static ssize_t ext2_quota_write(struct super_block *sb, int type, const char *data, size_t len, loff_t off);
fs/ext2/super.c
347
static int ext2_quota_on(struct super_block *sb, int type, int format_id,
fs/ext2/super.c
384
static struct inode *ext2_nfs_get_inode(struct super_block *sb,
fs/ext2/super.c
389
if (ino < EXT2_FIRST_INO(sb) && ino != EXT2_ROOT_INO)
fs/ext2/super.c
391
if (ino > le32_to_cpu(EXT2_SB(sb)->s_es->s_inodes_count))
fs/ext2/super.c
399
inode = ext2_iget(sb, ino);
fs/ext2/super.c
410
static struct dentry *ext2_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/ext2/super.c
413
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/ext2/super.c
417
static struct dentry *ext2_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/ext2/super.c
420
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/ext2/super.c
43
static void ext2_write_super(struct super_block *sb);
fs/ext2/super.c
45
static int ext2_sync_fs(struct super_block *sb, int wait);
fs/ext2/super.c
46
static int ext2_freeze(struct super_block *sb);
fs/ext2/super.c
47
static int ext2_unfreeze(struct super_block *sb);
fs/ext2/super.c
49
void ext2_error(struct super_block *sb, const char *function,
fs/ext2/super.c
54
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
57
if (!sb_rdonly(sb)) {
fs/ext2/super.c
62
ext2_sync_super(sb, es, 1);
fs/ext2/super.c
645
static int ext2_setup_super (struct super_block * sb,
fs/ext2/super.c
650
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
653
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
661
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
665
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
671
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
678
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
684
if (test_opt (sb, DEBUG))
fs/ext2/super.c
685
ext2_msg(sb, KERN_INFO, "%s, %s, bs=%lu, gc=%lu, "
fs/ext2/super.c
687
EXT2FS_VERSION, EXT2FS_DATE, sb->s_blocksize,
fs/ext2/super.c
689
EXT2_BLOCKS_PER_GROUP(sb),
fs/ext2/super.c
690
EXT2_INODES_PER_GROUP(sb),
fs/ext2/super.c
695
static int ext2_check_descriptors(struct super_block *sb)
fs/ext2/super.c
698
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
703
struct ext2_group_desc *gdp = ext2_get_group_desc(sb, i, NULL);
fs/ext2/super.c
704
ext2_fsblk_t first_block = ext2_group_first_block_no(sb, i);
fs/ext2/super.c
705
ext2_fsblk_t last_block = ext2_group_last_block_no(sb, i);
fs/ext2/super.c
71
sb->s_id, function, &vaf);
fs/ext2/super.c
710
ext2_error (sb, "ext2_check_descriptors",
fs/ext2/super.c
719
ext2_error (sb, "ext2_check_descriptors",
fs/ext2/super.c
729
ext2_error (sb, "ext2_check_descriptors",
fs/ext2/super.c
75
if (test_opt(sb, ERRORS_PANIC))
fs/ext2/super.c
77
if (!sb_rdonly(sb) && test_opt(sb, ERRORS_RO)) {
fs/ext2/super.c
78
ext2_msg(sb, KERN_CRIT,
fs/ext2/super.c
80
sb->s_flags |= SB_RDONLY;
fs/ext2/super.c
801
static unsigned long descriptor_loc(struct super_block *sb,
fs/ext2/super.c
805
struct ext2_sb_info *sbi = EXT2_SB(sb);
fs/ext2/super.c
810
if (!EXT2_HAS_INCOMPAT_FEATURE(sb, EXT2_FEATURE_INCOMPAT_META_BG) ||
fs/ext2/super.c
815
return ext2_group_first_block_no(sb, bg) + ext2_bg_has_super(sb, bg);
fs/ext2/super.c
877
static int ext2_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ext2/super.c
906
sb->s_fs_info = sbi;
fs/ext2/super.c
908
sbi->s_daxdev = fs_dax_get_by_bdev(sb->s_bdev, &sbi->s_dax_part_off,
fs/ext2/super.c
921
blocksize = sb_min_blocksize(sb, BLOCK_SIZE);
fs/ext2/super.c
923
ext2_msg(sb, KERN_ERR, "error: unable to set blocksize");
fs/ext2/super.c
938
if (!(bh = sb_bread(sb, logic_sb_block))) {
fs/ext2/super.c
939
ext2_msg(sb, KERN_ERR, "error: unable to read superblock");
fs/ext2/super.c
948
sb->s_magic = le16_to_cpu(es->s_magic);
fs/ext2/super.c
950
if (sb->s_magic != EXT2_SUPER_MAGIC)
fs/ext2/super.c
955
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/ext2/super.c
956
(test_opt(sb, POSIX_ACL) ? SB_POSIXACL : 0);
fs/ext2/super.c
957
sb->s_iflags |= SB_I_CGROUPWB;
fs/ext2/super.c
960
(EXT2_HAS_COMPAT_FEATURE(sb, ~0U) ||
fs/ext2/super.c
961
EXT2_HAS_RO_COMPAT_FEATURE(sb, ~0U) ||
fs/ext2/super.c
962
EXT2_HAS_INCOMPAT_FEATURE(sb, ~0U)))
fs/ext2/super.c
963
ext2_msg(sb, KERN_WARNING,
fs/ext2/super.c
971
features = EXT2_HAS_INCOMPAT_FEATURE(sb, ~EXT2_FEATURE_INCOMPAT_SUPP);
fs/ext2/super.c
973
ext2_msg(sb, KERN_ERR, "error: couldn't mount because of "
fs/ext2/super.c
978
if (!sb_rdonly(sb) && (features = EXT2_HAS_RO_COMPAT_FEATURE(sb, ~EXT2_FEATURE_RO_COMPAT_SUPP))){
fs/ext2/super.c
979
ext2_msg(sb, KERN_ERR, "error: couldn't mount RDWR because of "
fs/ext2/super.c
987
ext2_msg(sb, KERN_ERR,
fs/ext2/super.c
994
if (test_opt(sb, DAX)) {
fs/ext2/super.c
996
ext2_msg(sb, KERN_ERR,
fs/ext2/xattr.c
381
static void ext2_xattr_update_super_block(struct super_block *sb)
fs/ext2/xattr.c
383
if (EXT2_HAS_COMPAT_FEATURE(sb, EXT2_FEATURE_COMPAT_EXT_ATTR))
fs/ext2/xattr.c
386
spin_lock(&EXT2_SB(sb)->s_lock);
fs/ext2/xattr.c
387
ext2_update_dynamic_rev(sb);
fs/ext2/xattr.c
388
EXT2_SET_COMPAT_FEATURE(sb, EXT2_FEATURE_COMPAT_EXT_ATTR);
fs/ext2/xattr.c
389
spin_unlock(&EXT2_SB(sb)->s_lock);
fs/ext2/xattr.c
390
mark_buffer_dirty(EXT2_SB(sb)->s_sbh);
fs/ext2/xattr.c
409
struct super_block *sb = inode->i_sb;
fs/ext2/xattr.c
413
size_t name_len, free, min_offs = sb->s_blocksize;
fs/ext2/xattr.c
437
if (name_len > 255 || value_len > sb->s_blocksize)
fs/ext2/xattr.c
445
bh = sb_bread(sb, EXT2_I(inode)->i_file_acl);
fs/ext2/xattr.c
456
ext2_error(sb, "ext2_xattr_set",
fs/ext2/xattr.c
469
if (!ext2_xattr_entry_valid(last, end, sb->s_blocksize))
fs/ext2/xattr.c
492
free = sb->s_blocksize -
fs/ext2/xattr.c
551
header = kzalloc(sb->s_blocksize, GFP_KERNEL);
fs/ext2/xattr.c
709
struct super_block *sb = inode->i_sb;
fs/ext2/xattr.c
743
ext2_fsblk_t goal = ext2_group_first_block_no(sb,
fs/ext2/xattr.c
753
new_bh = sb_getblk(sb, block);
fs/ext2/xattr.c
766
ext2_xattr_update_super_block(sb);
fs/ext4/balloc.c
103
base_clusters = ext4_num_base_meta_clusters(sb, block_group);
fs/ext4/balloc.c
112
itbl_blk_start = ext4_inode_table(sb, gdp);
fs/ext4/balloc.c
136
if (ext4_block_in_group(sb, ext4_block_bitmap(sb, gdp), block_group)) {
fs/ext4/balloc.c
138
ext4_block_bitmap(sb, gdp) - start);
fs/ext4/balloc.c
145
if (ext4_block_in_group(sb, ext4_inode_bitmap(sb, gdp), block_group)) {
fs/ext4/balloc.c
147
ext4_inode_bitmap(sb, gdp) - start);
fs/ext4/balloc.c
162
static unsigned int num_clusters_in_group(struct super_block *sb,
fs/ext4/balloc.c
167
if (block_group == ext4_get_groups_count(sb) - 1) {
fs/ext4/balloc.c
174
blocks = ext4_blocks_count(EXT4_SB(sb)->s_es) -
fs/ext4/balloc.c
175
ext4_group_first_block_no(sb, block_group);
fs/ext4/balloc.c
177
blocks = EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/balloc.c
178
return EXT4_NUM_B2C(EXT4_SB(sb), blocks);
fs/ext4/balloc.c
182
static int ext4_init_block_bitmap(struct super_block *sb,
fs/ext4/balloc.c
188
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
193
if (!ext4_group_desc_csum_verify(sb, block_group, gdp)) {
fs/ext4/balloc.c
194
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
199
memset(bh->b_data, 0, sb->s_blocksize);
fs/ext4/balloc.c
201
bit_max = ext4_num_base_meta_clusters(sb, block_group);
fs/ext4/balloc.c
208
start = ext4_group_first_block_no(sb, block_group);
fs/ext4/balloc.c
211
tmp = ext4_block_bitmap(sb, gdp);
fs/ext4/balloc.c
212
if (ext4_block_in_group(sb, tmp, block_group))
fs/ext4/balloc.c
215
tmp = ext4_inode_bitmap(sb, gdp);
fs/ext4/balloc.c
216
if (ext4_block_in_group(sb, tmp, block_group))
fs/ext4/balloc.c
219
tmp = ext4_inode_table(sb, gdp);
fs/ext4/balloc.c
220
for (; tmp < ext4_inode_table(sb, gdp) +
fs/ext4/balloc.c
222
if (ext4_block_in_group(sb, tmp, block_group))
fs/ext4/balloc.c
231
ext4_mark_bitmap_end(num_clusters_in_group(sb, block_group),
fs/ext4/balloc.c
232
sb->s_blocksize * 8, bh->b_data);
fs/ext4/balloc.c
239
unsigned ext4_free_clusters_after_init(struct super_block *sb,
fs/ext4/balloc.c
243
return num_clusters_in_group(sb, block_group) -
fs/ext4/balloc.c
244
ext4_num_overhead_clusters(sb, block_group, gdp);
fs/ext4/balloc.c
265
struct ext4_group_desc * ext4_get_group_desc(struct super_block *sb,
fs/ext4/balloc.c
27
static unsigned ext4_num_base_meta_clusters(struct super_block *sb,
fs/ext4/balloc.c
271
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/balloc.c
273
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
277
sb, block_group, bh);
fs/ext4/balloc.c
280
ext4_error(sb, "block_group >= groups_count - block_group = %u,"
fs/ext4/balloc.c
286
group_desc = block_group >> EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/balloc.c
287
offset = block_group & (EXT4_DESC_PER_BLOCK(sb) - 1);
fs/ext4/balloc.c
296
ext4_error(sb, "Group descriptor not loaded - "
fs/ext4/balloc.c
304
offset * EXT4_DESC_SIZE(sb));
fs/ext4/balloc.c
310
static ext4_fsblk_t ext4_valid_block_bitmap_padding(struct super_block *sb,
fs/ext4/balloc.c
315
unsigned long bitmap_size = sb->s_blocksize * 8;
fs/ext4/balloc.c
316
unsigned int offset = num_clusters_in_group(sb, block_group);
fs/ext4/balloc.c
326
struct ext4_group_info *ext4_get_group_info(struct super_block *sb,
fs/ext4/balloc.c
332
if (unlikely(group >= EXT4_SB(sb)->s_groups_count))
fs/ext4/balloc.c
334
indexv = group >> (EXT4_DESC_PER_BLOCK_BITS(sb));
fs/ext4/balloc.c
335
indexh = group & ((EXT4_DESC_PER_BLOCK(sb)) - 1);
fs/ext4/balloc.c
336
grp_info = sbi_array_rcu_deref(EXT4_SB(sb), s_group_info, indexv);
fs/ext4/balloc.c
344
static ext4_fsblk_t ext4_valid_block_bitmap(struct super_block *sb,
fs/ext4/balloc.c
349
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
352
ext4_grpblk_t max_bit = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/balloc.c
356
if (ext4_has_feature_flex_bg(sb)) {
fs/ext4/balloc.c
36
ext4_group_t ext4_get_group_number(struct super_block *sb,
fs/ext4/balloc.c
365
group_first_block = ext4_group_first_block_no(sb, block_group);
fs/ext4/balloc.c
368
blk = ext4_block_bitmap(sb, desc);
fs/ext4/balloc.c
376
blk = ext4_inode_bitmap(sb, desc);
fs/ext4/balloc.c
384
blk = ext4_inode_table(sb, desc);
fs/ext4/balloc.c
399
static int ext4_validate_block_bitmap(struct super_block *sb,
fs/ext4/balloc.c
407
if (EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY)
fs/ext4/balloc.c
41
if (test_opt2(sb, STD_GROUP_SIZE))
fs/ext4/balloc.c
410
grp = ext4_get_group_info(sb, block_group);
fs/ext4/balloc.c
417
ext4_lock_group(sb, block_group);
fs/ext4/balloc.c
420
if (unlikely(!ext4_block_bitmap_csum_verify(sb, desc, bh) ||
fs/ext4/balloc.c
421
ext4_simulate_fail(sb, EXT4_SIM_BBITMAP_CRC))) {
fs/ext4/balloc.c
422
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
423
ext4_error(sb, "bg %u: bad block bitmap checksum", block_group);
fs/ext4/balloc.c
424
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
428
blk = ext4_valid_block_bitmap(sb, desc, block_group, bh);
fs/ext4/balloc.c
43
le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block)) >>
fs/ext4/balloc.c
430
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
431
ext4_error(sb, "bg %u: block %llu: invalid block bitmap",
fs/ext4/balloc.c
433
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
437
blk = ext4_valid_block_bitmap_padding(sb, block_group, bh);
fs/ext4/balloc.c
439
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
44
(EXT4_BLOCK_SIZE_BITS(sb) + EXT4_CLUSTER_BITS(sb) + 3);
fs/ext4/balloc.c
440
ext4_error(sb, "bg %u: block %llu: padding at end of block bitmap is not set",
fs/ext4/balloc.c
442
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
448
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
46
ext4_get_group_no_and_offset(sb, block, &group, NULL);
fs/ext4/balloc.c
464
ext4_read_block_bitmap_nowait(struct super_block *sb, ext4_group_t block_group,
fs/ext4/balloc.c
468
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
474
sb, block_group, ignore_locked);
fs/ext4/balloc.c
476
desc = ext4_get_group_desc(sb, block_group, NULL);
fs/ext4/balloc.c
479
bitmap_blk = ext4_block_bitmap(sb, desc);
fs/ext4/balloc.c
482
ext4_error(sb, "Invalid block bitmap block %llu in "
fs/ext4/balloc.c
484
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
488
bh = sb_getblk(sb, bitmap_blk);
fs/ext4/balloc.c
490
ext4_warning(sb, "Cannot get buffer for block bitmap - "
fs/ext4/balloc.c
510
ext4_lock_group(sb, block_group);
fs/ext4/balloc.c
511
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/balloc.c
514
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
516
ext4_error(sb, "Block bitmap for bg 0 marked "
fs/ext4/balloc.c
521
err = ext4_init_block_bitmap(sb, bh, block_group, desc);
fs/ext4/balloc.c
523
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
525
ext4_error(sb, "Failed to init block bitmap for group "
fs/ext4/balloc.c
532
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
536
ext4_unlock_group(sb, block_group);
fs/ext4/balloc.c
54
void ext4_get_group_no_and_offset(struct super_block *sb, ext4_fsblk_t blocknr,
fs/ext4/balloc.c
550
trace_ext4_read_block_bitmap_load(sb, block_group, ignore_locked);
fs/ext4/balloc.c
554
ext4_simulate_fail(sb, EXT4_SIM_BBITMAP_EIO));
fs/ext4/balloc.c
557
err = ext4_validate_block_bitmap(sb, desc, block_group, bh);
fs/ext4/balloc.c
567
int ext4_wait_block_bitmap(struct super_block *sb, ext4_group_t block_group,
fs/ext4/balloc.c
57
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/balloc.c
573
sb, block_group, bh);
fs/ext4/balloc.c
577
desc = ext4_get_group_desc(sb, block_group, NULL);
fs/ext4/balloc.c
582
ext4_error_err(sb, EIO, "Cannot read block bitmap - "
fs/ext4/balloc.c
585
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/balloc.c
591
return ext4_validate_block_bitmap(sb, desc, block_group, bh);
fs/ext4/balloc.c
595
ext4_read_block_bitmap(struct super_block *sb, ext4_group_t block_group)
fs/ext4/balloc.c
600
bh = ext4_read_block_bitmap_nowait(sb, block_group, false);
fs/ext4/balloc.c
603
err = ext4_wait_block_bitmap(sb, block_group, bh);
fs/ext4/balloc.c
61
offset = do_div(blocknr, EXT4_BLOCKS_PER_GROUP(sb)) >>
fs/ext4/balloc.c
62
EXT4_SB(sb)->s_cluster_bits;
fs/ext4/balloc.c
689
int ext4_should_retry_alloc(struct super_block *sb, int *retries)
fs/ext4/balloc.c
691
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
707
if (test_opt(sb, DISCARD)) {
fs/ext4/balloc.c
719
ext4_debug("%s: retrying operation after ENOSPC\n", sb->s_id);
fs/ext4/balloc.c
74
static inline int ext4_block_in_group(struct super_block *sb,
fs/ext4/balloc.c
770
ext4_fsblk_t ext4_count_free_clusters(struct super_block *sb)
fs/ext4/balloc.c
775
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/balloc.c
783
es = EXT4_SB(sb)->s_es;
fs/ext4/balloc.c
789
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/balloc.c
793
if (EXT4_SB(sb)->s_group_info)
fs/ext4/balloc.c
794
grp = ext4_get_group_info(sb, i);
fs/ext4/balloc.c
796
desc_count += ext4_free_group_clusters(sb, gdp);
fs/ext4/balloc.c
798
bitmap_bh = ext4_read_block_bitmap(sb, i);
fs/ext4/balloc.c
80
actual_group = ext4_get_group_number(sb, block);
fs/ext4/balloc.c
805
EXT4_CLUSTERS_PER_GROUP(sb) / 8);
fs/ext4/balloc.c
807
i, ext4_free_group_clusters(sb, gdp), x);
fs/ext4/balloc.c
813
EXT4_NUM_B2C(EXT4_SB(sb), ext4_free_blocks_count(es)),
fs/ext4/balloc.c
819
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/balloc.c
823
if (EXT4_SB(sb)->s_group_info)
fs/ext4/balloc.c
824
grp = ext4_get_group_info(sb, i);
fs/ext4/balloc.c
826
desc_count += ext4_free_group_clusters(sb, gdp);
fs/ext4/balloc.c
854
int ext4_bg_has_super(struct super_block *sb, ext4_group_t group)
fs/ext4/balloc.c
856
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/balloc.c
860
if (ext4_has_feature_sparse_super2(sb)) {
fs/ext4/balloc.c
866
if ((group <= 1) || !ext4_has_feature_sparse_super(sb))
fs/ext4/balloc.c
877
static unsigned long ext4_bg_num_gdb_meta(struct super_block *sb,
fs/ext4/balloc.c
88
static unsigned ext4_num_overhead_clusters(struct super_block *sb,
fs/ext4/balloc.c
880
unsigned long metagroup = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/balloc.c
881
ext4_group_t first = metagroup * EXT4_DESC_PER_BLOCK(sb);
fs/ext4/balloc.c
882
ext4_group_t last = first + EXT4_DESC_PER_BLOCK(sb) - 1;
fs/ext4/balloc.c
889
static unsigned long ext4_bg_num_gdb_nometa(struct super_block *sb,
fs/ext4/balloc.c
892
if (!ext4_bg_has_super(sb, group))
fs/ext4/balloc.c
895
if (ext4_has_feature_meta_bg(sb))
fs/ext4/balloc.c
896
return le32_to_cpu(EXT4_SB(sb)->s_es->s_first_meta_bg);
fs/ext4/balloc.c
898
return EXT4_SB(sb)->s_gdb_count;
fs/ext4/balloc.c
910
unsigned long ext4_bg_num_gdb(struct super_block *sb, ext4_group_t group)
fs/ext4/balloc.c
913
le32_to_cpu(EXT4_SB(sb)->s_es->s_first_meta_bg);
fs/ext4/balloc.c
914
unsigned long metagroup = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/balloc.c
916
if (!ext4_has_feature_meta_bg(sb) || metagroup < first_meta_bg)
fs/ext4/balloc.c
917
return ext4_bg_num_gdb_nometa(sb, group);
fs/ext4/balloc.c
919
return ext4_bg_num_gdb_meta(sb,group);
fs/ext4/balloc.c
927
unsigned int ext4_num_base_meta_blocks(struct super_block *sb,
fs/ext4/balloc.c
930
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/balloc.c
934
num = ext4_bg_has_super(sb, block_group);
fs/ext4/balloc.c
936
if (!ext4_has_feature_meta_bg(sb) ||
fs/ext4/balloc.c
940
num += ext4_bg_num_gdb_nometa(sb, block_group);
fs/ext4/balloc.c
944
num += ext4_bg_num_gdb_meta(sb, block_group);
fs/ext4/balloc.c
949
static unsigned int ext4_num_base_meta_clusters(struct super_block *sb,
fs/ext4/balloc.c
95
ext4_fsblk_t start = ext4_group_first_block_no(sb, block_group);
fs/ext4/balloc.c
952
return EXT4_NUM_B2C(EXT4_SB(sb), ext4_num_base_meta_blocks(sb, block_group));
fs/ext4/balloc.c
96
ext4_fsblk_t end = start + EXT4_BLOCKS_PER_GROUP(sb) - 1;
fs/ext4/balloc.c
98
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/bitmap.c
19
int ext4_inode_bitmap_csum_verify(struct super_block *sb,
fs/ext4/bitmap.c
25
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/bitmap.c
28
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/bitmap.c
31
sz = EXT4_INODES_PER_GROUP(sb) >> 3;
fs/ext4/bitmap.c
43
void ext4_inode_bitmap_csum_set(struct super_block *sb,
fs/ext4/bitmap.c
48
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/bitmap.c
51
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/bitmap.c
54
sz = EXT4_INODES_PER_GROUP(sb) >> 3;
fs/ext4/bitmap.c
61
int ext4_block_bitmap_csum_verify(struct super_block *sb,
fs/ext4/bitmap.c
67
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/bitmap.c
68
int sz = EXT4_CLUSTERS_PER_GROUP(sb) / 8;
fs/ext4/bitmap.c
70
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/bitmap.c
84
void ext4_block_bitmap_csum_set(struct super_block *sb,
fs/ext4/bitmap.c
88
int sz = EXT4_CLUSTERS_PER_GROUP(sb) / 8;
fs/ext4/bitmap.c
90
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/bitmap.c
92
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/block_validity.c
147
static int ext4_protect_reserved_inode(struct super_block *sb,
fs/ext4/block_validity.c
152
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/block_validity.c
160
inode = ext4_iget(sb, ino, EXT4_IGET_SPECIAL);
fs/ext4/block_validity.c
163
num = (inode->i_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/ext4/block_validity.c
211
int ext4_setup_system_zone(struct super_block *sb)
fs/ext4/block_validity.c
213
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/block_validity.c
214
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/block_validity.c
225
unsigned int meta_blks = ext4_num_base_meta_blocks(sb, i);
fs/ext4/block_validity.c
230
ext4_group_first_block_no(sb, i),
fs/ext4/block_validity.c
235
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/block_validity.c
237
ext4_block_bitmap(sb, gdp), 1, 0);
fs/ext4/block_validity.c
241
ext4_inode_bitmap(sb, gdp), 1, 0);
fs/ext4/block_validity.c
245
ext4_inode_table(sb, gdp),
fs/ext4/block_validity.c
250
if (ext4_has_feature_journal(sb) && sbi->s_es->s_journal_inum) {
fs/ext4/block_validity.c
251
ret = ext4_protect_reserved_inode(sb, system_blks,
fs/ext4/block_validity.c
264
if (test_opt(sb, DEBUG))
fs/ext4/block_validity.c
283
void ext4_release_system_zone(struct super_block *sb)
fs/ext4/block_validity.c
287
system_blks = rcu_dereference_protected(EXT4_SB(sb)->s_system_blks,
fs/ext4/block_validity.c
288
lockdep_is_held(&sb->s_umount));
fs/ext4/block_validity.c
289
rcu_assign_pointer(EXT4_SB(sb)->s_system_blks, NULL);
fs/ext4/block_validity.c
295
int ext4_sb_block_valid(struct super_block *sb, struct inode *inode,
fs/ext4/block_validity.c
298
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/crypto.c
102
err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh,
fs/ext4/crypto.c
108
ext4_superblock_csum_set(sb);
fs/ext4/crypto.c
226
static const union fscrypt_policy *ext4_get_dummy_policy(struct super_block *sb)
fs/ext4/crypto.c
228
return EXT4_SB(sb)->s_dummy_enc_policy.policy;
fs/ext4/crypto.c
231
static bool ext4_has_stable_inodes(struct super_block *sb)
fs/ext4/crypto.c
233
return ext4_has_feature_stable_inodes(sb);
fs/ext4/crypto.c
85
struct super_block *sb = file_inode(filp)->i_sb;
fs/ext4/crypto.c
86
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/crypto.c
90
if (!ext4_has_feature_encrypt(sb))
fs/ext4/crypto.c
97
handle = ext4_journal_start_sb(sb, EXT4_HT_MISC, 1);
fs/ext4/dir.c
137
struct super_block *sb = inode->i_sb;
fs/ext4/dir.c
152
if (!ext4_has_feature_metadata_csum(sb)) {
fs/ext4/dir.c
183
offset = ctx->pos & (sb->s_blocksize - 1);
fs/ext4/dir.c
184
map.m_lblk = ctx->pos >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/dir.c
192
ctx->pos += map.m_len * sb->s_blocksize;
fs/ext4/dir.c
200
sb->s_bdev->bd_mapping,
fs/ext4/dir.c
202
1 << EXT4_SB(sb)->s_min_folio_order);
fs/ext4/dir.c
216
ctx->pos += sb->s_blocksize - offset;
fs/ext4/dir.c
226
ctx->pos += sb->s_blocksize - offset;
fs/ext4/dir.c
238
for (i = 0; i < sb->s_blocksize && i < offset; ) {
fs/ext4/dir.c
248
sb->s_blocksize) < ext4_dir_rec_len(1,
fs/ext4/dir.c
252
sb->s_blocksize);
fs/ext4/dir.c
255
ctx->pos = (ctx->pos & ~(sb->s_blocksize - 1))
fs/ext4/dir.c
261
&& offset < sb->s_blocksize) {
fs/ext4/dir.c
270
(sb->s_blocksize - 1)) + 1;
fs/ext4/dir.c
274
sb->s_blocksize);
fs/ext4/dir.c
280
get_dtype(sb, de->file_type)))
fs/ext4/dir.c
308
get_dtype(sb, de->file_type)))
fs/ext4/dir.c
313
sb->s_blocksize);
fs/ext4/dir.c
48
struct super_block *sb = inode->i_sb;
fs/ext4/dir.c
52
((inode->i_size >> sb->s_blocksize_bits) == 1) ||
fs/ext4/dir.c
535
struct super_block *sb = inode->i_sb;
fs/ext4/dir.c
538
ext4_msg(sb, KERN_ERR, "%s:%d: inode #%lu: comm %s: "
fs/ext4/dir.c
548
get_dtype(sb, fname->file_type))) {
fs/ext4/ext4.h
1286
#define clear_opt(sb, opt) EXT4_SB(sb)->s_mount_opt &= \
fs/ext4/ext4.h
1288
#define set_opt(sb, opt) EXT4_SB(sb)->s_mount_opt |= \
fs/ext4/ext4.h
1290
#define test_opt(sb, opt) (EXT4_SB(sb)->s_mount_opt & \
fs/ext4/ext4.h
1293
#define clear_opt2(sb, opt) EXT4_SB(sb)->s_mount_opt2 &= \
fs/ext4/ext4.h
1295
#define set_opt2(sb, opt) EXT4_SB(sb)->s_mount_opt2 |= \
fs/ext4/ext4.h
1297
#define test_opt2(sb, opt) (EXT4_SB(sb)->s_mount_opt2 & \
fs/ext4/ext4.h
1484
struct super_block *sb;
fs/ext4/ext4.h
1501
static inline int ext4_inodes_per_orphan_block(struct super_block *sb)
fs/ext4/ext4.h
1503
return (sb->s_blocksize - sizeof(struct ext4_orphan_block_tail)) /
fs/ext4/ext4.h
1803
static inline struct ext4_sb_info *EXT4_SB(struct super_block *sb)
fs/ext4/ext4.h
1805
return sb->s_fs_info;
fs/ext4/ext4.h
1812
static inline int ext4_writepages_down_read(struct super_block *sb)
fs/ext4/ext4.h
1814
percpu_down_read(&EXT4_SB(sb)->s_writepages_rwsem);
fs/ext4/ext4.h
1818
static inline void ext4_writepages_up_read(struct super_block *sb, int ctx)
fs/ext4/ext4.h
1821
percpu_up_read(&EXT4_SB(sb)->s_writepages_rwsem);
fs/ext4/ext4.h
1824
static inline int ext4_writepages_down_write(struct super_block *sb)
fs/ext4/ext4.h
1826
percpu_down_write(&EXT4_SB(sb)->s_writepages_rwsem);
fs/ext4/ext4.h
1830
static inline void ext4_writepages_up_write(struct super_block *sb, int ctx)
fs/ext4/ext4.h
1833
percpu_up_write(&EXT4_SB(sb)->s_writepages_rwsem);
fs/ext4/ext4.h
1836
static inline int ext4_fc_lock(struct super_block *sb)
fs/ext4/ext4.h
1838
mutex_lock(&EXT4_SB(sb)->s_fc_lock);
fs/ext4/ext4.h
1842
static inline void ext4_fc_unlock(struct super_block *sb, int ctx)
fs/ext4/ext4.h
1845
mutex_unlock(&EXT4_SB(sb)->s_fc_lock);
fs/ext4/ext4.h
1848
static inline int ext4_valid_inum(struct super_block *sb, unsigned long ino)
fs/ext4/ext4.h
1851
(ino >= EXT4_FIRST_INO(sb) &&
fs/ext4/ext4.h
1852
ino <= le32_to_cpu(EXT4_SB(sb)->s_es->s_inodes_count));
fs/ext4/ext4.h
1893
static inline void ext4_set_mount_flag(struct super_block *sb, int bit)
fs/ext4/ext4.h
1895
set_bit(bit, &EXT4_SB(sb)->s_mount_flags);
fs/ext4/ext4.h
1898
static inline void ext4_clear_mount_flag(struct super_block *sb, int bit)
fs/ext4/ext4.h
1900
clear_bit(bit, &EXT4_SB(sb)->s_mount_flags);
fs/ext4/ext4.h
1903
static inline int ext4_test_mount_flag(struct super_block *sb, int bit)
fs/ext4/ext4.h
1905
return test_bit(bit, &EXT4_SB(sb)->s_mount_flags);
fs/ext4/ext4.h
1921
static inline bool ext4_simulate_fail(struct super_block *sb,
fs/ext4/ext4.h
1925
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ext4.h
2025
#define EXT4_SB(sb) (sb)
fs/ext4/ext4.h
2130
extern void ext4_update_dynamic_rev(struct super_block *sb);
fs/ext4/ext4.h
2133
static inline bool ext4_has_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2135
return ((EXT4_SB(sb)->s_es->s_feature_compat & \
fs/ext4/ext4.h
2138
static inline void ext4_set_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2140
ext4_update_dynamic_rev(sb); \
fs/ext4/ext4.h
2141
EXT4_SB(sb)->s_es->s_feature_compat |= \
fs/ext4/ext4.h
2144
static inline void ext4_clear_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2146
EXT4_SB(sb)->s_es->s_feature_compat &= \
fs/ext4/ext4.h
2151
static inline bool ext4_has_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2153
return ((EXT4_SB(sb)->s_es->s_feature_ro_compat & \
fs/ext4/ext4.h
2156
static inline void ext4_set_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2158
ext4_update_dynamic_rev(sb); \
fs/ext4/ext4.h
2159
EXT4_SB(sb)->s_es->s_feature_ro_compat |= \
fs/ext4/ext4.h
2162
static inline void ext4_clear_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2164
EXT4_SB(sb)->s_es->s_feature_ro_compat &= \
fs/ext4/ext4.h
2169
static inline bool ext4_has_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2171
return ((EXT4_SB(sb)->s_es->s_feature_incompat & \
fs/ext4/ext4.h
2174
static inline void ext4_set_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2176
ext4_update_dynamic_rev(sb); \
fs/ext4/ext4.h
2177
EXT4_SB(sb)->s_es->s_feature_incompat |= \
fs/ext4/ext4.h
2180
static inline void ext4_clear_feature_##name(struct super_block *sb) \
fs/ext4/ext4.h
2182
EXT4_SB(sb)->s_es->s_feature_incompat &= \
fs/ext4/ext4.h
2274
static inline bool ext4_has_unknown_ext##ver##_compat_features(struct super_block *sb) \
fs/ext4/ext4.h
2276
return ((EXT4_SB(sb)->s_es->s_feature_compat & \
fs/ext4/ext4.h
2279
static inline bool ext4_has_unknown_ext##ver##_ro_compat_features(struct super_block *sb) \
fs/ext4/ext4.h
2281
return ((EXT4_SB(sb)->s_es->s_feature_ro_compat & \
fs/ext4/ext4.h
2284
static inline bool ext4_has_unknown_ext##ver##_incompat_features(struct super_block *sb) \
fs/ext4/ext4.h
2286
return ((EXT4_SB(sb)->s_es->s_feature_incompat & \
fs/ext4/ext4.h
2294
static inline bool ext4_has_compat_features(struct super_block *sb)
fs/ext4/ext4.h
2296
return (EXT4_SB(sb)->s_es->s_feature_compat != 0);
fs/ext4/ext4.h
2298
static inline bool ext4_has_ro_compat_features(struct super_block *sb)
fs/ext4/ext4.h
2300
return (EXT4_SB(sb)->s_es->s_feature_ro_compat != 0);
fs/ext4/ext4.h
2302
static inline bool ext4_has_incompat_features(struct super_block *sb)
fs/ext4/ext4.h
2304
return (EXT4_SB(sb)->s_es->s_feature_incompat != 0);
fs/ext4/ext4.h
2307
extern int ext4_feature_set_ok(struct super_block *sb, int readonly);
fs/ext4/ext4.h
2319
static inline int ext4_forced_shutdown(struct super_block *sb)
fs/ext4/ext4.h
2321
return test_bit(EXT4_FLAGS_SHUTDOWN, &EXT4_SB(sb)->s_ext4_flags);
fs/ext4/ext4.h
2324
static inline int ext4_emergency_ro(struct super_block *sb)
fs/ext4/ext4.h
2326
return test_bit(EXT4_FLAGS_EMERGENCY_RO, &EXT4_SB(sb)->s_ext4_flags);
fs/ext4/ext4.h
2329
static inline int ext4_emergency_state(struct super_block *sb)
fs/ext4/ext4.h
2331
if (unlikely(ext4_forced_shutdown(sb)))
fs/ext4/ext4.h
2333
if (unlikely(ext4_emergency_ro(sb)))
fs/ext4/ext4.h
2633
ext4_group_first_block_no(struct super_block *sb, ext4_group_t group_no)
fs/ext4/ext4.h
2635
return group_no * (ext4_fsblk_t)EXT4_BLOCKS_PER_GROUP(sb) +
fs/ext4/ext4.h
2636
le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block);
fs/ext4/ext4.h
2648
static inline int ext4_dir_htree_level(struct super_block *sb)
fs/ext4/ext4.h
2650
return ext4_has_feature_largedir(sb) ?
fs/ext4/ext4.h
2730
struct super_block *sb; /* super block of the fs */
fs/ext4/ext4.h
2767
void ext4_inode_bitmap_csum_set(struct super_block *sb,
fs/ext4/ext4.h
2770
int ext4_inode_bitmap_csum_verify(struct super_block *sb,
fs/ext4/ext4.h
2773
void ext4_block_bitmap_csum_set(struct super_block *sb,
fs/ext4/ext4.h
2776
int ext4_block_bitmap_csum_verify(struct super_block *sb,
fs/ext4/ext4.h
2781
extern void ext4_get_group_no_and_offset(struct super_block *sb,
fs/ext4/ext4.h
2785
extern ext4_group_t ext4_get_group_number(struct super_block *sb,
fs/ext4/ext4.h
2788
extern int ext4_bg_has_super(struct super_block *sb, ext4_group_t group);
fs/ext4/ext4.h
2789
extern unsigned long ext4_bg_num_gdb(struct super_block *sb,
fs/ext4/ext4.h
2799
extern struct ext4_group_desc * ext4_get_group_desc(struct super_block * sb,
fs/ext4/ext4.h
2802
extern struct ext4_group_info *ext4_get_group_info(struct super_block *sb,
fs/ext4/ext4.h
2804
extern int ext4_should_retry_alloc(struct super_block *sb, int *retries);
fs/ext4/ext4.h
2806
extern struct buffer_head *ext4_read_block_bitmap_nowait(struct super_block *sb,
fs/ext4/ext4.h
2809
extern int ext4_wait_block_bitmap(struct super_block *sb,
fs/ext4/ext4.h
2812
extern struct buffer_head *ext4_read_block_bitmap(struct super_block *sb,
fs/ext4/ext4.h
2814
extern unsigned ext4_free_clusters_after_init(struct super_block *sb,
fs/ext4/ext4.h
2923
static inline unsigned char get_dtype(struct super_block *sb, int filetype)
fs/ext4/ext4.h
2925
if (!ext4_has_feature_filetype(sb) || filetype >= EXT4_FT_MAX)
fs/ext4/ext4.h
2941
extern int ext4_mark_inode_used(struct super_block *sb, int ino);
fs/ext4/ext4.h
2963
extern int ext4_init_inode_table(struct super_block *sb,
fs/ext4/ext4.h
2969
void ext4_fc_init(struct super_block *sb, journal_t *journal);
fs/ext4/ext4.h
2983
void ext4_fc_mark_ineligible(struct super_block *sb, int reason, handle_t *handle);
fs/ext4/ext4.h
2985
bool ext4_fc_replay_check_excluded(struct super_block *sb, ext4_fsblk_t block);
fs/ext4/ext4.h
2986
void ext4_fc_replay_cleanup(struct super_block *sb);
fs/ext4/ext4.h
2990
int ext4_fc_record_regions(struct super_block *sb, int ino,
fs/ext4/ext4.h
3005
extern ext4_group_t ext4_mb_prefetch(struct super_block *sb,
fs/ext4/ext4.h
3008
extern void ext4_mb_prefetch_fini(struct super_block *sb, ext4_group_t group,
fs/ext4/ext4.h
3014
extern int ext4_mb_alloc_groupinfo(struct super_block *sb,
fs/ext4/ext4.h
3016
extern int ext4_mb_add_groupinfo(struct super_block *sb,
fs/ext4/ext4.h
3018
extern int ext4_group_add_blocks(handle_t *handle, struct super_block *sb,
fs/ext4/ext4.h
3021
extern void ext4_process_freed_data(struct super_block *sb, tid_t commit_tid);
fs/ext4/ext4.h
3022
extern void ext4_mb_mark_bb(struct super_block *sb, ext4_fsblk_t block,
fs/ext4/ext4.h
3066
extern struct inode *__ext4_iget(struct super_block *sb, unsigned long ino,
fs/ext4/ext4.h
3070
#define ext4_iget(sb, ino, flags) \
fs/ext4/ext4.h
3071
__ext4_iget((sb), (ino), (flags), __func__, __LINE__)
fs/ext4/ext4.h
3086
extern int ext4_get_fc_inode_loc(struct super_block *sb, unsigned long ino,
fs/ext4/ext4.h
3114
static inline bool is_special_ino(struct super_block *sb, unsigned long ino)
fs/ext4/ext4.h
3116
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/ext4.h
3118
return (ino < EXT4_FIRST_INO(sb) && ino != EXT4_ROOT_INO) ||
fs/ext4/ext4.h
3140
int ext4_update_overhead(struct super_block *sb, bool force);
fs/ext4/ext4.h
3141
int ext4_force_shutdown(struct super_block *sb, u32 flags);
fs/ext4/ext4.h
3171
extern int ext4_group_add(struct super_block *sb,
fs/ext4/ext4.h
3173
extern int ext4_group_extend(struct super_block *sb,
fs/ext4/ext4.h
3176
extern int ext4_resize_fs(struct super_block *sb, ext4_fsblk_t n_blocks_count);
fs/ext4/ext4.h
3177
extern unsigned int ext4_list_backups(struct super_block *sb,
fs/ext4/ext4.h
3182
extern struct buffer_head *ext4_sb_bread(struct super_block *sb,
fs/ext4/ext4.h
3184
extern struct buffer_head *ext4_sb_bread_unmovable(struct super_block *sb,
fs/ext4/ext4.h
3186
extern struct buffer_head *ext4_sb_bread_nofail(struct super_block *sb,
fs/ext4/ext4.h
3193
extern void ext4_sb_breadahead_unmovable(struct super_block *sb, sector_t block);
fs/ext4/ext4.h
3195
extern int ext4_calculate_overhead(struct super_block *sb);
fs/ext4/ext4.h
3197
extern void ext4_superblock_csum_set(struct super_block *sb);
fs/ext4/ext4.h
3198
extern int ext4_alloc_flex_bg_array(struct super_block *sb,
fs/ext4/ext4.h
3200
extern const char *ext4_decode_error(struct super_block *sb, int errno,
fs/ext4/ext4.h
3202
extern void ext4_mark_group_bitmap_corrupted(struct super_block *sb,
fs/ext4/ext4.h
3205
extern unsigned int ext4_num_base_meta_blocks(struct super_block *sb,
fs/ext4/ext4.h
3249
#define ext4_abort(sb, err, fmt, a...) \
fs/ext4/ext4.h
3250
__ext4_error((sb), __func__, __LINE__, true, (err), 0, (fmt), ## a)
fs/ext4/ext4.h
3261
#define ext4_error(sb, fmt, ...) \
fs/ext4/ext4.h
3262
__ext4_error((sb), __func__, __LINE__, false, 0, 0, (fmt), \
fs/ext4/ext4.h
3264
#define ext4_error_err(sb, err, fmt, ...) \
fs/ext4/ext4.h
3265
__ext4_error((sb), __func__, __LINE__, false, (err), 0, (fmt), \
fs/ext4/ext4.h
3267
#define ext4_warning(sb, fmt, ...) \
fs/ext4/ext4.h
3268
__ext4_warning(sb, __func__, __LINE__, fmt, ##__VA_ARGS__)
fs/ext4/ext4.h
3271
#define ext4_msg(sb, level, fmt, ...) \
fs/ext4/ext4.h
3272
__ext4_msg(sb, level, fmt, ##__VA_ARGS__)
fs/ext4/ext4.h
3273
#define dump_mmp_msg(sb, mmp, msg) \
fs/ext4/ext4.h
3274
__dump_mmp_msg(sb, mmp, __func__, __LINE__, msg)
fs/ext4/ext4.h
3275
#define ext4_grp_locked_error(sb, grp, ino, block, fmt, ...) \
fs/ext4/ext4.h
3276
__ext4_grp_locked_error(__func__, __LINE__, sb, grp, ino, block, \
fs/ext4/ext4.h
3296
#define ext4_error(sb, fmt, ...) \
fs/ext4/ext4.h
3299
__ext4_error(sb, "", 0, false, 0, 0, " "); \
fs/ext4/ext4.h
3301
#define ext4_error_err(sb, err, fmt, ...) \
fs/ext4/ext4.h
3304
__ext4_error(sb, "", 0, false, err, 0, " "); \
fs/ext4/ext4.h
3306
#define ext4_warning(sb, fmt, ...) \
fs/ext4/ext4.h
3309
__ext4_warning(sb, "", 0, " "); \
fs/ext4/ext4.h
3316
#define ext4_msg(sb, level, fmt, ...) \
fs/ext4/ext4.h
3319
__ext4_msg(sb, "", " "); \
fs/ext4/ext4.h
3321
#define dump_mmp_msg(sb, mmp, msg) \
fs/ext4/ext4.h
3322
__dump_mmp_msg(sb, mmp, "", 0, "")
fs/ext4/ext4.h
3323
#define ext4_grp_locked_error(sb, grp, ino, block, fmt, ...) \
fs/ext4/ext4.h
3326
__ext4_grp_locked_error("", 0, sb, grp, ino, block, " "); \
fs/ext4/ext4.h
3331
extern ext4_fsblk_t ext4_block_bitmap(struct super_block *sb,
fs/ext4/ext4.h
3333
extern ext4_fsblk_t ext4_inode_bitmap(struct super_block *sb,
fs/ext4/ext4.h
3335
extern ext4_fsblk_t ext4_inode_table(struct super_block *sb,
fs/ext4/ext4.h
3337
extern __u32 ext4_free_group_clusters(struct super_block *sb,
fs/ext4/ext4.h
3339
extern __u32 ext4_free_inodes_count(struct super_block *sb,
fs/ext4/ext4.h
3341
extern __u32 ext4_used_dirs_count(struct super_block *sb,
fs/ext4/ext4.h
3343
extern __u32 ext4_itable_unused_count(struct super_block *sb,
fs/ext4/ext4.h
3345
extern void ext4_block_bitmap_set(struct super_block *sb,
fs/ext4/ext4.h
3347
extern void ext4_inode_bitmap_set(struct super_block *sb,
fs/ext4/ext4.h
3349
extern void ext4_inode_table_set(struct super_block *sb,
fs/ext4/ext4.h
3351
extern void ext4_free_group_clusters_set(struct super_block *sb,
fs/ext4/ext4.h
3354
extern void ext4_free_inodes_set(struct super_block *sb,
fs/ext4/ext4.h
3356
extern void ext4_used_dirs_set(struct super_block *sb,
fs/ext4/ext4.h
3358
extern void ext4_itable_unused_set(struct super_block *sb,
fs/ext4/ext4.h
3360
extern int ext4_group_desc_csum_verify(struct super_block *sb, __u32 group,
fs/ext4/ext4.h
3362
extern void ext4_group_desc_csum_set(struct super_block *sb, __u32 group,
fs/ext4/ext4.h
3364
extern int ext4_register_li_request(struct super_block *sb,
fs/ext4/ext4.h
3367
static inline int ext4_has_group_desc_csum(struct super_block *sb)
fs/ext4/ext4.h
3369
return ext4_has_feature_gdt_csum(sb) ||
fs/ext4/ext4.h
3370
ext4_has_feature_metadata_csum(sb);
fs/ext4/ext4.h
3414
static inline loff_t ext4_isize(struct super_block *sb,
fs/ext4/ext4.h
3417
if (ext4_has_feature_largedir(sb) ||
fs/ext4/ext4.h
3436
static inline ext4_group_t ext4_get_groups_count(struct super_block *sb)
fs/ext4/ext4.h
3438
ext4_group_t ngroups = EXT4_SB(sb)->s_groups_count;
fs/ext4/ext4.h
3462
#define ext4_std_error(sb, errno) \
fs/ext4/ext4.h
3465
__ext4_std_error((sb), __func__, __LINE__, (errno)); \
fs/ext4/ext4.h
3561
static inline spinlock_t *ext4_group_lock_ptr(struct super_block *sb,
fs/ext4/ext4.h
3564
return bgl_lock_ptr(EXT4_SB(sb)->s_blockgroup_lock, group);
fs/ext4/ext4.h
3576
static inline bool ext4_try_lock_group(struct super_block *sb, ext4_group_t group)
fs/ext4/ext4.h
3578
if (!spin_trylock(ext4_group_lock_ptr(sb, group)))
fs/ext4/ext4.h
3584
atomic_add_unless(&EXT4_SB(sb)->s_lock_busy, -1, 0);
fs/ext4/ext4.h
3588
static inline void ext4_lock_group(struct super_block *sb, ext4_group_t group)
fs/ext4/ext4.h
3590
if (!ext4_try_lock_group(sb, group)) {
fs/ext4/ext4.h
3595
atomic_add_unless(&EXT4_SB(sb)->s_lock_busy, 1,
fs/ext4/ext4.h
3597
spin_lock(ext4_group_lock_ptr(sb, group));
fs/ext4/ext4.h
3601
static inline void ext4_unlock_group(struct super_block *sb,
fs/ext4/ext4.h
3604
spin_unlock(ext4_group_lock_ptr(sb, group));
fs/ext4/ext4.h
3608
static inline bool ext4_quota_capable(struct super_block *sb)
fs/ext4/ext4.h
3610
return (test_opt(sb, QUOTA) || ext4_has_feature_quota(sb));
fs/ext4/ext4.h
3613
static inline bool ext4_is_quota_journalled(struct super_block *sb)
fs/ext4/ext4.h
3615
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ext4.h
3617
return (ext4_has_feature_quota(sb) ||
fs/ext4/ext4.h
3620
int ext4_enable_quotas(struct super_block *sb);
fs/ext4/ext4.h
3736
static inline void ext4_set_de_type(struct super_block *sb,
fs/ext4/ext4.h
3739
if (ext4_has_feature_filetype(sb))
fs/ext4/ext4.h
3756
extern int ext4_register_sysfs(struct super_block *sb);
fs/ext4/ext4.h
3757
extern void ext4_unregister_sysfs(struct super_block *sb);
fs/ext4/ext4.h
3762
extern void ext4_release_system_zone(struct super_block *sb);
fs/ext4/ext4.h
3763
extern int ext4_setup_system_zone(struct super_block *sb);
fs/ext4/ext4.h
3771
extern int ext4_sb_block_valid(struct super_block *sb, struct inode *inode,
fs/ext4/ext4.h
3879
extern void ext4_orphan_cleanup(struct super_block *sb,
fs/ext4/ext4.h
3881
extern void ext4_release_orphan_info(struct super_block *sb);
fs/ext4/ext4.h
3882
extern int ext4_init_orphan_info(struct super_block *sb);
fs/ext4/ext4.h
3883
extern int ext4_orphan_file_empty(struct super_block *sb);
fs/ext4/ext4.h
3906
extern int ext4_resize_begin(struct super_block *sb);
fs/ext4/ext4.h
3907
extern int ext4_resize_end(struct super_block *sb, bool update_backups);
fs/ext4/ext4_jbd2.c
104
trace_ext4_journal_start_sb(sb, blocks, rsv_blocks,
fs/ext4/ext4_jbd2.c
107
err = ext4_journal_check_start(sb);
fs/ext4/ext4_jbd2.c
111
journal = EXT4_SB(sb)->s_journal;
fs/ext4/ext4_jbd2.c
112
if (!journal || (EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY))
fs/ext4/ext4_jbd2.c
120
struct super_block *sb;
fs/ext4/ext4_jbd2.c
135
sb = handle->h_transaction->t_journal->j_private;
fs/ext4/ext4_jbd2.c
141
__ext4_std_error(sb, where, line, err);
fs/ext4/ext4_jbd2.c
148
struct super_block *sb;
fs/ext4/ext4_jbd2.c
154
sb = handle->h_journal->j_private;
fs/ext4/ext4_jbd2.c
155
trace_ext4_journal_start_reserved(sb,
fs/ext4/ext4_jbd2.c
157
err = ext4_journal_check_start(sb);
fs/ext4/ext4_jbd2.c
209
static void ext4_check_bdev_write_error(struct super_block *sb)
fs/ext4/ext4_jbd2.c
211
struct address_space *mapping = sb->s_bdev->bd_mapping;
fs/ext4/ext4_jbd2.c
212
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ext4_jbd2.c
226
ext4_error_err(sb, -err,
fs/ext4/ext4_jbd2.c
232
handle_t *handle, struct super_block *sb,
fs/ext4/ext4_jbd2.c
248
ext4_check_bdev_write_error(sb);
fs/ext4/ext4_jbd2.c
250
!ext4_has_feature_metadata_csum(sb))
fs/ext4/ext4_jbd2.c
254
&EXT4_SB(sb)->s_journal_triggers[trigger_type].tr_triggers);
fs/ext4/ext4_jbd2.c
329
handle_t *handle, struct super_block *sb,
fs/ext4/ext4_jbd2.c
345
!ext4_has_feature_metadata_csum(sb))
fs/ext4/ext4_jbd2.c
349
&EXT4_SB(sb)->s_journal_triggers[trigger_type].tr_triggers);
fs/ext4/ext4_jbd2.c
64
static int ext4_journal_check_start(struct super_block *sb)
fs/ext4/ext4_jbd2.c
71
ret = ext4_emergency_state(sb);
fs/ext4/ext4_jbd2.c
75
if (WARN_ON_ONCE(sb_rdonly(sb)))
fs/ext4/ext4_jbd2.c
78
WARN_ON(sb->s_writers.frozen == SB_FREEZE_COMPLETE);
fs/ext4/ext4_jbd2.c
79
journal = EXT4_SB(sb)->s_journal;
fs/ext4/ext4_jbd2.c
86
ext4_abort(sb, -journal->j_errno, "Detected aborted journal");
fs/ext4/ext4_jbd2.c
93
struct super_block *sb, unsigned int line,
fs/ext4/ext4_jbd2.h
100
#define EXT4_QUOTA_TRANS_BLOCKS(sb) 0
fs/ext4/ext4_jbd2.h
101
#define EXT4_QUOTA_INIT_BLOCKS(sb) 0
fs/ext4/ext4_jbd2.h
102
#define EXT4_QUOTA_DEL_BLOCKS(sb) 0
fs/ext4/ext4_jbd2.h
104
#define EXT4_MAXQUOTAS_TRANS_BLOCKS(sb) (EXT4_MAXQUOTAS*EXT4_QUOTA_TRANS_BLOCKS(sb))
fs/ext4/ext4_jbd2.h
105
#define EXT4_MAXQUOTAS_INIT_BLOCKS(sb) (EXT4_MAXQUOTAS*EXT4_QUOTA_INIT_BLOCKS(sb))
fs/ext4/ext4_jbd2.h
106
#define EXT4_MAXQUOTAS_DEL_BLOCKS(sb) (EXT4_MAXQUOTAS*EXT4_QUOTA_DEL_BLOCKS(sb))
fs/ext4/ext4_jbd2.h
150
handle_t *handle, struct super_block *sb,
fs/ext4/ext4_jbd2.h
159
handle_t *handle, struct super_block *sb,
fs/ext4/ext4_jbd2.h
167
#define ext4_journal_get_write_access(handle, sb, bh, trigger_type) \
fs/ext4/ext4_jbd2.h
168
__ext4_journal_get_write_access(__func__, __LINE__, (handle), (sb), \
fs/ext4/ext4_jbd2.h
173
#define ext4_journal_get_create_access(handle, sb, bh, trigger_type) \
fs/ext4/ext4_jbd2.h
174
__ext4_journal_get_create_access(__func__, __LINE__, (handle), (sb), \
fs/ext4/ext4_jbd2.h
180
handle_t *__ext4_journal_start_sb(struct inode *inode, struct super_block *sb,
fs/ext4/ext4_jbd2.h
209
static inline int ext4_free_metadata_revoke_credits(struct super_block *sb,
fs/ext4/ext4_jbd2.h
213
return blocks * EXT4_SB(sb)->s_cluster_ratio;
fs/ext4/ext4_jbd2.h
216
static inline int ext4_trans_default_revoke_credits(struct super_block *sb)
fs/ext4/ext4_jbd2.h
218
return ext4_free_metadata_revoke_credits(sb, 8);
fs/ext4/ext4_jbd2.h
221
#define ext4_journal_start_sb(sb, type, nblocks) \
fs/ext4/ext4_jbd2.h
222
__ext4_journal_start_sb(NULL, (sb), __LINE__, (type), (nblocks), 0,\
fs/ext4/ext4_jbd2.h
223
ext4_trans_default_revoke_credits(sb))
fs/ext4/ext4_jbd2.h
33
#define EXT4_SINGLEDATA_TRANS_BLOCKS(sb) \
fs/ext4/ext4_jbd2.h
34
(ext4_has_feature_extents(sb) ? 20U : 8U)
fs/ext4/ext4_jbd2.h
368
int ext4_force_commit(struct super_block *sb);
fs/ext4/ext4_jbd2.h
48
#define EXT4_DATA_TRANS_BLOCKS(sb) (EXT4_SINGLEDATA_TRANS_BLOCKS(sb) + \
fs/ext4/ext4_jbd2.h
50
EXT4_MAXQUOTAS_TRANS_BLOCKS(sb))
fs/ext4/ext4_jbd2.h
57
#define EXT4_META_TRANS_BLOCKS(sb) (EXT4_XATTR_TRANS_BLOCKS + \
fs/ext4/ext4_jbd2.h
58
EXT4_MAXQUOTAS_TRANS_BLOCKS(sb))
fs/ext4/ext4_jbd2.h
89
#define EXT4_QUOTA_TRANS_BLOCKS(sb) ((ext4_quota_capable(sb)) ? 1 : 0)
fs/ext4/ext4_jbd2.h
92
#define EXT4_QUOTA_INIT_BLOCKS(sb) ((ext4_quota_capable(sb)) ?\
fs/ext4/ext4_jbd2.h
93
(DQUOT_INIT_ALLOC*(EXT4_SINGLEDATA_TRANS_BLOCKS(sb)-3)\
fs/ext4/ext4_jbd2.h
96
#define EXT4_QUOTA_DEL_BLOCKS(sb) ((ext4_quota_capable(sb)) ?\
fs/ext4/ext4_jbd2.h
97
(DQUOT_DEL_ALLOC*(EXT4_SINGLEDATA_TRANS_BLOCKS(sb)-3)\
fs/ext4/extents-test.c
128
static void ext_kill_sb(struct super_block *sb)
fs/ext4/extents-test.c
130
generic_shutdown_super(sb);
fs/ext4/extents-test.c
133
static int ext_set(struct super_block *sb, void *data)
fs/ext4/extents-test.c
145
struct super_block *sb = k_ctx.k_ei->vfs_inode.i_sb;
fs/ext4/extents-test.c
146
struct ext4_sb_info *sbi = sb->s_fs_info;
fs/ext4/extents-test.c
221
struct super_block *sb;
fs/ext4/extents-test.c
227
sb = sget(&ext_fs_type, NULL, ext_set, 0, NULL);
fs/ext4/extents-test.c
228
if (IS_ERR(sb))
fs/ext4/extents-test.c
229
return PTR_ERR(sb);
fs/ext4/extents-test.c
231
sb->s_blocksize = 4096;
fs/ext4/extents-test.c
232
sb->s_blocksize_bits = 12;
fs/ext4/extents-test.c
238
sbi->s_sb = sb;
fs/ext4/extents-test.c
239
sb->s_fs_info = sbi;
fs/ext4/extents-test.c
263
<< sb->s_blocksize_bits;
fs/ext4/extents-test.c
266
inode->i_sb = sb;
fs/ext4/extents.c
3096
void ext4_ext_init(struct super_block *sb)
fs/ext4/extents.c
3102
if (ext4_has_feature_extents(sb)) {
fs/ext4/extents.c
3117
spin_lock_init(&EXT4_SB(sb)->s_ext_stats_lock);
fs/ext4/extents.c
3118
EXT4_SB(sb)->s_ext_min = 1 << 30;
fs/ext4/extents.c
3119
EXT4_SB(sb)->s_ext_max = 0;
fs/ext4/extents.c
3127
void ext4_ext_release(struct super_block *sb)
fs/ext4/extents.c
3129
if (!ext4_has_feature_extents(sb))
fs/ext4/extents.c
3133
if (EXT4_SB(sb)->s_ext_blocks && EXT4_SB(sb)->s_ext_extents) {
fs/ext4/extents.c
3134
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/extents.c
4123
static int get_implied_cluster_alloc(struct super_block *sb,
fs/ext4/extents.c
4128
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/extents.c
4176
trace_ext4_get_implied_cluster_alloc_exit(sb, map, 1);
fs/ext4/extents.c
4180
trace_ext4_get_implied_cluster_alloc_exit(sb, map, 0);
fs/ext4/extents.c
4545
struct super_block *sb = inode->i_sb;
fs/ext4/extents.c
4561
last_block = (inode->i_size + sb->s_blocksize - 1)
fs/ext4/extents.c
4562
>> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/extents.c
5519
struct super_block *sb = inode->i_sb;
fs/ext4/extents.c
5535
if (!IS_ALIGNED(offset | len, EXT4_CLUSTER_SIZE(sb)))
fs/ext4/extents.c
5566
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_FALLOC_RANGE, handle);
fs/ext4/extents.c
5620
struct super_block *sb = inode->i_sb;
fs/ext4/extents.c
5637
if (!IS_ALIGNED(offset | len, EXT4_CLUSTER_SIZE(sb)))
fs/ext4/extents.c
5662
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_FALLOC_RANGE, handle);
fs/ext4/fast_commit.c
1007
if (!ext4_fc_add_dentry_tlv(sb, crc, fc_dentry))
fs/ext4/fast_commit.c
1036
if (!ext4_fc_add_dentry_tlv(sb, crc, fc_dentry))
fs/ext4/fast_commit.c
1044
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
1045
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
1059
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
1064
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
1074
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
1091
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
1108
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
1113
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
1124
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
1134
if (!ext4_fc_add_tlv(sb, EXT4_FC_TAG_HEAD, sizeof(head),
fs/ext4/fast_commit.c
1160
ret = ext4_fc_write_tail(sb, crc);
fs/ext4/fast_commit.c
1163
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
1168
static void ext4_fc_update_stats(struct super_block *sb, int status,
fs/ext4/fast_commit.c
1171
struct ext4_fc_stats *stats = &EXT4_SB(sb)->s_fc_stats;
fs/ext4/fast_commit.c
1192
trace_ext4_fc_commit_stop(sb, nblks, status, commit_tid);
fs/ext4/fast_commit.c
1203
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
1204
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
1211
if (!test_opt2(sb, JOURNAL_FAST_COMMIT))
fs/ext4/fast_commit.c
1214
trace_ext4_fc_commit_start(sb, commit_tid);
fs/ext4/fast_commit.c
1226
ext4_fc_update_stats(sb, EXT4_FC_STATUS_SKIPPED, 0, 0,
fs/ext4/fast_commit.c
1234
ext4_fc_update_stats(sb, EXT4_FC_STATUS_FAILED, 0, 0,
fs/ext4/fast_commit.c
1243
if (ext4_test_mount_flag(sb, EXT4_MF_FC_INELIGIBLE)) {
fs/ext4/fast_commit.c
1277
ext4_fc_update_stats(sb, status, commit_time, nblks, commit_tid);
fs/ext4/fast_commit.c
1283
ext4_fc_update_stats(sb, status, 0, 0, commit_tid);
fs/ext4/fast_commit.c
1293
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
1294
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
1305
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
1359
ext4_clear_mount_flag(sb, EXT4_MF_FC_INELIGIBLE);
fs/ext4/fast_commit.c
1364
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
1365
trace_ext4_fc_stats(sb);
fs/ext4/fast_commit.c
1405
static int ext4_fc_replay_unlink(struct super_block *sb,
fs/ext4/fast_commit.c
1415
trace_ext4_fc_replay(sb, EXT4_FC_TAG_UNLINK, darg.ino,
fs/ext4/fast_commit.c
1420
inode = ext4_iget(sb, darg.ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1427
old_parent = ext4_iget(sb, darg.parent_ino,
fs/ext4/fast_commit.c
1444
static int ext4_fc_replay_link_internal(struct super_block *sb,
fs/ext4/fast_commit.c
1453
dir = ext4_iget(sb, darg->parent_ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1503
static int ext4_fc_replay_link(struct super_block *sb,
fs/ext4/fast_commit.c
1511
trace_ext4_fc_replay(sb, EXT4_FC_TAG_LINK, darg.ino,
fs/ext4/fast_commit.c
1514
inode = ext4_iget(sb, darg.ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1520
ret = ext4_fc_replay_link_internal(sb, &darg, inode);
fs/ext4/fast_commit.c
1529
static int ext4_fc_record_modified_inode(struct super_block *sb, int ino)
fs/ext4/fast_commit.c
1534
state = &EXT4_SB(sb)->s_fc_replay_state;
fs/ext4/fast_commit.c
1558
static int ext4_fc_replay_inode(struct super_block *sb,
fs/ext4/fast_commit.c
1573
trace_ext4_fc_replay(sb, tag, ino, 0, 0);
fs/ext4/fast_commit.c
1575
inode = ext4_iget(sb, ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1582
ret = ext4_fc_record_modified_inode(sb, ino);
fs/ext4/fast_commit.c
1588
ret = ext4_get_fc_inode_loc(sb, ino, &iloc);
fs/ext4/fast_commit.c
1620
ret = ext4_mark_inode_used(sb, ino);
fs/ext4/fast_commit.c
1625
inode = ext4_iget(sb, ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1652
blkdev_issue_flush(sb->s_bdev);
fs/ext4/fast_commit.c
1664
static int ext4_fc_replay_create(struct super_block *sb,
fs/ext4/fast_commit.c
1674
trace_ext4_fc_replay(sb, EXT4_FC_TAG_CREAT, darg.ino,
fs/ext4/fast_commit.c
1678
ret = ext4_mark_inode_used(sb, darg.ino);
fs/ext4/fast_commit.c
1682
inode = ext4_iget(sb, darg.ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1695
dir = ext4_iget(sb, darg.parent_ino, EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1707
ret = ext4_fc_replay_link_internal(sb, &darg, inode);
fs/ext4/fast_commit.c
1722
int ext4_fc_record_regions(struct super_block *sb, int ino,
fs/ext4/fast_commit.c
1728
state = &EXT4_SB(sb)->s_fc_replay_state;
fs/ext4/fast_commit.c
1762
static int ext4_fc_replay_add_range(struct super_block *sb,
fs/ext4/fast_commit.c
1778
trace_ext4_fc_replay(sb, EXT4_FC_TAG_ADD_RANGE,
fs/ext4/fast_commit.c
1782
inode = ext4_iget(sb, le32_to_cpu(fc_add_ex.fc_ino), EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1788
ret = ext4_fc_record_modified_inode(sb, inode->i_ino);
fs/ext4/fast_commit.c
1874
sb->s_blocksize_bits);
fs/ext4/fast_commit.c
1883
ext4_fc_replay_del_range(struct super_block *sb,
fs/ext4/fast_commit.c
1896
trace_ext4_fc_replay(sb, EXT4_FC_TAG_DEL_RANGE,
fs/ext4/fast_commit.c
1899
inode = ext4_iget(sb, le32_to_cpu(lrange.fc_ino), EXT4_IGET_NORMAL);
fs/ext4/fast_commit.c
1905
ret = ext4_fc_record_modified_inode(sb, inode->i_ino);
fs/ext4/fast_commit.c
1937
i_size_read(inode) >> sb->s_blocksize_bits);
fs/ext4/fast_commit.c
1944
static void ext4_fc_set_bitmaps_and_counters(struct super_block *sb)
fs/ext4/fast_commit.c
1953
state = &EXT4_SB(sb)->s_fc_replay_state;
fs/ext4/fast_commit.c
1955
inode = ext4_iget(sb, state->fc_modified_inodes[i],
fs/ext4/fast_commit.c
2003
bool ext4_fc_replay_check_excluded(struct super_block *sb, ext4_fsblk_t blk)
fs/ext4/fast_commit.c
2008
state = &EXT4_SB(sb)->s_fc_replay_state;
fs/ext4/fast_commit.c
2021
void ext4_fc_replay_cleanup(struct super_block *sb)
fs/ext4/fast_commit.c
2023
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
2078
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
2079
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
2129
ret = ext4_fc_record_regions(sb,
fs/ext4/fast_commit.c
2189
trace_ext4_fc_replay_scan(sb, ret, off);
fs/ext4/fast_commit.c
2200
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
2201
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
221
static bool ext4_fc_disabled(struct super_block *sb)
fs/ext4/fast_commit.c
2219
ext4_fc_set_bitmaps_and_counters(sb);
fs/ext4/fast_commit.c
223
return (!test_opt2(sb, JOURNAL_FAST_COMMIT) ||
fs/ext4/fast_commit.c
224
(EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY));
fs/ext4/fast_commit.c
2240
ext4_fc_set_bitmaps_and_counters(sb);
fs/ext4/fast_commit.c
2248
ret = ext4_fc_replay_link(sb, &tl, val);
fs/ext4/fast_commit.c
2251
ret = ext4_fc_replay_unlink(sb, &tl, val);
fs/ext4/fast_commit.c
2254
ret = ext4_fc_replay_add_range(sb, &tl, val);
fs/ext4/fast_commit.c
2257
ret = ext4_fc_replay_create(sb, &tl, val);
fs/ext4/fast_commit.c
2260
ret = ext4_fc_replay_del_range(sb, &tl, val);
fs/ext4/fast_commit.c
2263
ret = ext4_fc_replay_inode(sb, &tl, val);
fs/ext4/fast_commit.c
2266
trace_ext4_fc_replay(sb, EXT4_FC_TAG_PAD, 0,
fs/ext4/fast_commit.c
2270
trace_ext4_fc_replay(sb, EXT4_FC_TAG_TAIL,
fs/ext4/fast_commit.c
2278
trace_ext4_fc_replay(sb, tl.fc_tag, 0, tl.fc_len, 0);
fs/ext4/fast_commit.c
2289
void ext4_fc_init(struct super_block *sb, journal_t *journal)
fs/ext4/fast_commit.c
2297
if (!test_opt2(sb, JOURNAL_FAST_COMMIT))
fs/ext4/fast_commit.c
312
void ext4_fc_mark_ineligible(struct super_block *sb, int reason, handle_t *handle)
fs/ext4/fast_commit.c
314
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
320
if (ext4_fc_disabled(sb))
fs/ext4/fast_commit.c
333
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
334
is_ineligible = ext4_test_mount_flag(sb, EXT4_MF_FC_INELIGIBLE);
fs/ext4/fast_commit.c
337
ext4_set_mount_flag(sb, EXT4_MF_FC_INELIGIBLE);
fs/ext4/fast_commit.c
338
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
405
struct super_block *sb = inode->i_sb;
fs/ext4/fast_commit.c
406
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
412
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_ENCRYPTED_FILENAME,
fs/ext4/fast_commit.c
420
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_NOMEM, handle);
fs/ext4/fast_commit.c
431
alloc_ctx = ext4_fc_lock(sb);
fs/ext4/fast_commit.c
452
ext4_fc_unlock(sb, alloc_ctx);
fs/ext4/fast_commit.c
667
static void ext4_fc_submit_bh(struct super_block *sb, bool is_tail)
fs/ext4/fast_commit.c
670
struct buffer_head *bh = EXT4_SB(sb)->s_fc_bh;
fs/ext4/fast_commit.c
673
if (test_opt(sb, BARRIER) && is_tail)
fs/ext4/fast_commit.c
680
EXT4_SB(sb)->s_fc_bh = NULL;
fs/ext4/fast_commit.c
696
static u8 *ext4_fc_reserve_space(struct super_block *sb, int len, u32 *crc)
fs/ext4/fast_commit.c
699
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
714
ret = jbd2_fc_get_buf(EXT4_SB(sb)->s_journal, &bh);
fs/ext4/fast_commit.c
742
ext4_fc_submit_bh(sb, false);
fs/ext4/fast_commit.c
744
ret = jbd2_fc_get_buf(EXT4_SB(sb)->s_journal, &bh);
fs/ext4/fast_commit.c
760
static int ext4_fc_write_tail(struct super_block *sb, u32 crc)
fs/ext4/fast_commit.c
762
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
772
dst = ext4_fc_reserve_space(sb, EXT4_FC_TAG_BASE_LEN + sizeof(tail), &crc);
fs/ext4/fast_commit.c
794
ext4_fc_submit_bh(sb, true);
fs/ext4/fast_commit.c
803
static bool ext4_fc_add_tlv(struct super_block *sb, u16 tag, u16 len, u8 *val,
fs/ext4/fast_commit.c
809
dst = ext4_fc_reserve_space(sb, EXT4_FC_TAG_BASE_LEN + len, crc);
fs/ext4/fast_commit.c
823
static bool ext4_fc_add_dentry_tlv(struct super_block *sb, u32 *crc,
fs/ext4/fast_commit.c
829
u8 *dst = ext4_fc_reserve_space(sb,
fs/ext4/fast_commit.c
972
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
973
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fast_commit.c
995
struct super_block *sb = journal->j_private;
fs/ext4/fast_commit.c
996
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/file.c
199
struct super_block *sb = inode->i_sb;
fs/ext4/file.c
200
unsigned long blockmask = sb->s_blocksize - 1;
fs/ext4/file.c
730
struct super_block *sb = inode->i_sb;
fs/ext4/file.c
749
sb_start_pagefault(sb);
fs/ext4/file.c
753
handle = ext4_journal_start_sb(sb, EXT4_HT_WRITE_PAGE,
fs/ext4/file.c
754
EXT4_DATA_TRANS_BLOCKS(sb));
fs/ext4/file.c
757
sb_end_pagefault(sb);
fs/ext4/file.c
768
ext4_should_retry_alloc(sb, &retries))
fs/ext4/file.c
774
sb_end_pagefault(sb);
fs/ext4/file.c
834
static int ext4_sample_last_mounted(struct super_block *sb,
fs/ext4/file.c
837
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/file.c
843
if (likely(ext4_test_mount_flag(sb, EXT4_MF_MNTDIR_SAMPLED)))
fs/ext4/file.c
846
if (ext4_emergency_state(sb) || sb_rdonly(sb) ||
fs/ext4/file.c
847
!sb_start_intwrite_trylock(sb))
fs/ext4/file.c
850
ext4_set_mount_flag(sb, EXT4_MF_MNTDIR_SAMPLED);
fs/ext4/file.c
864
handle = ext4_journal_start_sb(sb, EXT4_HT_MISC, 1);
fs/ext4/file.c
869
err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh,
fs/ext4/file.c
875
ext4_superblock_csum_set(sb);
fs/ext4/file.c
881
sb_end_intwrite(sb);
fs/ext4/fsmap.c
138
ext4_get_group_no_and_offset(sb, info->gfi_next_fsblk,
fs/ext4/fsmap.c
140
trace_ext4_fsmap_mapping(sb, info->gfi_dev, agno,
fs/ext4/fsmap.c
16
void ext4_fsmap_from_internal(struct super_block *sb, struct fsmap *dest,
fs/ext4/fsmap.c
163
ext4_get_group_no_and_offset(sb, rec_fsblk, &agno, &cno);
fs/ext4/fsmap.c
164
trace_ext4_fsmap_mapping(sb, info->gfi_dev, agno, EXT4_C2B(sbi, cno),
fs/ext4/fsmap.c
189
static int ext4_getfsmap_meta_helper(struct super_block *sb,
fs/ext4/fsmap.c
196
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fsmap.c
201
ext4_group_first_block_no(sb, agno));
fs/ext4/fsmap.c
21
dest->fmr_physical = src->fmr_physical << sb->s_blocksize_bits;
fs/ext4/fsmap.c
218
error = ext4_getfsmap_helper(sb, info,
fs/ext4/fsmap.c
224
error = ext4_getfsmap_helper(sb, info, p);
fs/ext4/fsmap.c
24
dest->fmr_length = src->fmr_length << sb->s_blocksize_bits;
fs/ext4/fsmap.c
243
static int ext4_getfsmap_datadev_helper(struct super_block *sb,
fs/ext4/fsmap.c
251
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fsmap.c
256
fsb = (EXT4_C2B(sbi, start) + ext4_group_first_block_no(sb, agno));
fs/ext4/fsmap.c
271
error = ext4_getfsmap_helper(sb, info, &info->gfi_lastfree);
fs/ext4/fsmap.c
283
error = ext4_getfsmap_helper(sb, info, p);
fs/ext4/fsmap.c
300
ext4_group_first_block_no(sb, agno + 1)) {
fs/ext4/fsmap.c
306
return ext4_getfsmap_helper(sb, info, &irec);
fs/ext4/fsmap.c
31
void ext4_fsmap_to_internal(struct super_block *sb, struct ext4_fsmap *dest,
fs/ext4/fsmap.c
310
static int ext4_getfsmap_logdev(struct super_block *sb, struct ext4_fsmap *keys,
fs/ext4/fsmap.c
313
journal_t *journal = EXT4_SB(sb)->s_journal;
fs/ext4/fsmap.c
322
trace_ext4_fsmap_low_key(sb, info->gfi_dev, 0,
fs/ext4/fsmap.c
327
trace_ext4_fsmap_high_key(sb, info->gfi_dev, 0,
fs/ext4/fsmap.c
341
return ext4_getfsmap_helper(sb, info, &irec);
fs/ext4/fsmap.c
36
dest->fmr_physical = src->fmr_physical >> sb->s_blocksize_bits;
fs/ext4/fsmap.c
368
static unsigned int ext4_getfsmap_find_sb(struct super_block *sb,
fs/ext4/fsmap.c
372
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fsmap.c
373
ext4_fsblk_t fsb = ext4_group_first_block_no(sb, agno);
fs/ext4/fsmap.c
376
unsigned long metagroup = agno / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/fsmap.c
38
dest->fmr_length = src->fmr_length >> sb->s_blocksize_bits;
fs/ext4/fsmap.c
380
if (ext4_bg_has_super(sb, agno)) {
fs/ext4/fsmap.c
388
len = ext4_bg_num_gdb(sb, agno);
fs/ext4/fsmap.c
398
if (!ext4_has_feature_meta_bg(sb) || metagroup < first_meta_bg) {
fs/ext4/fsmap.c
470
static int ext4_getfsmap_find_fixed_metadata(struct super_block *sb,
fs/ext4/fsmap.c
480
for (agno = 0; agno < EXT4_SB(sb)->s_groups_count; agno++) {
fs/ext4/fsmap.c
481
gdp = ext4_get_group_desc(sb, agno, NULL);
fs/ext4/fsmap.c
488
error = ext4_getfsmap_find_sb(sb, agno, meta_list);
fs/ext4/fsmap.c
494
ext4_block_bitmap(sb, gdp), 1,
fs/ext4/fsmap.c
501
ext4_inode_bitmap(sb, gdp), 1,
fs/ext4/fsmap.c
508
ext4_inode_table(sb, gdp),
fs/ext4/fsmap.c
509
EXT4_SB(sb)->s_itb_per_group,
fs/ext4/fsmap.c
528
static int ext4_getfsmap_datadev(struct super_block *sb,
fs/ext4/fsmap.c
532
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fsmap.c
558
ext4_get_group_no_and_offset(sb, start_fsb, &start_ag, &first_cluster);
fs/ext4/fsmap.c
559
ext4_get_group_no_and_offset(sb, end_fsb, &end_ag, &last_cluster);
fs/ext4/fsmap.c
573
error = ext4_getfsmap_find_fixed_metadata(sb, &info->gfi_meta_list);
fs/ext4/fsmap.c
58
int (*gfd_fn)(struct super_block *sb,
fs/ext4/fsmap.c
592
trace_ext4_fsmap_low_key(sb, info->gfi_dev, info->gfi_agno,
fs/ext4/fsmap.c
597
trace_ext4_fsmap_high_key(sb, info->gfi_dev, info->gfi_agno,
fs/ext4/fsmap.c
602
error = ext4_mballoc_query_range(sb, info->gfi_agno,
fs/ext4/fsmap.c
620
error = ext4_getfsmap_helper(sb, info, &info->gfi_lastfree);
fs/ext4/fsmap.c
636
error = ext4_getfsmap_helper(sb, info, &irec);
fs/ext4/fsmap.c
646
static bool ext4_getfsmap_is_valid_device(struct super_block *sb,
fs/ext4/fsmap.c
650
fm->fmr_device == new_encode_dev(sb->s_bdev->bd_dev))
fs/ext4/fsmap.c
652
if (EXT4_SB(sb)->s_journal_bdev_file &&
fs/ext4/fsmap.c
654
new_encode_dev(file_bdev(EXT4_SB(sb)->s_journal_bdev_file)->bd_dev))
fs/ext4/fsmap.c
703
int ext4_getfsmap(struct super_block *sb, struct ext4_fsmap_head *head,
fs/ext4/fsmap.c
714
if (!ext4_getfsmap_is_valid_device(sb, &head->fmh_keys[0]) ||
fs/ext4/fsmap.c
715
!ext4_getfsmap_is_valid_device(sb, &head->fmh_keys[1]))
fs/ext4/fsmap.c
722
handlers[0].gfd_dev = new_encode_dev(sb->s_bdev->bd_dev);
fs/ext4/fsmap.c
724
if (EXT4_SB(sb)->s_journal_bdev_file) {
fs/ext4/fsmap.c
726
file_bdev(EXT4_SB(sb)->s_journal_bdev_file)->bd_dev);
fs/ext4/fsmap.c
784
error = handlers[i].gfd_fn(sb, dkeys, &info);
fs/ext4/fsmap.c
85
static int ext4_getfsmap_helper(struct super_block *sb,
fs/ext4/fsmap.c
90
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/fsmap.h
31
void ext4_fsmap_from_internal(struct super_block *sb, struct fsmap *dest,
fs/ext4/fsmap.h
33
void ext4_fsmap_to_internal(struct super_block *sb, struct ext4_fsmap *dest,
fs/ext4/fsmap.h
39
int ext4_getfsmap(struct super_block *sb, struct ext4_fsmap_head *head,
fs/ext4/ialloc.c
100
blk = ext4_inode_bitmap(sb, desc);
fs/ext4/ialloc.c
101
if (!ext4_inode_bitmap_csum_verify(sb, desc, bh) ||
fs/ext4/ialloc.c
1017
group = (goal - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1018
ino = (goal - 1) % EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
102
ext4_simulate_fail(sb, EXT4_SIM_IBITMAP_CRC)) {
fs/ext4/ialloc.c
1024
ret2 = find_group_orlov(sb, dir, &group, mode, qstr);
fs/ext4/ialloc.c
1026
ret2 = find_group_other(sb, dir, &group, mode);
fs/ext4/ialloc.c
103
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
104
ext4_error(sb, "Corrupt inode bitmap - block_group = %u, "
fs/ext4/ialloc.c
1042
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
1049
if (ext4_free_inodes_count(sb, gdp) == 0)
fs/ext4/ialloc.c
1053
grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
106
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/ialloc.c
1063
inode_bitmap_bh = ext4_read_inode_bitmap(sb, group);
fs/ext4/ialloc.c
1073
ret2 = find_inode_bit(sb, group, inode_bitmap_bh, &ino);
fs/ext4/ialloc.c
1077
if (group == 0 && (ino + 1) < EXT4_FIRST_INO(sb)) {
fs/ext4/ialloc.c
1078
ext4_error(sb, "reserved inode found cleared - "
fs/ext4/ialloc.c
1080
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/ialloc.c
1089
ext4_trans_default_revoke_credits(sb));
fs/ext4/ialloc.c
1092
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1097
err = ext4_journal_get_write_access(handle, sb, inode_bitmap_bh,
fs/ext4/ialloc.c
1100
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1103
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1109
ret2 = find_inode_bit(sb, group, inode_bitmap_bh, &ino);
fs/ext4/ialloc.c
1117
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
112
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
1133
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1138
err = ext4_journal_get_write_access(handle, sb, group_desc_bh,
fs/ext4/ialloc.c
1141
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1146
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
1150
block_bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/ialloc.c
1156
err = ext4_journal_get_write_access(handle, sb, block_bitmap_bh,
fs/ext4/ialloc.c
1160
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1168
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1169
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
1172
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/ialloc.c
1173
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/ialloc.c
1174
ext4_block_bitmap_csum_set(sb, gdp, block_bitmap_bh);
fs/ext4/ialloc.c
1175
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1177
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
1181
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1187
if (ext4_has_group_desc_csum(sb)) {
fs/ext4/ialloc.c
1192
grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
1202
ext4_lock_group(sb, group); /* while we modify the bg desc */
fs/ext4/ialloc.c
1203
free = EXT4_INODES_PER_GROUP(sb) -
fs/ext4/ialloc.c
1204
ext4_itable_unused_count(sb, gdp);
fs/ext4/ialloc.c
1215
ext4_itable_unused_set(sb, gdp,
fs/ext4/ialloc.c
1216
(EXT4_INODES_PER_GROUP(sb) - ino));
fs/ext4/ialloc.c
1220
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1223
ext4_free_inodes_set(sb, gdp, ext4_free_inodes_count(sb, gdp) - 1);
fs/ext4/ialloc.c
1225
ext4_used_dirs_set(sb, gdp, ext4_used_dirs_count(sb, gdp) + 1);
fs/ext4/ialloc.c
123
ext4_read_inode_bitmap(struct super_block *sb, ext4_group_t block_group)
fs/ext4/ialloc.c
1233
if (ext4_has_group_desc_csum(sb)) {
fs/ext4/ialloc.c
1234
ext4_inode_bitmap_csum_set(sb, gdp, inode_bitmap_bh);
fs/ext4/ialloc.c
1235
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1237
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
1242
ext4_std_error(sb, err);
fs/ext4/ialloc.c
1256
inode->i_ino = ino + group * EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
126
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ialloc.c
1284
ext4_error(sb, "failed to insert inode %lu: doubly allocated?",
fs/ext4/ialloc.c
1286
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/ialloc.c
1293
if (ext4_has_feature_metadata_csum(sb)) {
fs/ext4/ialloc.c
1306
if (ext4_has_feature_inline_data(sb) &&
fs/ext4/ialloc.c
131
desc = ext4_get_group_desc(sb, block_group, NULL);
fs/ext4/ialloc.c
1335
if (ext4_has_feature_extents(sb)) {
fs/ext4/ialloc.c
1349
ext4_std_error(sb, err);
fs/ext4/ialloc.c
135
bitmap_blk = ext4_inode_bitmap(sb, desc);
fs/ext4/ialloc.c
1372
struct inode *ext4_orphan_get(struct super_block *sb, unsigned long ino)
fs/ext4/ialloc.c
1374
unsigned long max_ino = le32_to_cpu(EXT4_SB(sb)->s_es->s_inodes_count);
fs/ext4/ialloc.c
138
ext4_error(sb, "Invalid inode bitmap blk %llu in "
fs/ext4/ialloc.c
1381
if (ino < EXT4_FIRST_INO(sb) || ino > max_ino)
fs/ext4/ialloc.c
1384
block_group = (ino - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1385
bit = (ino - 1) % EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1386
bitmap_bh = ext4_read_inode_bitmap(sb, block_group);
fs/ext4/ialloc.c
1397
inode = ext4_iget(sb, ino, EXT4_IGET_NORMAL);
fs/ext4/ialloc.c
140
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/ialloc.c
1400
ext4_error_err(sb, -err,
fs/ext4/ialloc.c
1423
ext4_error(sb, "bad orphan inode %lu", ino);
fs/ext4/ialloc.c
144
bh = sb_getblk(sb, bitmap_blk);
fs/ext4/ialloc.c
1444
unsigned long ext4_count_free_inodes(struct super_block *sb)
fs/ext4/ialloc.c
1448
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/ialloc.c
1454
es = EXT4_SB(sb)->s_es;
fs/ext4/ialloc.c
1459
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/ialloc.c
146
ext4_warning(sb, "Cannot read inode bitmap - "
fs/ext4/ialloc.c
1462
desc_count += ext4_free_inodes_count(sb, gdp);
fs/ext4/ialloc.c
1464
bitmap_bh = ext4_read_inode_bitmap(sb, i);
fs/ext4/ialloc.c
1471
EXT4_INODES_PER_GROUP(sb) / 8);
fs/ext4/ialloc.c
1473
(unsigned long) i, ext4_free_inodes_count(sb, gdp), x);
fs/ext4/ialloc.c
1484
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/ialloc.c
1487
desc_count += ext4_free_inodes_count(sb, gdp);
fs/ext4/ialloc.c
1495
unsigned long ext4_count_dirs(struct super_block * sb)
fs/ext4/ialloc.c
1498
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/ialloc.c
1501
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/ialloc.c
1504
count += ext4_used_dirs_count(sb, gdp);
fs/ext4/ialloc.c
1517
int ext4_init_inode_table(struct super_block *sb, ext4_group_t group,
fs/ext4/ialloc.c
1520
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/ialloc.c
1521
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ialloc.c
1529
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
1540
handle = ext4_journal_start_sb(sb, EXT4_HT_MISC, 1);
fs/ext4/ialloc.c
1553
used_inos = EXT4_INODES_PER_GROUP(sb) -
fs/ext4/ialloc.c
1554
ext4_itable_unused_count(sb, gdp);
fs/ext4/ialloc.c
1559
ext4_error(sb, "Something is wrong with group %u: "
fs/ext4/ialloc.c
1563
ext4_itable_unused_count(sb, gdp));
fs/ext4/ialloc.c
1568
used_inos += group * EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
1574
(used_inos < EXT4_FIRST_INO(sb))) {
fs/ext4/ialloc.c
1575
ext4_error(sb, "Something is wrong with group %u: "
fs/ext4/ialloc.c
1578
group, ext4_itable_unused_count(sb, gdp),
fs/ext4/ialloc.c
1585
blk = ext4_inode_table(sb, gdp) + used_blks;
fs/ext4/ialloc.c
1589
ret = ext4_journal_get_write_access(handle, sb, group_desc_bh,
fs/ext4/ialloc.c
160
ext4_lock_group(sb, block_group);
fs/ext4/ialloc.c
1604
ret = sb_issue_zeroout(sb, blk, num, GFP_NOFS);
fs/ext4/ialloc.c
1608
blkdev_issue_flush(sb->s_bdev);
fs/ext4/ialloc.c
161
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
1611
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
1613
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
1614
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
164
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
166
ext4_error(sb, "Inode bitmap for bg 0 marked "
fs/ext4/ialloc.c
171
memset(bh->b_data, 0, (EXT4_INODES_PER_GROUP(sb) + 7) / 8);
fs/ext4/ialloc.c
172
ext4_mark_bitmap_end(EXT4_INODES_PER_GROUP(sb),
fs/ext4/ialloc.c
173
sb->s_blocksize * 8, bh->b_data);
fs/ext4/ialloc.c
177
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
181
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
195
trace_ext4_load_inode_bitmap(sb, block_group);
fs/ext4/ialloc.c
198
ext4_simulate_fail(sb, EXT4_SIM_IBITMAP_EIO));
fs/ext4/ialloc.c
201
ext4_error_err(sb, EIO, "Cannot read inode bitmap - "
fs/ext4/ialloc.c
204
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/ialloc.c
210
err = ext4_validate_inode_bitmap(sb, desc, block_group, bh);
fs/ext4/ialloc.c
237
struct super_block *sb = inode->i_sb;
fs/ext4/ialloc.c
250
if (!sb) {
fs/ext4/ialloc.c
256
ext4_msg(sb, KERN_ERR, "%s:%d: inode #%lu: count=%d",
fs/ext4/ialloc.c
262
ext4_msg(sb, KERN_ERR, "%s:%d: inode #%lu: nlink=%d\n",
fs/ext4/ialloc.c
266
sbi = EXT4_SB(sb);
fs/ext4/ialloc.c
281
if (ino < EXT4_FIRST_INO(sb) || ino > le32_to_cpu(es->s_inodes_count)) {
fs/ext4/ialloc.c
282
ext4_error(sb, "reserved or nonexistent inode %lu", ino);
fs/ext4/ialloc.c
285
block_group = (ino - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
286
bit = (ino - 1) % EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
287
bitmap_bh = ext4_read_inode_bitmap(sb, block_group);
fs/ext4/ialloc.c
295
grp = ext4_get_group_info(sb, block_group);
fs/ext4/ialloc.c
303
fatal = ext4_journal_get_write_access(handle, sb, bitmap_bh,
fs/ext4/ialloc.c
309
gdp = ext4_get_group_desc(sb, block_group, &bh2);
fs/ext4/ialloc.c
312
fatal = ext4_journal_get_write_access(handle, sb, bh2,
fs/ext4/ialloc.c
315
ext4_lock_group(sb, block_group);
fs/ext4/ialloc.c
318
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
322
count = ext4_free_inodes_count(sb, gdp) + 1;
fs/ext4/ialloc.c
323
ext4_free_inodes_set(sb, gdp, count);
fs/ext4/ialloc.c
325
count = ext4_used_dirs_count(sb, gdp) - 1;
fs/ext4/ialloc.c
326
ext4_used_dirs_set(sb, gdp, count);
fs/ext4/ialloc.c
330
ext4_inode_bitmap_csum_set(sb, gdp, bitmap_bh);
fs/ext4/ialloc.c
331
ext4_group_desc_csum_set(sb, block_group, gdp);
fs/ext4/ialloc.c
332
ext4_unlock_group(sb, block_group);
fs/ext4/ialloc.c
354
ext4_error(sb, "bit already cleared for inode %lu", ino);
fs/ext4/ialloc.c
355
ext4_mark_group_bitmap_corrupted(sb, block_group,
fs/ext4/ialloc.c
361
ext4_std_error(sb, fatal);
fs/ext4/ialloc.c
375
static void get_orlov_stats(struct super_block *sb, ext4_group_t g,
fs/ext4/ialloc.c
381
struct flex_groups *fg = sbi_array_rcu_deref(EXT4_SB(sb),
fs/ext4/ialloc.c
389
desc = ext4_get_group_desc(sb, g, NULL);
fs/ext4/ialloc.c
391
stats->free_inodes = ext4_free_inodes_count(sb, desc);
fs/ext4/ialloc.c
392
stats->free_clusters = ext4_free_group_clusters(sb, desc);
fs/ext4/ialloc.c
393
stats->used_dirs = ext4_used_dirs_count(sb, desc);
fs/ext4/ialloc.c
422
static int find_group_orlov(struct super_block *sb, struct inode *parent,
fs/ext4/ialloc.c
427
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ialloc.c
428
ext4_group_t real_ngroups = ext4_get_groups_count(sb);
fs/ext4/ialloc.c
429
int inodes_per_group = EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
456
((parent == d_inode(sb->s_root)) ||
fs/ext4/ialloc.c
470
get_orlov_stats(sb, g, flex_size, &stats);
fs/ext4/ialloc.c
502
desc = ext4_get_group_desc(sb, grp+i, NULL);
fs/ext4/ialloc.c
503
if (desc && ext4_free_inodes_count(sb, desc)) {
fs/ext4/ialloc.c
515
min_clusters = avefreec - EXT4_CLUSTERS_PER_GROUP(sb)*flex_size / 4;
fs/ext4/ialloc.c
531
get_orlov_stats(sb, grp, flex_size, &stats);
fs/ext4/ialloc.c
548
desc = ext4_get_group_desc(sb, grp, NULL);
fs/ext4/ialloc.c
550
grp_free = ext4_free_inodes_count(sb, desc);
fs/ext4/ialloc.c
570
static int find_group_other(struct super_block *sb, struct inode *parent,
fs/ext4/ialloc.c
574
ext4_group_t i, last, ngroups = ext4_get_groups_count(sb);
fs/ext4/ialloc.c
576
int flex_size = ext4_flex_bg_size(EXT4_SB(sb));
fs/ext4/ialloc.c
594
desc = ext4_get_group_desc(sb, i, NULL);
fs/ext4/ialloc.c
595
if (desc && ext4_free_inodes_count(sb, desc)) {
fs/ext4/ialloc.c
613
return find_group_orlov(sb, parent, group, mode, NULL);
fs/ext4/ialloc.c
620
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
621
if (desc && ext4_free_inodes_count(sb, desc) &&
fs/ext4/ialloc.c
622
ext4_free_group_clusters(sb, desc))
fs/ext4/ialloc.c
644
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
645
if (desc && ext4_free_inodes_count(sb, desc) &&
fs/ext4/ialloc.c
646
ext4_free_group_clusters(sb, desc))
fs/ext4/ialloc.c
658
desc = ext4_get_group_desc(sb, *group, NULL);
fs/ext4/ialloc.c
659
if (desc && ext4_free_inodes_count(sb, desc))
fs/ext4/ialloc.c
675
static int recently_deleted(struct super_block *sb, ext4_group_t group, int ino)
fs/ext4/ialloc.c
680
int inodes_per_block = EXT4_SB(sb)->s_inodes_per_block;
fs/ext4/ialloc.c
685
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/ialloc.c
690
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
692
ino >= EXT4_INODES_PER_GROUP(sb) - ext4_itable_unused_count(sb, gdp)))
fs/ext4/ialloc.c
695
bh = sb_find_get_block(sb, ext4_inode_table(sb, gdp) +
fs/ext4/ialloc.c
705
offset = (ino % inodes_per_block) * EXT4_INODE_SIZE(sb);
fs/ext4/ialloc.c
725
static int find_inode_bit(struct super_block *sb, ext4_group_t group,
fs/ext4/ialloc.c
728
bool check_recently_deleted = EXT4_SB(sb)->s_journal == NULL;
fs/ext4/ialloc.c
729
unsigned long recently_deleted_ino = EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
734
EXT4_INODES_PER_GROUP(sb), *ino);
fs/ext4/ialloc.c
735
if (*ino >= EXT4_INODES_PER_GROUP(sb))
fs/ext4/ialloc.c
738
if (check_recently_deleted && recently_deleted(sb, group, *ino)) {
fs/ext4/ialloc.c
741
if (*ino < EXT4_INODES_PER_GROUP(sb))
fs/ext4/ialloc.c
747
if (recently_deleted_ino >= EXT4_INODES_PER_GROUP(sb))
fs/ext4/ialloc.c
759
int ext4_mark_inode_used(struct super_block *sb, int ino)
fs/ext4/ialloc.c
761
unsigned long max_ino = le32_to_cpu(EXT4_SB(sb)->s_es->s_inodes_count);
fs/ext4/ialloc.c
768
if (ino < EXT4_FIRST_INO(sb) || ino > max_ino)
fs/ext4/ialloc.c
771
group = (ino - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
772
bit = (ino - 1) % EXT4_INODES_PER_GROUP(sb);
fs/ext4/ialloc.c
773
inode_bitmap_bh = ext4_read_inode_bitmap(sb, group);
fs/ext4/ialloc.c
782
gdp = ext4_get_group_desc(sb, group, &group_desc_bh);
fs/ext4/ialloc.c
79
static int ext4_validate_inode_bitmap(struct super_block *sb,
fs/ext4/ialloc.c
793
ext4_std_error(sb, err);
fs/ext4/ialloc.c
798
ext4_std_error(sb, err);
fs/ext4/ialloc.c
803
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
807
block_bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/ialloc.c
818
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
819
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/ialloc.c
822
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/ialloc.c
823
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/ialloc.c
824
ext4_block_bitmap_csum_set(sb, gdp, block_bitmap_bh);
fs/ext4/ialloc.c
825
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
827
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
831
ext4_std_error(sb, err);
fs/ext4/ialloc.c
837
if (ext4_has_group_desc_csum(sb)) {
fs/ext4/ialloc.c
840
ext4_lock_group(sb, group); /* while we modify the bg desc */
fs/ext4/ialloc.c
841
free = EXT4_INODES_PER_GROUP(sb) -
fs/ext4/ialloc.c
842
ext4_itable_unused_count(sb, gdp);
fs/ext4/ialloc.c
854
ext4_itable_unused_set(sb, gdp,
fs/ext4/ialloc.c
855
(EXT4_INODES_PER_GROUP(sb) - bit - 1));
fs/ext4/ialloc.c
857
ext4_lock_group(sb, group);
fs/ext4/ialloc.c
860
ext4_free_inodes_set(sb, gdp, ext4_free_inodes_count(sb, gdp) - 1);
fs/ext4/ialloc.c
861
if (ext4_has_group_desc_csum(sb)) {
fs/ext4/ialloc.c
862
ext4_inode_bitmap_csum_set(sb, gdp, inode_bitmap_bh);
fs/ext4/ialloc.c
863
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/ialloc.c
866
ext4_unlock_group(sb, group);
fs/ext4/ialloc.c
87
if (EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY)
fs/ext4/ialloc.c
877
struct super_block *sb = dir->i_sb;
fs/ext4/ialloc.c
888
__ext4_xattr_set_credits(sb, NULL /* inode */,
fs/ext4/ialloc.c
907
__ext4_xattr_set_credits(sb, NULL /* inode */,
fs/ext4/ialloc.c
913
nblocks += __ext4_xattr_set_credits(sb,
fs/ext4/ialloc.c
93
grp = ext4_get_group_info(sb, block_group);
fs/ext4/ialloc.c
938
struct super_block *sb;
fs/ext4/ialloc.c
958
sb = dir->i_sb;
fs/ext4/ialloc.c
959
sbi = EXT4_SB(sb);
fs/ext4/ialloc.c
961
ret2 = ext4_emergency_state(sb);
fs/ext4/ialloc.c
965
ngroups = ext4_get_groups_count(sb);
fs/ext4/ialloc.c
967
inode = new_inode(sb);
fs/ext4/ialloc.c
97
ext4_lock_group(sb, block_group);
fs/ext4/ialloc.c
981
} else if (test_opt(sb, GRPID)) {
fs/ext4/ialloc.c
988
if (ext4_has_feature_project(sb) &&
fs/ext4/indirect.c
148
struct super_block *sb = inode->i_sb;
fs/ext4/indirect.c
161
if (key > ext4_blocks_count(EXT4_SB(sb)->s_es)) {
fs/ext4/indirect.c
166
bh = sb_getblk(sb, key);
fs/ext4/inline.c
1386
struct super_block *sb;
fs/ext4/inline.c
1419
sb = inode->i_sb;
fs/ext4/inline.c
1499
get_dtype(sb, de->file_type)))
fs/ext4/inode.c
1775
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
1778
ext4_msg(sb, KERN_CRIT, "Total free blocks count %lld",
fs/ext4/inode.c
1780
ext4_count_free_clusters(sb)));
fs/ext4/inode.c
1781
ext4_msg(sb, KERN_CRIT, "Free/Dirty block details");
fs/ext4/inode.c
1782
ext4_msg(sb, KERN_CRIT, "free_blocks=%lld",
fs/ext4/inode.c
1783
(long long) EXT4_C2B(EXT4_SB(sb),
fs/ext4/inode.c
1785
ext4_msg(sb, KERN_CRIT, "dirty_blocks=%lld",
fs/ext4/inode.c
1786
(long long) EXT4_C2B(EXT4_SB(sb),
fs/ext4/inode.c
1788
ext4_msg(sb, KERN_CRIT, "Block reservation details");
fs/ext4/inode.c
1789
ext4_msg(sb, KERN_CRIT, "i_reserved_data_blocks=%u",
fs/ext4/inode.c
2479
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
2481
if (ext4_emergency_state(sb))
fs/ext4/inode.c
2489
(err == -ENOSPC && ext4_count_free_clusters(sb))) {
fs/ext4/inode.c
2502
ext4_msg(sb, KERN_CRIT,
fs/ext4/inode.c
2509
ext4_msg(sb, KERN_CRIT,
fs/ext4/inode.c
3015
struct super_block *sb = mapping->host->i_sb;
fs/ext4/inode.c
3024
ret = ext4_emergency_state(sb);
fs/ext4/inode.c
3028
alloc_ctx = ext4_writepages_down_read(sb);
fs/ext4/inode.c
3037
ext4_writepages_up_read(sb, alloc_ctx);
fs/ext4/inode.c
3081
static int ext4_nonda_switch(struct super_block *sb)
fs/ext4/inode.c
3084
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/inode.c
3102
try_to_writeback_inodes_sb(sb, WB_REASON_FS_FREE_SPACE);
fs/ext4/inode.c
4165
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
4172
partial_start = lstart & (sb->s_blocksize - 1);
fs/ext4/inode.c
4173
partial_end = byte_end & (sb->s_blocksize - 1);
fs/ext4/inode.c
4175
start = lstart >> sb->s_blocksize_bits;
fs/ext4/inode.c
4176
end = byte_end >> sb->s_blocksize_bits;
fs/ext4/inode.c
4180
(partial_start || (partial_end != sb->s_blocksize - 1))) {
fs/ext4/inode.c
4188
lstart, sb->s_blocksize);
fs/ext4/inode.c
4193
if (partial_end != sb->s_blocksize - 1)
fs/ext4/inode.c
4339
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
4341
loff_t max_end = sb->s_maxbytes;
fs/ext4/inode.c
4355
max_end = EXT4_SB(sb)->s_bitmap_maxbytes - sb->s_blocksize;
fs/ext4/inode.c
4366
end = round_up(inode->i_size, sb->s_blocksize);
fs/ext4/inode.c
4375
if (!IS_ALIGNED(offset | end, sb->s_blocksize)) {
fs/ext4/inode.c
4398
ext4_std_error(sb, ret);
fs/ext4/inode.c
4625
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
4643
if (!ext4_has_feature_huge_file(sb))
fs/ext4/inode.c
4765
static int __ext4_get_inode_loc(struct super_block *sb, unsigned long ino,
fs/ext4/inode.c
4777
ino > le32_to_cpu(EXT4_SB(sb)->s_es->s_inodes_count))
fs/ext4/inode.c
4780
iloc->block_group = (ino - 1) / EXT4_INODES_PER_GROUP(sb);
fs/ext4/inode.c
4781
gdp = ext4_get_group_desc(sb, iloc->block_group, NULL);
fs/ext4/inode.c
4788
inodes_per_block = EXT4_SB(sb)->s_inodes_per_block;
fs/ext4/inode.c
4790
EXT4_INODES_PER_GROUP(sb));
fs/ext4/inode.c
4791
iloc->offset = (inode_offset % inodes_per_block) * EXT4_INODE_SIZE(sb);
fs/ext4/inode.c
4793
block = ext4_inode_table(sb, gdp);
fs/ext4/inode.c
4794
if ((block <= le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block)) ||
fs/ext4/inode.c
4795
(block >= ext4_blocks_count(EXT4_SB(sb)->s_es))) {
fs/ext4/inode.c
4796
ext4_error(sb, "Invalid inode table block %llu in "
fs/ext4/inode.c
4802
bh = sb_getblk(sb, block);
fs/ext4/inode.c
4827
bitmap_bh = sb_getblk(sb, ext4_inode_bitmap(sb, gdp));
fs/ext4/inode.c
4867
if (EXT4_SB(sb)->s_inode_readahead_blks) {
fs/ext4/inode.c
4870
__u32 ra_blks = EXT4_SB(sb)->s_inode_readahead_blks;
fs/ext4/inode.c
4872
table = ext4_inode_table(sb, gdp);
fs/ext4/inode.c
4878
num = EXT4_INODES_PER_GROUP(sb);
fs/ext4/inode.c
4879
if (ext4_has_group_desc_csum(sb))
fs/ext4/inode.c
4880
num -= ext4_itable_unused_count(sb, gdp);
fs/ext4/inode.c
4885
ext4_sb_breadahead_unmovable(sb, b++);
fs/ext4/inode.c
4893
trace_ext4_load_inode(sb, ino);
fs/ext4/inode.c
4895
ext4_simulate_fail(sb, EXT4_SIM_INODE_EIO));
fs/ext4/inode.c
4941
int ext4_get_fc_inode_loc(struct super_block *sb, unsigned long ino,
fs/ext4/inode.c
4944
return __ext4_get_inode_loc(sb, ino, NULL, iloc, NULL);
fs/ext4/inode.c
5011
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
5013
if (ext4_has_feature_huge_file(sb)) {
fs/ext4/inode.c
5116
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
5119
max_order = EXT4_SB(sb)->s_max_folio_order;
fs/ext4/inode.c
5123
min_order = EXT4_SB(sb)->s_min_folio_order;
fs/ext4/inode.c
5133
struct inode *__ext4_iget(struct super_block *sb, unsigned long ino,
fs/ext4/inode.c
5140
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/inode.c
5142
journal_t *journal = EXT4_SB(sb)->s_journal;
fs/ext4/inode.c
5150
if ((!(flags & EXT4_IGET_SPECIAL) && is_special_ino(sb, ino)) ||
fs/ext4/inode.c
5155
__ext4_error(sb, function, line, false, EFSCORRUPTED, 0,
fs/ext4/inode.c
5161
inode = iget_locked(sb, ino);
fs/ext4/inode.c
5204
if (ext4_has_feature_metadata_csum(sb)) {
fs/ext4/inode.c
5215
ext4_simulate_fail(sb, EXT4_SIM_INODE_CRC)) &&
fs/ext4/inode.c
5216
(!(EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY))) {
fs/ext4/inode.c
5226
if (ext4_has_feature_project(sb) &&
fs/ext4/inode.c
5227
EXT4_INODE_SIZE(sb) > EXT4_GOOD_OLD_INODE_SIZE &&
fs/ext4/inode.c
5282
if (ext4_has_feature_64bit(sb))
fs/ext4/inode.c
5285
inode->i_size = ext4_isize(sb, raw_inode);
fs/ext4/inode.c
5298
if (!ext4_has_feature_dir_index(sb) &&
fs/ext4/inode.c
5299
ext4_has_feature_metadata_csum(sb) &&
fs/ext4/inode.c
5386
if (!(EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY) &&
fs/ext4/inode.c
5427
if ((EXT4_SB(sb)->s_mount_state & EXT4_ORPHAN_FS)) {
fs/ext4/inode.c
5504
static void __ext4_update_other_inode_time(struct super_block *sb,
fs/ext4/inode.c
5511
inode = find_inode_by_ino_rcu(sb, ino);
fs/ext4/inode.c
5541
static void ext4_update_other_inodes_time(struct super_block *sb,
fs/ext4/inode.c
5545
int i, inodes_per_block = EXT4_SB(sb)->s_inodes_per_block;
fs/ext4/inode.c
5546
int inode_size = EXT4_INODE_SIZE(sb);
fs/ext4/inode.c
5558
__ext4_update_other_inode_time(sb, orig_ino, ino,
fs/ext4/inode.c
5578
struct super_block *sb = inode->i_sb;
fs/ext4/inode.c
5594
if (!ext4_has_feature_large_file(sb) ||
fs/ext4/inode.c
5595
EXT4_SB(sb)->s_es->s_rev_level == cpu_to_le32(EXT4_GOOD_OLD_REV))
fs/ext4/inode.c
5616
BUFFER_TRACE(EXT4_SB(sb)->s_sbh, "get write access");
fs/ext4/inode.c
5617
err = ext4_journal_get_write_access(handle, sb,
fs/ext4/inode.c
5618
EXT4_SB(sb)->s_sbh,
fs/ext4/inode.c
5622
lock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/inode.c
5623
ext4_set_feature_large_file(sb);
fs/ext4/inode.c
5624
ext4_superblock_csum_set(sb);
fs/ext4/inode.c
5625
unlock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/inode.c
5628
EXT4_SB(sb)->s_sbh);
fs/ext4/ioctl.c
107
static int ext4_update_backup_sb(struct super_block *sb,
fs/ext4/ioctl.c
1071
struct super_block *sb = file_inode(filp)->i_sb;
fs/ext4/ioctl.c
1086
if (!EXT4_SB(sb)->s_journal)
fs/ext4/ioctl.c
1090
!bdev_max_discard_sectors(EXT4_SB(sb)->s_journal->j_dev))
fs/ext4/ioctl.c
1104
jbd2_journal_lock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1105
err = jbd2_journal_flush(EXT4_SB(sb)->s_journal, flush_flags);
fs/ext4/ioctl.c
1106
jbd2_journal_unlock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1116
struct super_block *sb = file_inode(filp)->i_sb;
fs/ext4/ioctl.c
1142
ret = ext4_update_superblocks_fn(sb, ext4_sb_setlabel, new_label);
fs/ext4/ioctl.c
117
if (!ext4_bg_has_super(sb, grp))
fs/ext4/ioctl.c
1203
struct super_block *sb = file_inode(filp)->i_sb;
fs/ext4/ioctl.c
1214
if (((ext4_has_feature_gdt_csum(sb) ||
fs/ext4/ioctl.c
1215
ext4_has_feature_metadata_csum(sb))
fs/ext4/ioctl.c
1216
&& !ext4_has_feature_csum_seed(sb))
fs/ext4/ioctl.c
1217
|| ext4_has_feature_stable_inodes(sb))
fs/ext4/ioctl.c
1233
ret = ext4_update_superblocks_fn(sb, ext4_sb_setuuid, &uuid);
fs/ext4/ioctl.c
126
offset = do_div(sb_block, sb->s_blocksize);
fs/ext4/ioctl.c
128
sb_block = ext4_group_first_block_no(sb, grp);
fs/ext4/ioctl.c
132
trace_ext4_update_sb(sb, sb_block, handle ? 1 : 0);
fs/ext4/ioctl.c
134
bh = ext4_sb_bread(sb, sb_block, 0);
fs/ext4/ioctl.c
1388
struct super_block *sb = file_inode(filp)->i_sb;
fs/ext4/ioctl.c
1389
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ioctl.c
140
err = ext4_journal_get_write_access(handle, sb,
fs/ext4/ioctl.c
149
if (ext4_has_feature_metadata_csum(sb) &&
fs/ext4/ioctl.c
151
ext4_msg(sb, KERN_ERR, "Invalid checksum for backup "
fs/ext4/ioctl.c
1523
ret = ext4_update_superblocks_fn(sb, ext4_sb_setparams, ¶ms);
fs/ext4/ioctl.c
1535
struct super_block *sb = inode->i_sb;
fs/ext4/ioctl.c
1542
return ext4_ioc_getfsmap(sb, (void __user *)arg);
fs/ext4/ioctl.c
1557
ext4_warning(sb, "Setting inode version is not "
fs/ext4/ioctl.c
156
func(EXT4_SB(sb), es, arg);
fs/ext4/ioctl.c
157
if (ext4_has_feature_metadata_csum(sb))
fs/ext4/ioctl.c
1595
err = ext4_resize_begin(sb);
fs/ext4/ioctl.c
1604
if (ext4_has_feature_bigalloc(sb)) {
fs/ext4/ioctl.c
1605
ext4_msg(sb, KERN_ERR,
fs/ext4/ioctl.c
1615
err = ext4_group_extend(sb, EXT4_SB(sb)->s_es, n_blocks_count);
fs/ext4/ioctl.c
1616
if (EXT4_SB(sb)->s_journal) {
fs/ext4/ioctl.c
1617
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_RESIZE,
fs/ext4/ioctl.c
1619
jbd2_journal_lock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1620
err2 = jbd2_journal_flush(EXT4_SB(sb)->s_journal, 0);
fs/ext4/ioctl.c
1621
jbd2_journal_unlock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1627
err2 = ext4_resize_end(sb, false);
fs/ext4/ioctl.c
1721
err = swap_inode_boot_loader(sb, idmap, inode);
fs/ext4/ioctl.c
1729
ext4_group_t o_group = EXT4_SB(sb)->s_groups_count;
fs/ext4/ioctl.c
1736
err = ext4_resize_begin(sb);
fs/ext4/ioctl.c
174
ext4_std_error(sb, err);
fs/ext4/ioctl.c
1744
err = ext4_resize_fs(sb, n_blocks_count);
fs/ext4/ioctl.c
1745
if (EXT4_SB(sb)->s_journal) {
fs/ext4/ioctl.c
1746
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_RESIZE, NULL);
fs/ext4/ioctl.c
1747
jbd2_journal_lock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1748
err2 = jbd2_journal_flush(EXT4_SB(sb)->s_journal, 0);
fs/ext4/ioctl.c
1749
jbd2_journal_unlock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
1754
if (!err && (o_group < EXT4_SB(sb)->s_groups_count) &&
fs/ext4/ioctl.c
1755
ext4_has_group_desc_csum(sb) &&
fs/ext4/ioctl.c
1756
test_opt(sb, INIT_INODE_TABLE))
fs/ext4/ioctl.c
1757
err = ext4_register_li_request(sb, o_group);
fs/ext4/ioctl.c
1760
err2 = ext4_resize_end(sb, true);
fs/ext4/ioctl.c
1774
if (!bdev_max_discard_sectors(sb->s_bdev))
fs/ext4/ioctl.c
1781
if (test_opt(sb, NOLOAD) && ext4_has_feature_journal(sb))
fs/ext4/ioctl.c
1788
ret = ext4_trim_fs(sb, &range);
fs/ext4/ioctl.c
1808
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1816
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1821
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1826
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1831
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1836
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1841
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1846
if (!ext4_has_feature_encrypt(sb))
fs/ext4/ioctl.c
1878
return ext4_ioctl_shutdown(sb, arg);
fs/ext4/ioctl.c
188
int ext4_update_superblocks_fn(struct super_block *sb,
fs/ext4/ioctl.c
1881
if (!ext4_has_feature_verity(sb))
fs/ext4/ioctl.c
1886
if (!ext4_has_feature_verity(sb))
fs/ext4/ioctl.c
1891
if (!ext4_has_feature_verity(sb))
fs/ext4/ioctl.c
1900
return ext4_ioctl_getlabel(EXT4_SB(sb), (void __user *)arg);
fs/ext4/ioctl.c
1907
return ext4_ioctl_getuuid(EXT4_SB(sb), (void __user *)arg);
fs/ext4/ioctl.c
1911
return ext4_ioctl_get_tune_sb(EXT4_SB(sb),
fs/ext4/ioctl.c
199
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ioctl.c
2008
int ext4_update_overhead(struct super_block *sb, bool force)
fs/ext4/ioctl.c
2010
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ioctl.c
2012
if (ext4_emergency_state(sb) || sb_rdonly(sb))
fs/ext4/ioctl.c
2018
return ext4_update_superblocks_fn(sb, set_overhead, &sbi->s_overhead);
fs/ext4/ioctl.c
206
ext4_msg(sb, KERN_ERR, "Can't modify superblock while"
fs/ext4/ioctl.c
215
handle = ext4_journal_start_sb(sb, EXT4_HT_MISC, 3);
fs/ext4/ioctl.c
222
err = ext4_update_primary_sb(sb, handle, func, arg);
fs/ext4/ioctl.c
224
ext4_msg(sb, KERN_ERR, "Failed to update primary "
fs/ext4/ioctl.c
229
primary_grp = ext4_get_group_number(sb, sbi->s_sbh->b_blocknr);
fs/ext4/ioctl.c
230
ngroups = ext4_get_groups_count(sb);
fs/ext4/ioctl.c
244
ret = ext4_update_backup_sb(sb, handle, grp, func, arg);
fs/ext4/ioctl.c
266
grp = ext4_list_backups(sb, &three, &five, &seven);
fs/ext4/ioctl.c
375
static long swap_inode_boot_loader(struct super_block *sb,
fs/ext4/ioctl.c
387
inode_bl = ext4_iget(sb, EXT4_BOOT_LOADER_INO,
fs/ext4/ioctl.c
433
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_SWAP_BOOT, handle);
fs/ext4/ioctl.c
449
if (ext4_has_feature_extents(sb)) {
fs/ext4/ioctl.c
59
int ext4_update_primary_sb(struct super_block *sb, handle_t *handle,
fs/ext4/ioctl.c
598
struct super_block *sb = inode->i_sb;
fs/ext4/ioctl.c
623
if (!ext4_has_feature_casefold(sb)) {
fs/ext4/ioctl.c
64
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ioctl.c
68
trace_ext4_update_sb(sb, bh->b_blocknr, 1);
fs/ext4/ioctl.c
71
err = ext4_journal_get_write_access(handle, sb,
fs/ext4/ioctl.c
718
struct super_block *sb = inode->i_sb;
fs/ext4/ioctl.c
727
if (!ext4_has_feature_project(sb)) {
fs/ext4/ioctl.c
734
if (EXT4_INODE_SIZE(sb) <= EXT4_GOOD_OLD_INODE_SIZE)
fs/ext4/ioctl.c
758
EXT4_SB(sb)->s_want_extra_isize,
fs/ext4/ioctl.c
767
EXT4_QUOTA_INIT_BLOCKS(sb) +
fs/ext4/ioctl.c
768
EXT4_QUOTA_DEL_BLOCKS(sb) + 3);
fs/ext4/ioctl.c
776
transfer_to[PRJQUOTA] = dqget(sb, make_kqid_projid(kprojid));
fs/ext4/ioctl.c
79
ext4_superblock_csum_set(sb);
fs/ext4/ioctl.c
810
int ext4_force_shutdown(struct super_block *sb, u32 flags)
fs/ext4/ioctl.c
812
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/ioctl.c
818
if (ext4_forced_shutdown(sb))
fs/ext4/ioctl.c
821
ext4_msg(sb, KERN_ALERT, "shut down requested (%d)", flags);
fs/ext4/ioctl.c
822
trace_ext4_shutdown(sb, flags);
fs/ext4/ioctl.c
826
ret = bdev_freeze(sb->s_bdev);
fs/ext4/ioctl.c
830
bdev_thaw(sb->s_bdev);
fs/ext4/ioctl.c
835
(void) ext4_force_commit(sb);
fs/ext4/ioctl.c
847
clear_opt(sb, DISCARD);
fs/ext4/ioctl.c
848
fserror_report_shutdown(sb, GFP_KERNEL);
fs/ext4/ioctl.c
852
static int ext4_ioctl_shutdown(struct super_block *sb, unsigned long arg)
fs/ext4/ioctl.c
862
return ext4_force_shutdown(sb, flags);
fs/ext4/ioctl.c
888
static int ext4_ioc_getfsmap(struct super_block *sb,
fs/ext4/ioctl.c
916
ext4_fsmap_to_internal(sb, &xhead.fmh_keys[0], &head.fmh_keys[0]);
fs/ext4/ioctl.c
917
ext4_fsmap_to_internal(sb, &xhead.fmh_keys[1], &head.fmh_keys[1]);
fs/ext4/ioctl.c
919
trace_ext4_getfsmap_low_key(sb, &xhead.fmh_keys[0]);
fs/ext4/ioctl.c
920
trace_ext4_getfsmap_high_key(sb, &xhead.fmh_keys[1]);
fs/ext4/ioctl.c
922
info.gi_sb = sb;
fs/ext4/ioctl.c
924
error = ext4_getfsmap(sb, &xhead, ext4_getfsmap_format, &info);
fs/ext4/ioctl.c
94
ext4_std_error(sb, err);
fs/ext4/ioctl.c
951
struct super_block *sb = file_inode(file)->i_sb;
fs/ext4/ioctl.c
954
err = ext4_resize_begin(sb);
fs/ext4/ioctl.c
958
if (ext4_has_feature_bigalloc(sb)) {
fs/ext4/ioctl.c
959
ext4_msg(sb, KERN_ERR,
fs/ext4/ioctl.c
969
err = ext4_group_add(sb, input);
fs/ext4/ioctl.c
970
if (EXT4_SB(sb)->s_journal) {
fs/ext4/ioctl.c
971
ext4_fc_mark_ineligible(sb, EXT4_FC_REASON_RESIZE, NULL);
fs/ext4/ioctl.c
972
jbd2_journal_lock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
973
err2 = jbd2_journal_flush(EXT4_SB(sb)->s_journal, 0);
fs/ext4/ioctl.c
974
jbd2_journal_unlock_updates(EXT4_SB(sb)->s_journal);
fs/ext4/ioctl.c
979
if (!err && ext4_has_group_desc_csum(sb) &&
fs/ext4/ioctl.c
980
test_opt(sb, INIT_INODE_TABLE))
fs/ext4/ioctl.c
981
err = ext4_register_li_request(sb, input->group);
fs/ext4/ioctl.c
983
err2 = ext4_resize_end(sb, false);
fs/ext4/mballoc-test.c
104
ret = percpu_counter_init(&EXT4_SB(sb)->s_dirtyclusters_counter, 0,
fs/ext4/mballoc-test.c
112
percpu_counter_destroy(&EXT4_SB(sb)->s_freeclusters_counter);
fs/ext4/mballoc-test.c
114
ext4_mb_release(sb);
fs/ext4/mballoc-test.c
116
kfree(sb->s_bdev->bd_queue);
fs/ext4/mballoc-test.c
117
kfree(sb->s_bdev);
fs/ext4/mballoc-test.c
121
static void mbt_mb_release(struct super_block *sb)
fs/ext4/mballoc-test.c
123
percpu_counter_destroy(&EXT4_SB(sb)->s_dirtyclusters_counter);
fs/ext4/mballoc-test.c
124
percpu_counter_destroy(&EXT4_SB(sb)->s_freeclusters_counter);
fs/ext4/mballoc-test.c
125
ext4_mb_release(sb);
fs/ext4/mballoc-test.c
126
kfree(sb->s_bdev->bd_queue);
fs/ext4/mballoc-test.c
127
kfree(sb->s_bdev);
fs/ext4/mballoc-test.c
130
static int mbt_set(struct super_block *sb, void *data)
fs/ext4/mballoc-test.c
138
struct super_block *sb;
fs/ext4/mballoc-test.c
145
sb = sget(&mbt_fs_type, NULL, mbt_set, 0, NULL);
fs/ext4/mballoc-test.c
146
if (IS_ERR(sb))
fs/ext4/mballoc-test.c
159
sbi->s_sb = sb;
fs/ext4/mballoc-test.c
160
sb->s_fs_info = sbi;
fs/ext4/mballoc-test.c
162
up_write(&sb->s_umount);
fs/ext4/mballoc-test.c
163
return sb;
fs/ext4/mballoc-test.c
166
deactivate_locked_super(sb);
fs/ext4/mballoc-test.c
172
static void mbt_ext4_free_super_block(struct super_block *sb)
fs/ext4/mballoc-test.c
174
struct mbt_ext4_super_block *fsb = MBT_SB(sb);
fs/ext4/mballoc-test.c
175
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc-test.c
178
deactivate_super(sb);
fs/ext4/mballoc-test.c
190
static void mbt_init_sb_layout(struct super_block *sb,
fs/ext4/mballoc-test.c
193
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc-test.c
196
sb->s_blocksize = 1UL << layout->blocksize_bits;
fs/ext4/mballoc-test.c
197
sb->s_blocksize_bits = layout->blocksize_bits;
fs/ext4/mballoc-test.c
207
sb->s_blocksize_bits - (fls(layout->desc_size) - 1);
fs/ext4/mballoc-test.c
215
static int mbt_grp_ctx_init(struct super_block *sb,
fs/ext4/mballoc-test.c
218
ext4_grpblk_t max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
220
grp_ctx->bitmap_bh.b_data = kzalloc(EXT4_BLOCK_SIZE(sb), GFP_KERNEL);
fs/ext4/mballoc-test.c
223
mb_set_bits(grp_ctx->bitmap_bh.b_data, max, sb->s_blocksize * 8 - max);
fs/ext4/mballoc-test.c
224
ext4_free_group_clusters_set(sb, &grp_ctx->desc, max);
fs/ext4/mballoc-test.c
235
static void mbt_ctx_mark_used(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc-test.c
238
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc-test.c
243
static void *mbt_ctx_bitmap(struct super_block *sb, ext4_group_t group)
fs/ext4/mballoc-test.c
245
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc-test.c
251
static int mbt_ctx_init(struct super_block *sb)
fs/ext4/mballoc-test.c
253
struct mbt_ctx *ctx = MBT_CTX(sb);
fs/ext4/mballoc-test.c
254
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc-test.c
261
if (mbt_grp_ctx_init(sb, &ctx->grp_ctx[i]))
fs/ext4/mballoc-test.c
270
ext4_free_group_clusters_set(sb, &ctx->grp_ctx[0].desc,
fs/ext4/mballoc-test.c
271
EXT4_CLUSTERS_PER_GROUP(sb) - 1);
fs/ext4/mballoc-test.c
281
static void mbt_ctx_release(struct super_block *sb)
fs/ext4/mballoc-test.c
283
struct mbt_ctx *ctx = MBT_CTX(sb);
fs/ext4/mballoc-test.c
284
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc-test.c
292
ext4_read_block_bitmap_nowait_stub(struct super_block *sb, ext4_group_t block_group,
fs/ext4/mballoc-test.c
295
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, block_group);
fs/ext4/mballoc-test.c
302
static int ext4_wait_block_bitmap_stub(struct super_block *sb,
fs/ext4/mballoc-test.c
317
ext4_get_group_desc_stub(struct super_block *sb, ext4_group_t block_group,
fs/ext4/mballoc-test.c
320
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, block_group);
fs/ext4/mballoc-test.c
329
ext4_mb_mark_context_stub(handle_t *handle, struct super_block *sb, bool state,
fs/ext4/mballoc-test.c
334
struct mbt_grp_ctx *grp_ctx = MBT_GRP_CTX(sb, group);
fs/ext4/mballoc-test.c
34
static struct inode *mbt_alloc_inode(struct super_block *sb)
fs/ext4/mballoc-test.c
350
struct super_block *sb;
fs/ext4/mballoc-test.c
353
sb = mbt_ext4_alloc_super_block();
fs/ext4/mballoc-test.c
354
if (sb == NULL)
fs/ext4/mballoc-test.c
357
mbt_init_sb_layout(sb, layout);
fs/ext4/mballoc-test.c
359
ret = mbt_ctx_init(sb);
fs/ext4/mballoc-test.c
361
mbt_ext4_free_super_block(sb);
fs/ext4/mballoc-test.c
365
test->priv = sb;
fs/ext4/mballoc-test.c
380
if (mbt_mb_init(sb) != 0) {
fs/ext4/mballoc-test.c
381
mbt_ctx_release(sb);
fs/ext4/mballoc-test.c
382
mbt_ext4_free_super_block(sb);
fs/ext4/mballoc-test.c
391
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
393
mbt_mb_release(sb);
fs/ext4/mballoc-test.c
394
mbt_ctx_release(sb);
fs/ext4/mballoc-test.c
395
mbt_ext4_free_super_block(sb);
fs/ext4/mballoc-test.c
400
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
406
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc-test.c
412
inode->i_sb = sb;
fs/ext4/mballoc-test.c
416
ar.goal = ext4_group_first_block_no(sb, goal_group);
fs/ext4/mballoc-test.c
423
ar.goal = ext4_group_first_block_no(sb, goal_group);
fs/ext4/mballoc-test.c
430
mbt_ctx_mark_used(sb, goal_group, 0, EXT4_CLUSTERS_PER_GROUP(sb));
fs/ext4/mballoc-test.c
431
ar.goal = ext4_group_first_block_no(sb, goal_group);
fs/ext4/mballoc-test.c
434
ext4_group_first_block_no(sb, goal_group + 1), found,
fs/ext4/mballoc-test.c
436
ext4_group_first_block_no(sb, goal_group + 1), found);
fs/ext4/mballoc-test.c
439
for (i = goal_group; i < ext4_get_groups_count(sb); i++)
fs/ext4/mballoc-test.c
440
mbt_ctx_mark_used(sb, i, 0, EXT4_CLUSTERS_PER_GROUP(sb));
fs/ext4/mballoc-test.c
441
ar.goal = ext4_group_first_block_no(sb, goal_group);
fs/ext4/mballoc-test.c
444
ext4_group_first_block_no(sb, 0) + EXT4_C2B(sbi, 1), found,
fs/ext4/mballoc-test.c
446
ext4_group_first_block_no(sb, 0 + EXT4_C2B(sbi, 1)), found);
fs/ext4/mballoc-test.c
449
for (i = 0; i < ext4_get_groups_count(sb); i++)
fs/ext4/mballoc-test.c
450
mbt_ctx_mark_used(sb, i, 0, EXT4_CLUSTERS_PER_GROUP(sb));
fs/ext4/mballoc-test.c
451
ar.goal = ext4_group_first_block_no(sb, goal_group);
fs/ext4/mballoc-test.c
465
mbt_generate_test_ranges(struct super_block *sb, struct test_range *ranges,
fs/ext4/mballoc-test.c
471
max = EXT4_CLUSTERS_PER_GROUP(sb) / count;
fs/ext4/mballoc-test.c
483
validate_free_blocks_simple(struct kunit *test, struct super_block *sb,
fs/ext4/mballoc-test.c
488
ext4_grpblk_t bit, max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
491
for (i = 0; i < ext4_get_groups_count(sb); i++) {
fs/ext4/mballoc-test.c
495
bitmap = mbt_ctx_bitmap(sb, i);
fs/ext4/mballoc-test.c
501
bitmap = mbt_ctx_bitmap(sb, goal_group);
fs/ext4/mballoc-test.c
513
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
514
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc-test.c
521
inode->i_sb = sb;
fs/ext4/mballoc-test.c
526
block = ext4_group_first_block_no(sb, goal_group) +
fs/ext4/mballoc-test.c
529
validate_free_blocks_simple(test, sb, goal_group, start, len);
fs/ext4/mballoc-test.c
530
mbt_ctx_mark_used(sb, goal_group, 0, EXT4_CLUSTERS_PER_GROUP(sb));
fs/ext4/mballoc-test.c
535
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
536
ext4_grpblk_t max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
540
for (i = 0; i < ext4_get_groups_count(sb); i++)
fs/ext4/mballoc-test.c
541
mbt_ctx_mark_used(sb, i, 0, max);
fs/ext4/mballoc-test.c
543
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
555
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
568
bitmap = mbt_ctx_bitmap(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
569
memset(bitmap, 0, sb->s_blocksize);
fs/ext4/mballoc-test.c
573
max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
584
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
590
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
595
inode->i_sb = sb;
fs/ext4/mballoc-test.c
598
ac.ac_sb = sb;
fs/ext4/mballoc-test.c
605
static void mbt_generate_buddy(struct super_block *sb, void *buddy,
fs/ext4/mballoc-test.c
608
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc-test.c
61
static void mbt_kill_sb(struct super_block *sb)
fs/ext4/mballoc-test.c
613
memset(buddy, 0xff, sb->s_blocksize);
fs/ext4/mballoc-test.c
615
bb_counters[MB_NUM_ORDERS(sb)]));
fs/ext4/mballoc-test.c
618
max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
63
generic_shutdown_super(sb);
fs/ext4/mballoc-test.c
639
for (order = 1; order < MB_NUM_ORDERS(sb) - 1; order++) {
fs/ext4/mballoc-test.c
659
max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
676
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
687
for (i = 1; i < MB_NUM_ORDERS(sb); i++) {
fs/ext4/mballoc-test.c
697
do_test_generate_buddy(struct kunit *test, struct super_block *sb, void *bitmap,
fs/ext4/mballoc-test.c
703
mbt_generate_buddy(sb, mbt_buddy, bitmap, mbt_grp);
fs/ext4/mballoc-test.c
705
for (i = 0; i < MB_NUM_ORDERS(sb); i++)
fs/ext4/mballoc-test.c
709
memset(ext4_buddy, 0xff, sb->s_blocksize);
fs/ext4/mballoc-test.c
71
static int mbt_mb_init(struct super_block *sb)
fs/ext4/mballoc-test.c
710
ext4_mb_generate_buddy_test(sb, ext4_buddy, bitmap, TEST_GOAL_GROUP,
fs/ext4/mballoc-test.c
713
KUNIT_ASSERT_EQ(test, memcmp(mbt_buddy, ext4_buddy, sb->s_blocksize),
fs/ext4/mballoc-test.c
720
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
726
bitmap = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
728
expected_bb = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
730
generate_bb = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
733
bb_counters[MB_NUM_ORDERS(sb)]), GFP_KERNEL);
fs/ext4/mballoc-test.c
735
generate_grp = ext4_get_group_info(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
738
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
741
do_test_generate_buddy(test, sb, bitmap, expected_bb,
fs/ext4/mballoc-test.c
751
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
763
ext4_lock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
765
ext4_unlock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
77
sb->s_bdev = kzalloc_obj(*sb->s_bdev);
fs/ext4/mballoc-test.c
770
memset(buddy, 0xff, sb->s_blocksize);
fs/ext4/mballoc-test.c
771
for (i = 0; i < MB_NUM_ORDERS(sb); i++)
fs/ext4/mballoc-test.c
773
ext4_mb_generate_buddy_test(sb, buddy, bitmap, 0, grp);
fs/ext4/mballoc-test.c
775
KUNIT_ASSERT_EQ(test, memcmp(buddy, e4b->bd_buddy, sb->s_blocksize),
fs/ext4/mballoc-test.c
78
if (sb->s_bdev == NULL)
fs/ext4/mballoc-test.c
783
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
791
if (sb->s_blocksize > PAGE_SIZE)
fs/ext4/mballoc-test.c
794
bitmap = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
796
buddy = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
799
bb_counters[MB_NUM_ORDERS(sb)]), GFP_KERNEL);
fs/ext4/mballoc-test.c
802
ret = ext4_mb_load_buddy_test(sb, TEST_GOAL_GROUP, &e4b);
fs/ext4/mballoc-test.c
805
grp->bb_free = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
808
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
81
sb->s_bdev->bd_queue = kzalloc_obj(struct request_queue);
fs/ext4/mballoc-test.c
82
if (sb->s_bdev->bd_queue == NULL) {
fs/ext4/mballoc-test.c
821
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
828
ext4_lock_group(sb, e4b->bd_group);
fs/ext4/mballoc-test.c
83
kfree(sb->s_bdev);
fs/ext4/mballoc-test.c
830
ext4_unlock_group(sb, e4b->bd_group);
fs/ext4/mballoc-test.c
835
memset(buddy, 0xff, sb->s_blocksize);
fs/ext4/mballoc-test.c
836
for (i = 0; i < MB_NUM_ORDERS(sb); i++)
fs/ext4/mballoc-test.c
838
ext4_mb_generate_buddy_test(sb, buddy, bitmap, 0, grp);
fs/ext4/mballoc-test.c
840
KUNIT_ASSERT_EQ(test, memcmp(buddy, e4b->bd_buddy, sb->s_blocksize),
fs/ext4/mballoc-test.c
849
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
858
if (sb->s_blocksize > PAGE_SIZE)
fs/ext4/mballoc-test.c
861
bitmap = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
863
buddy = kunit_kzalloc(test, sb->s_blocksize, GFP_KERNEL);
fs/ext4/mballoc-test.c
866
bb_counters[MB_NUM_ORDERS(sb)]), GFP_KERNEL);
fs/ext4/mballoc-test.c
869
ret = ext4_mb_load_buddy_test(sb, TEST_GOAL_GROUP, &e4b);
fs/ext4/mballoc-test.c
873
ex.fe_len = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc-test.c
876
ext4_lock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
878
ext4_unlock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
883
memset(bitmap, 0xff, sb->s_blocksize);
fs/ext4/mballoc-test.c
885
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
897
struct super_block *sb = (struct super_block *)test->priv;
fs/ext4/mballoc-test.c
905
if (sb->s_blocksize > PAGE_SIZE)
fs/ext4/mballoc-test.c
908
ret = ext4_mb_load_buddy_test(sb, TEST_GOAL_GROUP, &e4b);
fs/ext4/mballoc-test.c
91
INIT_LIST_HEAD(&sb->s_inodes);
fs/ext4/mballoc-test.c
913
mbt_generate_test_ranges(sb, ranges, TEST_RANGE_COUNT);
fs/ext4/mballoc-test.c
92
sb->s_op = &mbt_sops;
fs/ext4/mballoc-test.c
921
ext4_lock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
923
ext4_unlock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
932
ext4_lock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
935
ext4_unlock_group(sb, TEST_GOAL_GROUP);
fs/ext4/mballoc-test.c
94
ret = ext4_mb_init(sb);
fs/ext4/mballoc-test.c
98
block = ext4_count_free_clusters(sb);
fs/ext4/mballoc-test.c
99
ret = percpu_counter_init(&EXT4_SB(sb)->s_freeclusters_counter, block,
fs/ext4/mballoc.c
1171
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
1172
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
1245
mb_set_largest_free_order(struct super_block *sb, struct ext4_group_info *grp)
fs/ext4/mballoc.c
1247
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
1250
for (new = MB_NUM_ORDERS(sb) - 1; new >= 0; new--)
fs/ext4/mballoc.c
1266
if (test_opt2(sb, MB_OPTIMIZE_SCAN) && new >= 0 && grp->bb_free) {
fs/ext4/mballoc.c
1275
mb_debug(sb, "insert group: %u to s_mb_largest_free_orders[%d] failed, err %d",
fs/ext4/mballoc.c
1281
void ext4_mb_generate_buddy(struct super_block *sb,
fs/ext4/mballoc.c
1285
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
1286
ext4_grpblk_t max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc.c
1305
ext4_mb_mark_free_simple(sb, buddy, first, len, grp);
fs/ext4/mballoc.c
1314
ext4_grp_locked_error(sb, group, 0, 0,
fs/ext4/mballoc.c
1323
ext4_mark_group_bitmap_corrupted(sb, group,
fs/ext4/mballoc.c
1326
mb_set_largest_free_order(sb, grp);
fs/ext4/mballoc.c
1327
mb_update_avg_fragment_size(sb, grp);
fs/ext4/mballoc.c
1383
struct super_block *sb;
fs/ext4/mballoc.c
1392
sb = inode->i_sb;
fs/ext4/mballoc.c
1393
ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
1399
mb_debug(sb, "init folio %lu\n", folio->index);
fs/ext4/mballoc.c
1416
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1430
bh[i] = ext4_read_block_bitmap_nowait(sb, group, false);
fs/ext4/mballoc.c
1436
mb_debug(sb, "read bitmap for group %u\n", group);
fs/ext4/mballoc.c
1445
err2 = ext4_wait_block_bitmap(sb, group, bh[i]);
fs/ext4/mballoc.c
1478
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1486
mb_debug(sb, "put buddy for group %u in folio %lu/%x\n",
fs/ext4/mballoc.c
1488
trace_ext4_mb_buddy_bitmap_load(sb, group);
fs/ext4/mballoc.c
1492
(MB_NUM_ORDERS(sb)));
fs/ext4/mballoc.c
1496
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
1499
ext4_mb_generate_buddy(sb, data, incore, group, grinfo);
fs/ext4/mballoc.c
1500
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
1505
mb_debug(sb, "put bitmap for group %u in folio %lu/%x\n",
fs/ext4/mballoc.c
1507
trace_ext4_mb_bitmap_load(sb, group);
fs/ext4/mballoc.c
1510
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
1514
ext4_mb_generate_from_pa(sb, data, group);
fs/ext4/mballoc.c
1516
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
1542
static int ext4_mb_get_buddy_folio_lock(struct super_block *sb,
fs/ext4/mballoc.c
1545
struct inode *inode = EXT4_SB(sb)->s_buddy_cache;
fs/ext4/mballoc.c
1564
WARN_ON_ONCE(folio_size(folio) < sb->s_blocksize);
fs/ext4/mballoc.c
1582
WARN_ON_ONCE(folio_size(folio) < sb->s_blocksize);
fs/ext4/mballoc.c
1605
int ext4_mb_init_group(struct super_block *sb, ext4_group_t group, gfp_t gfp)
fs/ext4/mballoc.c
1614
mb_debug(sb, "init group %u\n", group);
fs/ext4/mballoc.c
1615
this_grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1628
ret = ext4_mb_get_buddy_folio_lock(sb, group, &e4b, gfp);
fs/ext4/mballoc.c
1675
ext4_mb_load_buddy_gfp(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
1683
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
1687
mb_debug(sb, "load group %u\n", group);
fs/ext4/mballoc.c
1689
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
1693
e4b->bd_blkbits = sb->s_blocksize_bits;
fs/ext4/mballoc.c
1695
e4b->bd_sb = sb;
fs/ext4/mballoc.c
1705
ret = ext4_mb_init_group(sb, group, gfp);
fs/ext4/mballoc.c
1831
static int ext4_mb_load_buddy(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
1834
return ext4_mb_load_buddy_gfp(sb, group, e4b, GFP_NOFS);
fs/ext4/mballoc.c
2007
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
2011
BUG_ON(last >= (sb->s_blocksize << 3));
fs/ext4/mballoc.c
2012
assert_spin_locked(ext4_group_lock_ptr(sb, e4b->bd_group));
fs/ext4/mballoc.c
2026
if (last + 1 < EXT4_SB(sb)->s_mb_maxs[0])
fs/ext4/mballoc.c
2030
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
2042
blocknr = ext4_group_first_block_no(sb, e4b->bd_group);
fs/ext4/mballoc.c
2044
ext4_mark_group_bitmap_corrupted(sb, e4b->bd_group,
fs/ext4/mballoc.c
2046
ext4_grp_locked_error(sb, e4b->bd_group,
fs/ext4/mballoc.c
2082
mb_set_largest_free_order(sb, e4b->bd_info);
fs/ext4/mballoc.c
2083
mb_update_avg_fragment_size(sb, e4b->bd_info);
fs/ext4/mballoc.c
2506
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
2514
for (i = ac->ac_2order; i < MB_NUM_ORDERS(sb); i++) {
fs/ext4/mballoc.c
2544
if (EXT4_SB(sb)->s_mb_stats)
fs/ext4/mballoc.c
2545
atomic_inc(&EXT4_SB(sb)->s_bal_2orders);
fs/ext4/mballoc.c
2560
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
2574
EXT4_CLUSTERS_PER_GROUP(sb), i);
fs/ext4/mballoc.c
2575
if (i >= EXT4_CLUSTERS_PER_GROUP(sb)) {
fs/ext4/mballoc.c
2581
ext4_mark_group_bitmap_corrupted(sb, e4b->bd_group,
fs/ext4/mballoc.c
2583
ext4_grp_locked_error(sb, e4b->bd_group, 0, 0,
fs/ext4/mballoc.c
2598
EXT4_CLUSTERS_PER_GROUP(sb), i);
fs/ext4/mballoc.c
2612
ext4_mark_group_bitmap_corrupted(sb, e4b->bd_group,
fs/ext4/mballoc.c
2614
ext4_grp_locked_error(sb, e4b->bd_group, 0, 0,
fs/ext4/mballoc.c
2643
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
2644
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
2655
first_group_block = ext4_group_first_block_no(sb, e4b->bd_group);
fs/ext4/mballoc.c
2663
while (i < EXT4_CLUSTERS_PER_GROUP(sb)) {
fs/ext4/mballoc.c
2781
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
2782
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
2792
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2793
__release(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2808
__acquire(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2809
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
2815
ext4_get_group_desc(sb, group, NULL);
fs/ext4/mballoc.c
2830
!(ext4_has_group_desc_csum(sb) &&
fs/ext4/mballoc.c
2833
ret = ext4_mb_init_group(sb, group, GFP_NOFS);
fs/ext4/mballoc.c
2839
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2840
__release(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2845
__acquire(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
2846
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
2855
ext4_group_t ext4_mb_prefetch(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
2858
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
2864
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, group,
fs/ext4/mballoc.c
2866
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
2877
ext4_free_group_clusters(sb, gdp) > 0 ) {
fs/ext4/mballoc.c
2878
bh = ext4_read_block_bitmap_nowait(sb, group, true);
fs/ext4/mballoc.c
2935
void ext4_mb_prefetch_fini(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
2943
group = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
2945
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/mballoc.c
2946
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
2949
ext4_free_group_clusters(sb, gdp) > 0) {
fs/ext4/mballoc.c
2950
if (ext4_mb_init_group(sb, group, GFP_NOFS))
fs/ext4/mballoc.c
2960
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
2966
if (cr < CR_ANY_FREE && spin_is_locked(ext4_group_lock_ptr(sb, group)))
fs/ext4/mballoc.c
2977
ret = ext4_mb_load_buddy(sb, group, ac->ac_e4b);
fs/ext4/mballoc.c
2983
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
2984
else if (!ext4_try_lock_group(sb, group))
fs/ext4/mballoc.c
2994
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
3005
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
3006
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3033
if (i >= sbi->s_mb_order2_reqs && i <= MB_NUM_ORDERS(sb)) {
fs/ext4/mballoc.c
3036
MB_NUM_ORDERS(sb));
fs/ext4/mballoc.c
3086
mb_debug(sb, "lost chunk, group: %u, start: %d, len: %d, lost: %d\n",
fs/ext4/mballoc.c
3110
mb_debug(sb, "Best len %d, origin len %d, ac_status %u, ac_flags 0x%x, cr %d ret %d\n",
fs/ext4/mballoc.c
3115
ext4_mb_prefetch_fini(sb, ac->ac_prefetch_grp, ac->ac_prefetch_nr);
fs/ext4/mballoc.c
3122
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3125
if (*pos < 0 || *pos >= ext4_get_groups_count(sb))
fs/ext4/mballoc.c
3133
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3137
if (*pos < 0 || *pos >= ext4_get_groups_count(sb))
fs/ext4/mballoc.c
3145
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3152
sb->s_blocksize_bits,
fs/ext4/mballoc.c
3166
grinfo = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
3171
err = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
3209
struct super_block *sb = seq->private;
fs/ext4/mballoc.c
3210
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3301
ext4_get_groups_count(sb));
fs/ext4/mballoc.c
3312
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3315
if (*pos < 0 || *pos >= 2*MB_NUM_ORDERS(sb))
fs/ext4/mballoc.c
3323
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3327
if (*pos < 0 || *pos >= 2*MB_NUM_ORDERS(sb))
fs/ext4/mballoc.c
3335
struct super_block *sb = pde_data(file_inode(seq->file));
fs/ext4/mballoc.c
3336
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3343
if (position >= MB_NUM_ORDERS(sb)) {
fs/ext4/mballoc.c
3344
position -= MB_NUM_ORDERS(sb);
fs/ext4/mballoc.c
3358
test_opt2(sb, MB_OPTIMIZE_SCAN) ? 1 : 0);
fs/ext4/mballoc.c
3394
int ext4_mb_alloc_groupinfo(struct super_block *sb, ext4_group_t ngroups)
fs/ext4/mballoc.c
3396
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3400
size = (ngroups + EXT4_DESC_PER_BLOCK(sb) - 1) >>
fs/ext4/mballoc.c
3401
EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/mballoc.c
3408
ext4_msg(sb, KERN_ERR, "can't allocate buddy meta group");
fs/ext4/mballoc.c
3427
int ext4_mb_add_groupinfo(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
3432
int idx = group >> EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/mballoc.c
3433
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3435
struct kmem_cache *cachep = get_groupinfo_cache(sb->s_blocksize_bits);
fs/ext4/mballoc.c
3442
if (group % EXT4_DESC_PER_BLOCK(sb) == 0) {
fs/ext4/mballoc.c
3444
EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/mballoc.c
3447
ext4_msg(sb, KERN_ERR, "can't allocate mem "
fs/ext4/mballoc.c
3457
i = group & (EXT4_DESC_PER_BLOCK(sb) - 1);
fs/ext4/mballoc.c
3461
ext4_msg(sb, KERN_ERR, "can't allocate buddy mem");
fs/ext4/mballoc.c
3471
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/mballoc.c
3474
ext4_free_clusters_after_init(sb, group, desc);
fs/ext4/mballoc.c
3477
ext4_free_group_clusters(sb, desc);
fs/ext4/mballoc.c
3487
mb_group_bb_bitmap_alloc(sb, meta_group_info[i], group);
fs/ext4/mballoc.c
3492
if (group % EXT4_DESC_PER_BLOCK(sb) == 0) {
fs/ext4/mballoc.c
3504
static int ext4_mb_init_backend(struct super_block *sb)
fs/ext4/mballoc.c
3506
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
3508
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3514
err = ext4_mb_alloc_groupinfo(sb, ngroups);
fs/ext4/mballoc.c
3518
sbi->s_buddy_cache = new_inode(sb);
fs/ext4/mballoc.c
3520
ext4_msg(sb, KERN_ERR, "can't get new inode");
fs/ext4/mballoc.c
3533
desc = ext4_get_group_desc(sb, i, NULL);
fs/ext4/mballoc.c
3535
ext4_msg(sb, KERN_ERR, "can't read descriptor %u", i);
fs/ext4/mballoc.c
3538
if (ext4_mb_add_groupinfo(sb, i, desc) != 0)
fs/ext4/mballoc.c
3542
if (ext4_has_feature_flex_bg(sb)) {
fs/ext4/mballoc.c
3548
ext4_msg(sb, KERN_ERR, "too many log groups per flexible block group");
fs/ext4/mballoc.c
3552
BLK_MAX_SEGMENT_SIZE >> (sb->s_blocksize_bits - 9));
fs/ext4/mballoc.c
3557
if (sbi->s_mb_prefetch > ext4_get_groups_count(sb))
fs/ext4/mballoc.c
3558
sbi->s_mb_prefetch = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
3568
if (sbi->s_mb_prefetch_limit > ext4_get_groups_count(sb))
fs/ext4/mballoc.c
3569
sbi->s_mb_prefetch_limit = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
3574
cachep = get_groupinfo_cache(sb->s_blocksize_bits);
fs/ext4/mballoc.c
3576
struct ext4_group_info *grp = ext4_get_group_info(sb, i);
fs/ext4/mballoc.c
3646
struct super_block *sb = sbi->s_sb;
fs/ext4/mballoc.c
3663
if ((sb->s_flags & SB_ACTIVE) && !err &&
fs/ext4/mballoc.c
3670
err = ext4_mb_load_buddy(sb, grp, &e4b);
fs/ext4/mballoc.c
3680
ext4_lock_group(sb, grp);
fs/ext4/mballoc.c
3681
ext4_try_to_trim_range(sb, &e4b, fd->efd_start_cluster,
fs/ext4/mballoc.c
3683
ext4_unlock_group(sb, grp);
fs/ext4/mballoc.c
3716
int ext4_mb_init(struct super_block *sb)
fs/ext4/mballoc.c
3718
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3724
i = MB_NUM_ORDERS(sb) * sizeof(*sbi->s_mb_offsets);
fs/ext4/mballoc.c
3732
i = MB_NUM_ORDERS(sb) * sizeof(*sbi->s_mb_maxs);
fs/ext4/mballoc.c
3739
ret = ext4_groupinfo_create_slab(sb->s_blocksize);
fs/ext4/mballoc.c
3744
sbi->s_mb_maxs[0] = sb->s_blocksize << 3;
fs/ext4/mballoc.c
3749
offset_incr = 1 << (sb->s_blocksize_bits - 1);
fs/ext4/mballoc.c
3750
max = sb->s_blocksize << 2;
fs/ext4/mballoc.c
3758
} while (i < MB_NUM_ORDERS(sb));
fs/ext4/mballoc.c
3761
kmalloc_objs(struct xarray, MB_NUM_ORDERS(sb));
fs/ext4/mballoc.c
3766
for (i = 0; i < MB_NUM_ORDERS(sb); i++)
fs/ext4/mballoc.c
3770
kmalloc_objs(struct xarray, MB_NUM_ORDERS(sb));
fs/ext4/mballoc.c
3775
for (i = 0; i < MB_NUM_ORDERS(sb); i++)
fs/ext4/mballoc.c
3843
if (bdev_nonrot(sb->s_bdev))
fs/ext4/mballoc.c
3848
ret = ext4_mb_init_backend(sb);
fs/ext4/mballoc.c
3886
void ext4_mb_release(struct super_block *sb)
fs/ext4/mballoc.c
3888
ext4_group_t ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
3892
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
3893
struct kmem_cache *cachep = get_groupinfo_cache(sb->s_blocksize_bits);
fs/ext4/mballoc.c
3906
grinfo = ext4_get_group_info(sb, i);
fs/ext4/mballoc.c
3910
ext4_lock_group(sb, i);
fs/ext4/mballoc.c
3913
mb_debug(sb, "mballoc: %d PAs left\n",
fs/ext4/mballoc.c
3915
ext4_unlock_group(sb, i);
fs/ext4/mballoc.c
3919
EXT4_DESC_PER_BLOCK(sb) - 1) >>
fs/ext4/mballoc.c
3920
EXT4_DESC_PER_BLOCK_BITS(sb);
fs/ext4/mballoc.c
3931
ext4_msg(sb, KERN_INFO,
fs/ext4/mballoc.c
3936
ext4_msg(sb, KERN_INFO,
fs/ext4/mballoc.c
3945
ext4_msg(sb, KERN_INFO,
fs/ext4/mballoc.c
3949
ext4_msg(sb, KERN_INFO,
fs/ext4/mballoc.c
3959
static inline int ext4_issue_discard(struct super_block *sb,
fs/ext4/mballoc.c
3964
discard_block = (EXT4_C2B(EXT4_SB(sb), cluster) +
fs/ext4/mballoc.c
3965
ext4_group_first_block_no(sb, block_group));
fs/ext4/mballoc.c
3966
count = EXT4_C2B(EXT4_SB(sb), count);
fs/ext4/mballoc.c
3967
trace_ext4_discard_blocks(sb,
fs/ext4/mballoc.c
3970
return sb_issue_discard(sb, discard_block, count, GFP_NOFS, 0);
fs/ext4/mballoc.c
3973
static void ext4_free_data_in_buddy(struct super_block *sb,
fs/ext4/mballoc.c
3980
mb_debug(sb, "gonna free %u blocks in group %u (0x%p):",
fs/ext4/mballoc.c
3983
err = ext4_mb_load_buddy(sb, entry->efd_group, &e4b);
fs/ext4/mballoc.c
3987
atomic_sub(entry->efd_count, &EXT4_SB(sb)->s_mb_free_pending);
fs/ext4/mballoc.c
3991
ext4_lock_group(sb, entry->efd_group);
fs/ext4/mballoc.c
4009
ext4_unlock_group(sb, entry->efd_group);
fs/ext4/mballoc.c
4012
mb_debug(sb, "freed %d blocks in 1 structures\n", count);
fs/ext4/mballoc.c
4019
void ext4_process_freed_data(struct super_block *sb, tid_t commit_tid)
fs/ext4/mballoc.c
4021
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
4030
ext4_free_data_in_buddy(sb, entry);
fs/ext4/mballoc.c
4032
if (test_opt(sb, DISCARD)) {
fs/ext4/mballoc.c
4088
ext4_mb_mark_context(handle_t *handle, struct super_block *sb, bool state,
fs/ext4/mballoc.c
4092
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
4100
handle, sb, state, group, blkoff, len,
fs/ext4/mballoc.c
4105
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
4111
err = ext4_journal_get_write_access(handle, sb, bitmap_bh,
fs/ext4/mballoc.c
4118
gdp = ext4_get_group_desc(sb, group, &gdp_bh);
fs/ext4/mballoc.c
4124
err = ext4_journal_get_write_access(handle, sb, gdp_bh,
fs/ext4/mballoc.c
4130
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
4131
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/mballoc.c
4134
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/mballoc.c
4135
ext4_free_clusters_after_init(sb, group, gdp));
fs/ext4/mballoc.c
4149
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/mballoc.c
4150
ext4_free_group_clusters(sb, gdp) - changed);
fs/ext4/mballoc.c
4153
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/mballoc.c
4154
ext4_free_group_clusters(sb, gdp) + changed);
fs/ext4/mballoc.c
4157
ext4_block_bitmap_csum_set(sb, gdp, bitmap_bh);
fs/ext4/mballoc.c
4158
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/mballoc.c
4159
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
4200
struct super_block *sb;
fs/ext4/mballoc.c
4209
sb = ac->ac_sb;
fs/ext4/mballoc.c
4210
sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
4212
gdp = ext4_get_group_desc(sb, ac->ac_b_ex.fe_group, NULL);
fs/ext4/mballoc.c
4216
ext4_free_group_clusters(sb, gdp));
fs/ext4/mballoc.c
4218
block = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
4221
ext4_error(sb, "Allocating blocks %llu-%llu which overlap "
fs/ext4/mballoc.c
4227
err = ext4_mb_mark_context(handle, sb, true,
fs/ext4/mballoc.c
424
static void ext4_mb_generate_from_pa(struct super_block *sb, void *bitmap,
fs/ext4/mballoc.c
4240
err = ext4_mb_mark_context(handle, sb, true, ac->ac_b_ex.fe_group,
fs/ext4/mballoc.c
4259
void ext4_mb_mark_bb(struct super_block *sb, ext4_fsblk_t block,
fs/ext4/mballoc.c
4262
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
4269
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
4279
thisgrp_len = min(len, EXT4_BLOCKS_PER_GROUP(sb) - EXT4_C2B(sbi, blkoff));
fs/ext4/mballoc.c
4282
if (!ext4_sb_block_valid(sb, NULL, block, thisgrp_len)) {
fs/ext4/mballoc.c
4283
ext4_error(sb, "Marking blocks in system zone - "
fs/ext4/mballoc.c
4289
err = ext4_mb_mark_context(NULL, sb, state,
fs/ext4/mballoc.c
431
static int ext4_try_to_trim_range(struct super_block *sb,
fs/ext4/mballoc.c
4314
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
4318
ac->ac_g_ex.fe_len = EXT4_SB(sb)->s_mb_group_prealloc;
fs/ext4/mballoc.c
4319
mb_debug(sb, "goal %u blocks for locality group\n", ac->ac_g_ex.fe_len);
fs/ext4/mballoc.c
5087
void ext4_mb_generate_from_pa(struct super_block *sb, void *bitmap,
fs/ext4/mballoc.c
5090
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
5112
ext4_get_group_no_and_offset(sb, pa->pa_pstart,
fs/ext4/mballoc.c
5122
mb_debug(sb, "preallocated %d for group %u\n", preallocated, group);
fs/ext4/mballoc.c
5125
static void ext4_mb_mark_pa_deleted(struct super_block *sb,
fs/ext4/mballoc.c
5131
ext4_warning(sb, "deleted pa, type:%d, pblk:%llu, lblk:%u, len:%d\n",
fs/ext4/mballoc.c
5166
struct super_block *sb, struct ext4_prealloc_space *pa)
fs/ext4/mballoc.c
5184
ext4_mb_mark_pa_deleted(sb, pa);
fs/ext4/mballoc.c
5195
grp = ext4_get_group_number(sb, grp_blk);
fs/ext4/mballoc.c
5211
ext4_lock_group(sb, grp);
fs/ext4/mballoc.c
5213
ext4_unlock_group(sb, grp);
fs/ext4/mballoc.c
5259
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
5260
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
5322
pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
5330
mb_debug(sb, "new inode pa %p: %llu/%d for %u\n", pa, pa->pa_pstart,
fs/ext4/mballoc.c
5338
grp = ext4_get_group_info(sb, ac->ac_b_ex.fe_group);
fs/ext4/mballoc.c
5359
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
5372
pa->pa_pstart = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
5382
mb_debug(sb, "new group pa %p: %llu/%d for %u\n", pa, pa->pa_pstart,
fs/ext4/mballoc.c
5387
atomic_add(pa->pa_free, &EXT4_SB(sb)->s_mb_preallocated);
fs/ext4/mballoc.c
5389
grp = ext4_get_group_info(sb, ac->ac_b_ex.fe_group);
fs/ext4/mballoc.c
5426
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
5427
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
5436
ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
fs/ext4/mballoc.c
5446
mb_debug(sb, "free preallocated %u/%u in group %u\n",
fs/ext4/mballoc.c
5447
(unsigned) ext4_group_first_block_no(sb, group) + bit,
fs/ext4/mballoc.c
5451
trace_ext4_mballoc_discard(sb, NULL, group, bit, next - bit);
fs/ext4/mballoc.c
5464
ext4_grp_locked_error(sb, group, 0, 0, "free %u, pa_free %u",
fs/ext4/mballoc.c
5478
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
5482
trace_ext4_mb_release_group_pa(sb, pa);
fs/ext4/mballoc.c
5484
ext4_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit);
fs/ext4/mballoc.c
5486
ext4_warning(sb, "bad group: expected %u, group %u, pa_start %llu",
fs/ext4/mballoc.c
5491
atomic_add(pa->pa_len, &EXT4_SB(sb)->s_mb_discarded);
fs/ext4/mballoc.c
5492
trace_ext4_mballoc_discard(sb, NULL, group, bit, pa->pa_len);
fs/ext4/mballoc.c
5505
ext4_mb_discard_group_preallocations(struct super_block *sb,
fs/ext4/mballoc.c
5508
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
5519
mb_debug(sb, "discard preallocation for group %u\n", group);
fs/ext4/mballoc.c
5523
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
5526
ext4_error_err(sb, -err,
fs/ext4/mballoc.c
5532
err = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
5534
ext4_warning(sb, "Error %d loading buddy information for %u",
fs/ext4/mballoc.c
5540
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
5555
ext4_mb_mark_pa_deleted(sb, pa);
fs/ext4/mballoc.c
5595
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
5599
mb_debug(sb, "discarded (%d) blocks preallocated for group %u bb_free (%d)\n",
fs/ext4/mballoc.c
561
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
5616
struct super_block *sb = inode->i_sb;
fs/ext4/mballoc.c
5628
if (EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY)
fs/ext4/mballoc.c
5631
mb_debug(sb, "discard preallocation for inode %lu\n",
fs/ext4/mballoc.c
565
assert_spin_locked(ext4_group_lock_ptr(sb, e4b->bd_group));
fs/ext4/mballoc.c
5651
ext4_msg(sb, KERN_ERR,
fs/ext4/mballoc.c
5659
ext4_mb_mark_pa_deleted(sb, pa);
fs/ext4/mballoc.c
5689
group = ext4_get_group_number(sb, pa->pa_pstart);
fs/ext4/mballoc.c
5691
err = ext4_mb_load_buddy_gfp(sb, group, &e4b,
fs/ext4/mballoc.c
5694
ext4_error_err(sb, -err, "Error %d loading buddy information for %u",
fs/ext4/mballoc.c
5699
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
570
blocknr = ext4_group_first_block_no(sb, e4b->bd_group);
fs/ext4/mballoc.c
5702
ext4_error_err(sb, -err, "Error %d reading block bitmap for %u",
fs/ext4/mballoc.c
5708
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
571
blocknr += EXT4_C2B(EXT4_SB(sb), first + i);
fs/ext4/mballoc.c
5711
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
572
ext4_mark_group_bitmap_corrupted(sb, e4b->bd_group,
fs/ext4/mballoc.c
574
ext4_grp_locked_error(sb, e4b->bd_group,
fs/ext4/mballoc.c
5751
static inline void ext4_mb_show_pa(struct super_block *sb)
fs/ext4/mballoc.c
5755
if (ext4_emergency_state(sb))
fs/ext4/mballoc.c
5758
ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
5759
mb_debug(sb, "groups: ");
fs/ext4/mballoc.c
5761
struct ext4_group_info *grp = ext4_get_group_info(sb, i);
fs/ext4/mballoc.c
5768
ext4_lock_group(sb, i);
fs/ext4/mballoc.c
5773
ext4_get_group_no_and_offset(sb, pa->pa_pstart,
fs/ext4/mballoc.c
5776
mb_debug(sb, "PA:%u:%d:%d\n", i, start,
fs/ext4/mballoc.c
5779
ext4_unlock_group(sb, i);
fs/ext4/mballoc.c
5780
mb_debug(sb, "%u: %d/%d\n", i, grp->bb_free,
fs/ext4/mballoc.c
5787
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
5789
if (ext4_emergency_state(sb))
fs/ext4/mballoc.c
5792
mb_debug(sb, "Can't allocate:"
fs/ext4/mballoc.c
5794
mb_debug(sb, "status %u flags 0x%x",
fs/ext4/mballoc.c
5796
mb_debug(sb, "orig %lu/%lu/%lu@%lu, "
fs/ext4/mballoc.c
5812
mb_debug(sb, "%u found", ac->ac_found);
fs/ext4/mballoc.c
5813
mb_debug(sb, "used pa: %s, ", str_yes_no(ac->ac_pa));
fs/ext4/mballoc.c
5815
mb_debug(sb, "pa_type %s\n", ac->ac_pa->pa_type == MB_GROUP_PA ?
fs/ext4/mballoc.c
5817
ext4_mb_show_pa(sb);
fs/ext4/mballoc.c
5820
static inline void ext4_mb_show_pa(struct super_block *sb)
fs/ext4/mballoc.c
5892
struct super_block *sb = ar->inode->i_sb;
fs/ext4/mballoc.c
5893
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
5904
if (len >= EXT4_CLUSTERS_PER_GROUP(sb))
fs/ext4/mballoc.c
5905
len = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc.c
5912
ext4_get_group_no_and_offset(sb, goal, &group, &block);
fs/ext4/mballoc.c
5917
ac->ac_sb = sb;
fs/ext4/mballoc.c
5931
mb_debug(sb, "init ac: %u blocks @ %u, goal %u, flags 0x%x, 2^%d, "
fs/ext4/mballoc.c
5941
ext4_mb_discard_lg_preallocations(struct super_block *sb,
fs/ext4/mballoc.c
5950
mb_debug(sb, "discard locality group preallocation\n");
fs/ext4/mballoc.c
5974
ext4_mb_mark_pa_deleted(sb, pa);
fs/ext4/mballoc.c
5996
group = ext4_get_group_number(sb, pa->pa_pstart);
fs/ext4/mballoc.c
5997
err = ext4_mb_load_buddy_gfp(sb, group, &e4b,
fs/ext4/mballoc.c
6000
ext4_error_err(sb, -err, "Error %d loading buddy information for %u",
fs/ext4/mballoc.c
6004
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
6007
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
6027
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
6065
ext4_mb_discard_lg_preallocations(sb, lg,
fs/ext4/mballoc.c
6111
static int ext4_mb_discard_preallocations(struct super_block *sb, int needed)
fs/ext4/mballoc.c
6113
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/mballoc.c
6118
trace_ext4_mb_discard_preallocations(sb, needed);
fs/ext4/mballoc.c
6121
needed = EXT4_CLUSTERS_PER_GROUP(sb) + 1;
fs/ext4/mballoc.c
6124
ret = ext4_mb_discard_group_preallocations(sb, i, &busy);
fs/ext4/mballoc.c
6138
static bool ext4_mb_discard_preallocations_should_retry(struct super_block *sb,
fs/ext4/mballoc.c
6145
freed = ext4_mb_discard_preallocations(sb, ac->ac_o_ex.fe_len);
fs/ext4/mballoc.c
6158
mb_debug(sb, "freed %d, retry ? %s\n", freed, str_yes_no(ret));
fs/ext4/mballoc.c
6171
struct super_block *sb = ar->inode->i_sb;
fs/ext4/mballoc.c
6172
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
6175
ext4_grpblk_t max = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc.c
6186
ext4_get_group_no_and_offset(sb, goal, &group, &blkoff);
fs/ext4/mballoc.c
6187
for (nr = ext4_get_groups_count(sb); nr > 0; nr--) {
fs/ext4/mballoc.c
6188
bitmap_bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
620
static void mb_group_bb_bitmap_alloc(struct super_block *sb,
fs/ext4/mballoc.c
6200
if (ext4_fc_replay_check_excluded(sb,
fs/ext4/mballoc.c
6201
ext4_group_first_block_no(sb, group) +
fs/ext4/mballoc.c
6211
if (++group >= ext4_get_groups_count(sb))
fs/ext4/mballoc.c
6222
block = ext4_group_first_block_no(sb, group) + EXT4_C2B(sbi, i);
fs/ext4/mballoc.c
6223
ext4_mb_mark_bb(sb, block, 1, true);
fs/ext4/mballoc.c
6240
struct super_block *sb;
fs/ext4/mballoc.c
6248
sb = ar->inode->i_sb;
fs/ext4/mballoc.c
6249
sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
625
grp->bb_bitmap = kmalloc(sb->s_blocksize, GFP_NOFS);
fs/ext4/mballoc.c
6272
ext4_mb_show_pa(sb);
fs/ext4/mballoc.c
629
bh = ext4_read_block_bitmap(sb, group);
fs/ext4/mballoc.c
6339
block = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
fs/ext4/mballoc.c
6344
ext4_mb_discard_preallocations_should_retry(sb, ac, &seq))
fs/ext4/mballoc.c
636
memcpy(grp->bb_bitmap, bh->b_data, sb->s_blocksize);
fs/ext4/mballoc.c
6447
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
6448
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
6477
ext4_grp_locked_error(sb, group, 0,
fs/ext4/mballoc.c
6478
ext4_group_first_block_no(sb, group) +
fs/ext4/mballoc.c
6516
struct super_block *sb = inode->i_sb;
fs/ext4/mballoc.c
6520
ext4_get_group_no_and_offset(sb, block, &group, &blkoff);
fs/ext4/mballoc.c
6521
ext4_mb_mark_context(NULL, sb, false, group, blkoff, count,
fs/ext4/mballoc.c
6540
struct super_block *sb = inode->i_sb;
fs/ext4/mballoc.c
6552
sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
6556
ext4_error(sb, "Freeing blocks in system zone - "
fs/ext4/mballoc.c
6565
ext4_get_group_no_and_offset(sb, block, &block_group, &bit);
fs/ext4/mballoc.c
6567
grp = ext4_get_group_info(sb, block_group);
fs/ext4/mballoc.c
6575
if (EXT4_C2B(sbi, bit) + count > EXT4_BLOCKS_PER_GROUP(sb)) {
fs/ext4/mballoc.c
6577
EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/mballoc.c
6583
trace_ext4_mballoc_free(sb, inode, block_group, bit, count_clusters);
fs/ext4/mballoc.c
6586
err = ext4_mb_load_buddy_gfp(sb, block_group, &e4b,
fs/ext4/mballoc.c
6593
ext4_error(sb, "Freeing blocks in system zone - "
fs/ext4/mballoc.c
6602
err = ext4_mb_mark_context(handle, sb, false, block_group, bit,
fs/ext4/mballoc.c
661
static inline void mb_group_bb_bitmap_alloc(struct super_block *sb,
fs/ext4/mballoc.c
6634
ext4_lock_group(sb, block_group);
fs/ext4/mballoc.c
6637
if (test_opt(sb, DISCARD)) {
fs/ext4/mballoc.c
6638
err = ext4_issue_discard(sb, block_group, bit,
fs/ext4/mballoc.c
6647
ext4_msg(sb, KERN_WARNING, "discard request in"
fs/ext4/mballoc.c
6655
ext4_lock_group(sb, block_group);
fs/ext4/mballoc.c
6659
ext4_unlock_group(sb, block_group);
fs/ext4/mballoc.c
6685
ext4_std_error(sb, err);
fs/ext4/mballoc.c
6701
struct super_block *sb = inode->i_sb;
fs/ext4/mballoc.c
6705
sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
6723
ext4_error(sb, "Freeing blocks not in datazone - "
fs/ext4/mballoc.c
6800
int ext4_group_add_blocks(handle_t *handle, struct super_block *sb,
fs/ext4/mballoc.c
6805
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
6818
ext4_get_group_no_and_offset(sb, block, &block_group, &bit);
fs/ext4/mballoc.c
6823
if (bit + cluster_count > EXT4_CLUSTERS_PER_GROUP(sb)) {
fs/ext4/mballoc.c
6824
ext4_warning(sb, "too many blocks added to group %u",
fs/ext4/mballoc.c
6830
err = ext4_mb_load_buddy(sb, block_group, &e4b);
fs/ext4/mballoc.c
6834
if (!ext4_sb_block_valid(sb, NULL, block, count)) {
fs/ext4/mballoc.c
6835
ext4_error(sb, "Adding blocks in system zones - "
fs/ext4/mballoc.c
6842
err = ext4_mb_mark_context(handle, sb, false, block_group, bit,
fs/ext4/mballoc.c
6849
ext4_error(sb, "bit already cleared in group %u", block_group);
fs/ext4/mballoc.c
6851
ext4_lock_group(sb, block_group);
fs/ext4/mballoc.c
6853
ext4_unlock_group(sb, block_group);
fs/ext4/mballoc.c
6860
ext4_std_error(sb, err);
fs/ext4/mballoc.c
6875
static int ext4_trim_extent(struct super_block *sb,
fs/ext4/mballoc.c
6884
trace_ext4_trim_extent(sb, group, start, count);
fs/ext4/mballoc.c
6886
assert_spin_locked(ext4_group_lock_ptr(sb, group));
fs/ext4/mballoc.c
6897
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
6898
ret = ext4_issue_discard(sb, group, start, count);
fs/ext4/mballoc.c
6899
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
6904
static ext4_grpblk_t ext4_last_grp_cluster(struct super_block *sb,
fs/ext4/mballoc.c
6909
if (grp < (ext4_get_groups_count(sb) - 1))
fs/ext4/mballoc.c
6910
nr_clusters_in_group = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/mballoc.c
6912
nr_clusters_in_group = (ext4_blocks_count(EXT4_SB(sb)->s_es) -
fs/ext4/mballoc.c
6913
ext4_group_first_block_no(sb, grp))
fs/ext4/mballoc.c
6914
>> EXT4_CLUSTER_BITS(sb);
fs/ext4/mballoc.c
6924
static int ext4_try_to_trim_range(struct super_block *sb,
fs/ext4/mballoc.c
6927
__acquires(ext4_group_lock_ptr(sb, e4b->bd_group))
fs/ext4/mballoc.c
6928
__releases(ext4_group_lock_ptr(sb, e4b->bd_group))
fs/ext4/mballoc.c
6937
last = ext4_last_grp_cluster(sb, e4b->bd_group);
fs/ext4/mballoc.c
6956
int ret = ext4_trim_extent(sb, start, next - start, e4b);
fs/ext4/mballoc.c
6969
ext4_unlock_group(sb, e4b->bd_group);
fs/ext4/mballoc.c
6971
ext4_lock_group(sb, e4b->bd_group);
fs/ext4/mballoc.c
6998
ext4_trim_all_free(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
7005
trace_ext4_trim_all_free(sb, group, start, max);
fs/ext4/mballoc.c
7007
ret = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
7009
ext4_warning(sb, "Error %d loading buddy information for %u",
fs/ext4/mballoc.c
7014
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7017
minblocks < EXT4_SB(sb)->s_last_trim_minblks)
fs/ext4/mballoc.c
7018
ret = ext4_try_to_trim_range(sb, &e4b, start, max, minblocks);
fs/ext4/mballoc.c
7022
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7043
int ext4_trim_fs(struct super_block *sb, struct fstrim_range *range)
fs/ext4/mballoc.c
7045
unsigned int discard_granularity = bdev_discard_granularity(sb->s_bdev);
fs/ext4/mballoc.c
7051
le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block);
fs/ext4/mballoc.c
7052
ext4_fsblk_t max_blks = ext4_blocks_count(EXT4_SB(sb)->s_es);
fs/ext4/mballoc.c
7055
start = range->start >> sb->s_blocksize_bits;
fs/ext4/mballoc.c
7056
end = start + (range->len >> sb->s_blocksize_bits) - 1;
fs/ext4/mballoc.c
7057
minlen = EXT4_NUM_B2C(EXT4_SB(sb),
fs/ext4/mballoc.c
7058
range->minlen >> sb->s_blocksize_bits);
fs/ext4/mballoc.c
706
struct super_block *sb = e4b->bd_sb;
fs/ext4/mballoc.c
7060
if (minlen > EXT4_CLUSTERS_PER_GROUP(sb) ||
fs/ext4/mballoc.c
7062
range->len < sb->s_blocksize)
fs/ext4/mballoc.c
7066
minlen = EXT4_NUM_B2C(EXT4_SB(sb),
fs/ext4/mballoc.c
7067
discard_granularity >> sb->s_blocksize_bits);
fs/ext4/mballoc.c
7068
if (minlen > EXT4_CLUSTERS_PER_GROUP(sb))
fs/ext4/mballoc.c
7079
ext4_get_group_no_and_offset(sb, (ext4_fsblk_t) start,
fs/ext4/mballoc.c
7081
ext4_get_group_no_and_offset(sb, (ext4_fsblk_t) end,
fs/ext4/mballoc.c
7085
end = EXT4_CLUSTERS_PER_GROUP(sb) - 1;
fs/ext4/mballoc.c
7090
grp = ext4_get_group_info(sb, group);
fs/ext4/mballoc.c
7095
ret = ext4_mb_init_group(sb, group, GFP_NOFS);
fs/ext4/mballoc.c
7109
cnt = ext4_trim_all_free(sb, group, first_cluster,
fs/ext4/mballoc.c
7126
EXT4_SB(sb)->s_last_trim_minblks = minlen;
fs/ext4/mballoc.c
7129
range->len = EXT4_C2B(EXT4_SB(sb), trimmed) << sb->s_blocksize_bits;
fs/ext4/mballoc.c
7136
struct super_block *sb,
fs/ext4/mballoc.c
7149
error = ext4_mb_load_buddy(sb, group, &e4b);
fs/ext4/mballoc.c
7154
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7157
if (end >= EXT4_CLUSTERS_PER_GROUP(sb))
fs/ext4/mballoc.c
7158
end = EXT4_CLUSTERS_PER_GROUP(sb) - 1;
fs/ext4/mballoc.c
7162
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7163
error = meta_formatter(sb, group, first, start - first,
fs/ext4/mballoc.c
7167
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7175
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7176
error = formatter(sb, group, start, next - start, priv);
fs/ext4/mballoc.c
7179
ext4_lock_group(sb, group);
fs/ext4/mballoc.c
7184
ext4_unlock_group(sb, group);
fs/ext4/mballoc.c
7243
void ext4_mb_generate_buddy_test(struct super_block *sb, void *buddy,
fs/ext4/mballoc.c
7247
ext4_mb_generate_buddy(sb, buddy, bitmap, group, grp);
fs/ext4/mballoc.c
7251
int ext4_mb_load_buddy_test(struct super_block *sb, ext4_group_t group,
fs/ext4/mballoc.c
7254
return ext4_mb_load_buddy(sb, group, e4b);
fs/ext4/mballoc.c
779
grp = ext4_get_group_info(sb, e4b->bd_group);
fs/ext4/mballoc.c
788
ext4_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &k);
fs/ext4/mballoc.c
807
static void ext4_mb_mark_free_simple(struct super_block *sb,
fs/ext4/mballoc.c
811
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
817
BUG_ON(len > EXT4_CLUSTERS_PER_GROUP(sb));
fs/ext4/mballoc.c
819
border = 2 << sb->s_blocksize_bits;
fs/ext4/mballoc.c
843
static int mb_avg_fragment_size_order(struct super_block *sb, ext4_grpblk_t len)
fs/ext4/mballoc.c
854
if (order == MB_NUM_ORDERS(sb))
fs/ext4/mballoc.c
856
if (WARN_ON_ONCE(order > MB_NUM_ORDERS(sb)))
fs/ext4/mballoc.c
857
order = MB_NUM_ORDERS(sb) - 1;
fs/ext4/mballoc.c
863
mb_update_avg_fragment_size(struct super_block *sb, struct ext4_group_info *grp)
fs/ext4/mballoc.c
865
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.c
868
if (!test_opt2(sb, MB_OPTIMIZE_SCAN))
fs/ext4/mballoc.c
873
mb_avg_fragment_size_order(sb, grp->bb_free / grp->bb_fragments);
fs/ext4/mballoc.c
890
mb_debug(sb, "insert group: %u to s_mb_avg_fragment_size[%d] failed, err %d",
fs/ext4/mballoc.c
914
struct super_block *sb = ac->ac_sb;
fs/ext4/mballoc.c
915
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mballoc.h
235
static inline ext4_fsblk_t ext4_grp_offs_to_block(struct super_block *sb,
fs/ext4/mballoc.h
238
return ext4_group_first_block_no(sb, fex->fe_group) +
fs/ext4/mballoc.h
239
(fex->fe_start << EXT4_SB(sb)->s_cluster_bits);
fs/ext4/mballoc.h
257
struct super_block *sb,
fs/ext4/mballoc.h
265
struct super_block *sb,
fs/ext4/mballoc.h
274
struct super_block *sb, bool state,
fs/ext4/mballoc.h
292
extern void ext4_mb_generate_buddy_test(struct super_block *sb,
fs/ext4/mballoc.h
295
extern int ext4_mb_load_buddy_test(struct super_block *sb,
fs/ext4/mballoc.h
30
#define mb_debug(sb, fmt, ...) \
fs/ext4/mballoc.h
32
current->comm, task_pid_nr(current), sb->s_id, \
fs/ext4/mballoc.h
35
#define mb_debug(sb, fmt, ...) no_printk(fmt, ##__VA_ARGS__)
fs/ext4/mballoc.h
98
#define MB_NUM_ORDERS(sb) ((sb)->s_blocksize_bits + 2)
fs/ext4/migrate.c
185
struct super_block *sb = inode->i_sb;
fs/ext4/migrate.c
189
bh = ext4_sb_bread(sb, le32_to_cpu(i_data), 0);
fs/ext4/migrate.c
198
ext4_free_metadata_revoke_credits(sb, 1));
fs/ext4/migrate.c
211
ext4_free_metadata_revoke_credits(sb, 1));
fs/ext4/mmp.c
102
if (!ext4_mmp_csum_verify(sb, mmp)) {
fs/ext4/mmp.c
11
static __le32 ext4_mmp_csum(struct super_block *sb, struct mmp_struct *mmp)
fs/ext4/mmp.c
110
ext4_warning(sb, "Error %d while reading MMP block %llu",
fs/ext4/mmp.c
118
void __dump_mmp_msg(struct super_block *sb, struct mmp_struct *mmp,
fs/ext4/mmp.c
121
__ext4_warning(sb, function, line, "%s", msg);
fs/ext4/mmp.c
122
__ext4_warning(sb, function, line,
fs/ext4/mmp.c
13
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/mmp.c
134
struct super_block *sb = data;
fs/ext4/mmp.c
135
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/mmp.c
136
struct buffer_head *bh = EXT4_SB(sb)->s_mmp_bh;
fs/ext4/mmp.c
161
while (!kthread_should_stop() && !ext4_emergency_state(sb)) {
fs/ext4/mmp.c
162
if (!ext4_has_feature_mmp(sb)) {
fs/ext4/mmp.c
163
ext4_warning(sb, "kmmpd being stopped since MMP feature"
fs/ext4/mmp.c
174
retval = write_mmp_block(sb, bh);
fs/ext4/mmp.c
181
ext4_error_err(sb, -retval,
fs/ext4/mmp.c
202
retval = read_mmp_block(sb, &bh_check, mmp_block);
fs/ext4/mmp.c
204
ext4_error_err(sb, -retval,
fs/ext4/mmp.c
214
dump_mmp_msg(sb, mmp_check,
fs/ext4/mmp.c
218
ext4_error_err(sb, EBUSY, "abort");
fs/ext4/mmp.c
22
static int ext4_mmp_csum_verify(struct super_block *sb, struct mmp_struct *mmp)
fs/ext4/mmp.c
24
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/mmp.c
242
retval = write_mmp_block(sb, bh);
fs/ext4/mmp.c
27
return mmp->mmp_checksum == ext4_mmp_csum(sb, mmp);
fs/ext4/mmp.c
275
int ext4_multi_mount_protect(struct super_block *sb,
fs/ext4/mmp.c
278
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/mmp.c
288
ext4_warning(sb, "Invalid MMP block in superblock");
fs/ext4/mmp.c
293
retval = read_mmp_block(sb, &bh, mmp_block);
fs/ext4/mmp.c
30
static void ext4_mmp_csum_set(struct super_block *sb, struct mmp_struct *mmp)
fs/ext4/mmp.c
314
dump_mmp_msg(sb, mmp, "fsck is running on the filesystem");
fs/ext4/mmp.c
32
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/mmp.c
324
ext4_warning(sb, "MMP interval %u higher than expected, please"
fs/ext4/mmp.c
328
ext4_warning(sb, "MMP startup interrupted, failing mount\n");
fs/ext4/mmp.c
333
retval = read_mmp_block(sb, &bh, mmp_block);
fs/ext4/mmp.c
338
dump_mmp_msg(sb, mmp,
fs/ext4/mmp.c
35
mmp->mmp_checksum = ext4_mmp_csum(sb, mmp);
fs/ext4/mmp.c
355
retval = write_mmp_block_thawed(sb, bh);
fs/ext4/mmp.c
363
ext4_warning(sb, "MMP startup interrupted, failing mount");
fs/ext4/mmp.c
368
retval = read_mmp_block(sb, &bh, mmp_block);
fs/ext4/mmp.c
373
dump_mmp_msg(sb, mmp,
fs/ext4/mmp.c
379
EXT4_SB(sb)->s_mmp_bh = bh;
fs/ext4/mmp.c
388
EXT4_SB(sb)->s_mmp_tsk = kthread_run(kmmpd, sb, "kmmpd-%.*s",
fs/ext4/mmp.c
391
if (IS_ERR(EXT4_SB(sb)->s_mmp_tsk)) {
fs/ext4/mmp.c
392
EXT4_SB(sb)->s_mmp_tsk = NULL;
fs/ext4/mmp.c
393
ext4_warning(sb, "Unable to create kmmpd thread for %s.",
fs/ext4/mmp.c
394
sb->s_id);
fs/ext4/mmp.c
42
static int write_mmp_block_thawed(struct super_block *sb,
fs/ext4/mmp.c
47
ext4_mmp_csum_set(sb, mmp);
fs/ext4/mmp.c
58
static int write_mmp_block(struct super_block *sb, struct buffer_head *bh)
fs/ext4/mmp.c
64
scoped_guard(super_write, sb)
fs/ext4/mmp.c
65
return write_mmp_block_thawed(sb, bh);
fs/ext4/mmp.c
72
static int read_mmp_block(struct super_block *sb, struct buffer_head **bh,
fs/ext4/mmp.c
85
*bh = sb_getblk(sb, mmp_block);
fs/ext4/move_extent.c
419
struct super_block *sb = orig_inode->i_sb;
fs/ext4/move_extent.c
442
if (ext4_has_feature_bigalloc(sb)) {
fs/ext4/move_extent.c
443
ext4_msg(sb, KERN_ERR,
fs/ext4/move_extent.c
449
ext4_msg(sb, KERN_ERR,
fs/ext4/move_extent.c
460
ext4_msg(sb, KERN_ERR,
fs/ext4/move_extent.c
466
ext4_msg(sb, KERN_ERR,
fs/ext4/move_extent.c
474
ext4_msg(sb, KERN_ERR,
fs/ext4/move_extent.c
580
struct super_block *sb = orig_inode->i_sb;
fs/ext4/move_extent.c
581
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/move_extent.c
634
ext4_should_retry_alloc(sb, &retries))
fs/ext4/namei.c
1493
struct super_block *sb = dir->i_sb;
fs/ext4/namei.c
1500
ext4_rec_len_from_disk(de->rec_len, sb->s_blocksize) ==
fs/ext4/namei.c
1501
sb->s_blocksize)
fs/ext4/namei.c
1522
struct super_block *sb;
fs/ext4/namei.c
1535
sb = dir->i_sb;
fs/ext4/namei.c
1578
nblocks = dir->i_size >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/namei.c
1655
nblocks = dir->i_size >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/namei.c
2362
struct super_block *sb;
fs/ext4/namei.c
2373
sb = dir->i_sb;
fs/ext4/namei.c
2374
blocksize = sb->s_blocksize;
fs/ext4/namei.c
2401
if (ext4_has_feature_metadata_csum(sb)) {
fs/ext4/namei.c
2413
blocks = dir->i_size >> sb->s_blocksize_bits;
fs/ext4/namei.c
2432
ext4_has_feature_dir_index(sb)) {
fs/ext4/namei.c
2472
struct super_block *sb = dir->i_sb;
fs/ext4/namei.c
2492
err = ext4_journal_get_write_access(handle, sb, bh, EXT4_JTR_NONE);
fs/ext4/namei.c
2525
if (add_level && levels == ext4_dir_htree_level(sb)) {
fs/ext4/namei.c
2526
ext4_warning(sb, "Directory (ino: %lu) index full, "
fs/ext4/namei.c
2529
if (ext4_dir_htree_level(sb) < EXT4_HTREE_LEVEL) {
fs/ext4/namei.c
2530
ext4_warning(sb, "Large directory feature is "
fs/ext4/namei.c
2546
node2->fake.rec_len = ext4_rec_len_to_disk(sb->s_blocksize,
fs/ext4/namei.c
2547
sb->s_blocksize);
fs/ext4/namei.c
2549
err = ext4_journal_get_write_access(handle, sb, frame->bh,
fs/ext4/namei.c
2562
err = ext4_journal_get_write_access(handle, sb,
fs/ext4/namei.c
3061
struct super_block *sb;
fs/ext4/namei.c
3072
sb = inode->i_sb;
fs/ext4/namei.c
3091
offset = ext4_rec_len_from_disk(de->rec_len, sb->s_blocksize);
fs/ext4/namei.c
3092
de = ext4_next_entry(de, sb->s_blocksize);
fs/ext4/namei.c
3101
offset += ext4_rec_len_from_disk(de->rec_len, sb->s_blocksize);
fs/ext4/namei.c
3103
if (!(offset & (sb->s_blocksize - 1))) {
fs/ext4/namei.c
3106
lblock = offset >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/namei.c
3109
offset += sb->s_blocksize;
fs/ext4/namei.c
3116
(offset & (sb->s_blocksize - 1)));
fs/ext4/namei.c
3123
offset += ext4_rec_len_from_disk(de->rec_len, sb->s_blocksize);
fs/ext4/namei.c
3994
struct super_block *sb = old.inode->i_sb;
fs/ext4/namei.c
3998
if (test_opt2(sb, JOURNAL_FAST_COMMIT) &&
fs/ext4/namei.c
3999
!(EXT4_SB(sb)->s_mount_state & EXT4_FC_REPLAY) &&
fs/ext4/namei.c
4000
!(ext4_test_mount_flag(sb, EXT4_MF_FC_INELIGIBLE))) {
fs/ext4/orphan.c
103
struct super_block *sb = inode->i_sb;
fs/ext4/orphan.c
104
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/orphan.c
137
err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh,
fs/ext4/orphan.c
157
ext4_superblock_csum_set(sb);
fs/ext4/orphan.c
186
ext4_std_error(sb, err);
fs/ext4/orphan.c
312
static int ext4_quota_on_mount(struct super_block *sb, int type)
fs/ext4/orphan.c
314
return dquot_quota_on_mount(sb,
fs/ext4/orphan.c
315
rcu_dereference_protected(EXT4_SB(sb)->s_qf_names[type],
fs/ext4/orphan.c
316
lockdep_is_held(&sb->s_umount)),
fs/ext4/orphan.c
317
EXT4_SB(sb)->s_jquota_fmt, type);
fs/ext4/orphan.c
324
struct super_block *sb = inode->i_sb;
fs/ext4/orphan.c
329
if (test_opt(sb, DEBUG))
fs/ext4/orphan.c
330
ext4_msg(sb, KERN_DEBUG,
fs/ext4/orphan.c
350
if (test_opt(sb, DEBUG))
fs/ext4/orphan.c
351
ext4_msg(sb, KERN_DEBUG,
fs/ext4/orphan.c
378
void ext4_orphan_cleanup(struct super_block *sb, struct ext4_super_block *es)
fs/ext4/orphan.c
380
unsigned int s_flags = sb->s_flags;
fs/ext4/orphan.c
388
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
389
int inodes_per_ob = ext4_inodes_per_orphan_block(sb);
fs/ext4/orphan.c
396
if (bdev_read_only(sb->s_bdev)) {
fs/ext4/orphan.c
397
ext4_msg(sb, KERN_ERR, "write access "
fs/ext4/orphan.c
403
if (!ext4_feature_set_ok(sb, 0)) {
fs/ext4/orphan.c
404
ext4_msg(sb, KERN_INFO, "Skipping orphan cleanup due to "
fs/ext4/orphan.c
409
if (EXT4_SB(sb)->s_mount_state & EXT4_ERROR_FS) {
fs/ext4/orphan.c
412
ext4_msg(sb, KERN_INFO, "Errors on filesystem, "
fs/ext4/orphan.c
421
ext4_msg(sb, KERN_INFO, "orphan cleanup on readonly fs");
fs/ext4/orphan.c
422
sb->s_flags &= ~SB_RDONLY;
fs/ext4/orphan.c
429
if (ext4_has_feature_quota(sb) && (s_flags & SB_RDONLY)) {
fs/ext4/orphan.c
430
int ret = ext4_enable_quotas(sb);
fs/ext4/orphan.c
435
ext4_msg(sb, KERN_ERR,
fs/ext4/orphan.c
441
if (EXT4_SB(sb)->s_qf_names[i]) {
fs/ext4/orphan.c
442
int ret = ext4_quota_on_mount(sb, i);
fs/ext4/orphan.c
447
ext4_msg(sb, KERN_ERR,
fs/ext4/orphan.c
459
if (EXT4_SB(sb)->s_mount_state & EXT4_ERROR_FS) {
fs/ext4/orphan.c
465
inode = ext4_orphan_get(sb, le32_to_cpu(es->s_last_orphan));
fs/ext4/orphan.c
471
list_add(&EXT4_I(inode)->i_orphan, &EXT4_SB(sb)->s_orphan);
fs/ext4/orphan.c
480
inode = ext4_orphan_get(sb, le32_to_cpu(bdata[j]));
fs/ext4/orphan.c
492
ext4_msg(sb, KERN_INFO, "%d orphan inode%s deleted",
fs/ext4/orphan.c
495
ext4_msg(sb, KERN_INFO, "%d truncate%s cleaned up",
fs/ext4/orphan.c
501
if (sb_dqopt(sb)->files[i])
fs/ext4/orphan.c
502
dquot_quota_off(sb, i);
fs/ext4/orphan.c
506
sb->s_flags = s_flags; /* Restore SB_RDONLY status */
fs/ext4/orphan.c
509
void ext4_release_orphan_info(struct super_block *sb)
fs/ext4/orphan.c
512
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
522
struct super_block *sb,
fs/ext4/orphan.c
525
return (struct ext4_orphan_block_tail *)(bh->b_data + sb->s_blocksize -
fs/ext4/orphan.c
529
static int ext4_orphan_file_block_csum_verify(struct super_block *sb,
fs/ext4/orphan.c
533
int inodes_per_ob = ext4_inodes_per_orphan_block(sb);
fs/ext4/orphan.c
534
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
538
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/orphan.c
541
ot = ext4_orphan_block_tail(sb, bh);
fs/ext4/orphan.c
554
struct super_block *sb = EXT4_TRIGGER(triggers)->sb;
fs/ext4/orphan.c
556
int inodes_per_ob = ext4_inodes_per_orphan_block(sb);
fs/ext4/orphan.c
557
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
564
ot = ext4_orphan_block_tail(sb, bh);
fs/ext4/orphan.c
568
int ext4_init_orphan_info(struct super_block *sb)
fs/ext4/orphan.c
570
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
576
int inodes_per_ob = ext4_inodes_per_orphan_block(sb);
fs/ext4/orphan.c
578
ino_t orphan_ino = le32_to_cpu(EXT4_SB(sb)->s_es->s_orphan_file_inum);
fs/ext4/orphan.c
580
if (!ext4_has_feature_orphan_file(sb))
fs/ext4/orphan.c
583
inode = ext4_iget(sb, orphan_ino, EXT4_IGET_SPECIAL);
fs/ext4/orphan.c
585
ext4_msg(sb, KERN_ERR, "get orphan inode failed");
fs/ext4/orphan.c
594
ext4_msg(sb, KERN_ERR, "orphan file too big: %llu",
fs/ext4/orphan.c
599
oi->of_blocks = inode->i_size >> sb->s_blocksize_bits;
fs/ext4/orphan.c
616
ot = ext4_orphan_block_tail(sb, oi->of_binfo[i].ob_bh);
fs/ext4/orphan.c
618
ext4_error(sb, "orphan file block %d: bad magic", i);
fs/ext4/orphan.c
622
if (!ext4_orphan_file_block_csum_verify(sb,
fs/ext4/orphan.c
624
ext4_error(sb, "orphan file block %d: bad checksum", i);
fs/ext4/orphan.c
646
int ext4_orphan_file_empty(struct super_block *sb)
fs/ext4/orphan.c
648
struct ext4_orphan_info *oi = &EXT4_SB(sb)->s_orphan_info;
fs/ext4/orphan.c
650
int inodes_per_ob = ext4_inodes_per_orphan_block(sb);
fs/ext4/orphan.c
652
if (!ext4_has_feature_orphan_file(sb))
fs/ext4/page-io.c
180
struct super_block *sb = inode->i_sb;
fs/ext4/page-io.c
197
if (test_opt(sb, DATA_ERR_ABORT))
fs/ext4/page-io.c
198
jbd2_journal_abort(EXT4_SB(sb)->s_journal, ret);
fs/ext4/page-io.c
202
if (ret < 0 && !ext4_emergency_state(sb) &&
fs/ext4/page-io.c
204
ext4_msg(sb, KERN_EMERG,
fs/ext4/resize.c
100
clear_bit_unlock(EXT4_FLAGS_RESIZING, &EXT4_SB(sb)->s_ext4_flags);
fs/ext4/resize.c
1000
EXT4_SB(sb)->s_gdb_count++;
fs/ext4/resize.c
1021
struct super_block *sb = inode->i_sb;
fs/ext4/resize.c
1022
int reserved_gdb =le16_to_cpu(EXT4_SB(sb)->s_es->s_reserved_gdt_blocks);
fs/ext4/resize.c
1023
int cluster_bits = EXT4_SB(sb)->s_cluster_bits;
fs/ext4/resize.c
103
return ext4_update_overhead(sb, true);
fs/ext4/resize.c
1038
dind = ext4_sb_bread(sb, le32_to_cpu(*data), 0);
fs/ext4/resize.c
1045
blk = EXT4_SB(sb)->s_sbh->b_blocknr + 1 + EXT4_SB(sb)->s_gdb_count;
fs/ext4/resize.c
1046
data = (__le32 *)dind->b_data + (EXT4_SB(sb)->s_gdb_count %
fs/ext4/resize.c
1047
EXT4_ADDR_PER_BLOCK(sb));
fs/ext4/resize.c
1048
end = (__le32 *)dind->b_data + EXT4_ADDR_PER_BLOCK(sb);
fs/ext4/resize.c
1053
ext4_warning(sb, "reserved block %llu"
fs/ext4/resize.c
1060
primary[res] = ext4_sb_bread(sb, blk, 0);
fs/ext4/resize.c
1066
gdbackups = verify_reserved_gdb(sb, group, primary[res]);
fs/ext4/resize.c
107
static ext4_grpblk_t ext4_group_overhead_blocks(struct super_block *sb,
fs/ext4/resize.c
1078
if ((err = ext4_journal_get_write_access(handle, sb, primary[i],
fs/ext4/resize.c
1090
blk = group * EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/resize.c
110
overhead = ext4_bg_num_gdb(sb, group);
fs/ext4/resize.c
1100
inode->i_blocks += reserved_gdb * sb->s_blocksize >> (9 - cluster_bits);
fs/ext4/resize.c
111
if (ext4_bg_has_super(sb, group))
fs/ext4/resize.c
1114
static inline void ext4_set_block_group_nr(struct super_block *sb, char *data,
fs/ext4/resize.c
1120
if (ext4_has_feature_metadata_csum(sb))
fs/ext4/resize.c
113
le16_to_cpu(EXT4_SB(sb)->s_es->s_reserved_gdt_blocks);
fs/ext4/resize.c
1140
static void update_backups(struct super_block *sb, sector_t blk_off, char *data,
fs/ext4/resize.c
1143
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1145
const int bpg = EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/resize.c
1150
int rest = sb->s_blocksize - size;
fs/ext4/resize.c
1154
handle = ext4_journal_start_sb(sb, EXT4_HT_RESIZE, EXT4_MAX_TRANS_DATA);
fs/ext4/resize.c
1162
group = ext4_list_backups(sb, &three, &five, &seven);
fs/ext4/resize.c
1165
group = ext4_get_group_number(sb, blk_off) + 1;
fs/ext4/resize.c
1166
last = (ext4_group_t)(group + EXT4_DESC_PER_BLOCK(sb) - 2);
fs/ext4/resize.c
1172
int has_super = ext4_bg_has_super(sb, group);
fs/ext4/resize.c
1173
ext4_fsblk_t first_block = ext4_group_first_block_no(sb, group);
fs/ext4/resize.c
1185
bh = sb_getblk(sb, backup_block);
fs/ext4/resize.c
1192
ext4_group_first_block_no(sb, group));
fs/ext4/resize.c
1194
if ((err = ext4_journal_get_write_access(handle, sb, bh,
fs/ext4/resize.c
120
static int verify_group_input(struct super_block *sb,
fs/ext4/resize.c
1204
ext4_set_block_group_nr(sb, bh->b_data, group);
fs/ext4/resize.c
1209
ext4_std_error(sb, err);
fs/ext4/resize.c
1213
group = ext4_list_backups(sb, &three, &five, &seven);
fs/ext4/resize.c
123
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1234
ext4_warning(sb, "can't update backup for group %u (err %d), "
fs/ext4/resize.c
1252
static int ext4_add_new_descs(handle_t *handle, struct super_block *sb,
fs/ext4/resize.c
1256
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1262
meta_bg = ext4_has_feature_meta_bg(sb);
fs/ext4/resize.c
1264
int reserved_gdb = ext4_bg_has_super(sb, group) ?
fs/ext4/resize.c
1267
gdb_off = group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1268
gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1280
err = ext4_journal_get_write_access(handle, sb, gdb_bh,
fs/ext4/resize.c
1283
if (!err && reserved_gdb && ext4_bg_num_gdb(sb, group))
fs/ext4/resize.c
1286
err = add_new_gdb_meta_bg(sb, handle, group);
fs/ext4/resize.c
1296
static struct buffer_head *ext4_get_bitmap(struct super_block *sb, __u64 block)
fs/ext4/resize.c
1298
struct buffer_head *bh = sb_getblk(sb, block);
fs/ext4/resize.c
1311
static int ext4_set_bitmap_checksums(struct super_block *sb,
fs/ext4/resize.c
1317
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/resize.c
1320
bh = ext4_get_bitmap(sb, group_data->inode_bitmap);
fs/ext4/resize.c
1323
ext4_inode_bitmap_csum_set(sb, gdp, bh);
fs/ext4/resize.c
1326
bh = ext4_get_bitmap(sb, group_data->block_bitmap);
fs/ext4/resize.c
1329
ext4_block_bitmap_csum_set(sb, gdp, bh);
fs/ext4/resize.c
1338
static int ext4_setup_new_descs(handle_t *handle, struct super_block *sb,
fs/ext4/resize.c
1343
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1353
gdb_off = group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1354
gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
136
ext4_warning(sb, "Cannot add at group %u (only %u groups)",
fs/ext4/resize.c
1362
gdb_off * EXT4_DESC_SIZE(sb));
fs/ext4/resize.c
1364
memset(gdp, 0, EXT4_DESC_SIZE(sb));
fs/ext4/resize.c
1365
ext4_block_bitmap_set(sb, gdp, group_data->block_bitmap);
fs/ext4/resize.c
1366
ext4_inode_bitmap_set(sb, gdp, group_data->inode_bitmap);
fs/ext4/resize.c
1367
err = ext4_set_bitmap_checksums(sb, gdp, group_data);
fs/ext4/resize.c
1369
ext4_std_error(sb, err);
fs/ext4/resize.c
1373
ext4_inode_table_set(sb, gdp, group_data->inode_table);
fs/ext4/resize.c
1374
ext4_free_group_clusters_set(sb, gdp,
fs/ext4/resize.c
1376
ext4_free_inodes_set(sb, gdp, EXT4_INODES_PER_GROUP(sb));
fs/ext4/resize.c
1377
if (ext4_has_group_desc_csum(sb))
fs/ext4/resize.c
1378
ext4_itable_unused_set(sb, gdp,
fs/ext4/resize.c
1379
EXT4_INODES_PER_GROUP(sb));
fs/ext4/resize.c
1381
ext4_group_desc_csum_set(sb, group, gdp);
fs/ext4/resize.c
1385
ext4_std_error(sb, err);
fs/ext4/resize.c
1393
err = ext4_mb_add_groupinfo(sb, group, gdp);
fs/ext4/resize.c
1400
static void ext4_add_overhead(struct super_block *sb,
fs/ext4/resize.c
1403
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
141
overhead = ext4_group_overhead_blocks(sb, group);
fs/ext4/resize.c
1418
static void ext4_update_super(struct super_block *sb,
fs/ext4/resize.c
1425
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1453
le32_add_cpu(&es->s_inodes_count, EXT4_INODES_PER_GROUP(sb) *
fs/ext4/resize.c
1455
le32_add_cpu(&es->s_free_inodes_count, EXT4_INODES_PER_GROUP(sb) *
fs/ext4/resize.c
147
if (test_opt(sb, DEBUG))
fs/ext4/resize.c
1482
(EXT4_MAX_BLOCK_FILE_PHYS / EXT4_BLOCKS_PER_GROUP(sb)));
fs/ext4/resize.c
1493
EXT4_INODES_PER_GROUP(sb) * flex_gd->count);
fs/ext4/resize.c
1497
if (ext4_has_feature_flex_bg(sb) && sbi->s_log_groups_per_flex) {
fs/ext4/resize.c
150
ext4_bg_has_super(sb, input->group) ? "normal" :
fs/ext4/resize.c
1505
atomic_add(EXT4_INODES_PER_GROUP(sb) * flex_gd->count,
fs/ext4/resize.c
1516
if (ext4_has_feature_bigalloc(sb) && (sbi->s_overhead != 0))
fs/ext4/resize.c
1517
ext4_add_overhead(sb,
fs/ext4/resize.c
1520
ext4_calculate_overhead(sb);
fs/ext4/resize.c
1523
ext4_superblock_csum_set(sb);
fs/ext4/resize.c
1525
if (test_opt(sb, DEBUG))
fs/ext4/resize.c
1535
static int ext4_flex_group_add(struct super_block *sb,
fs/ext4/resize.c
1539
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
154
ext4_get_group_no_and_offset(sb, start, NULL, &offset);
fs/ext4/resize.c
1552
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
1555
err = setup_new_flex_group_blocks(sb, flex_gd);
fs/ext4/resize.c
156
ext4_warning(sb, "Last group not full");
fs/ext4/resize.c
1567
credit += 1 + DIV_ROUND_UP(flex_gd->count, EXT4_DESC_PER_BLOCK(sb));
fs/ext4/resize.c
1569
handle = ext4_journal_start_sb(sb, EXT4_HT_RESIZE, credit);
fs/ext4/resize.c
1576
err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh,
fs/ext4/resize.c
158
ext4_warning(sb, "Reserved blocks too high (%u)",
fs/ext4/resize.c
1583
err = ext4_add_new_descs(handle, sb, group,
fs/ext4/resize.c
1588
err = ext4_setup_new_descs(handle, sb, flex_gd);
fs/ext4/resize.c
1592
ext4_update_super(sb, flex_gd);
fs/ext4/resize.c
1602
int gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1604
EXT4_DESC_PER_BLOCK(sb));
fs/ext4/resize.c
1605
int meta_bg = ext4_has_feature_meta_bg(sb) &&
fs/ext4/resize.c
1608
ext4_group_first_block_no(sb, 0);
fs/ext4/resize.c
161
ext4_warning(sb, "Bad blocks count %u",
fs/ext4/resize.c
1610
update_backups(sb, ext4_group_first_block_no(sb, 0),
fs/ext4/resize.c
1617
update_backups(sb, gdb_bh->b_blocknr - padding_blocks,
fs/ext4/resize.c
1625
static int ext4_setup_next_flex_gd(struct super_block *sb,
fs/ext4/resize.c
1629
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
163
else if (IS_ERR(bh = ext4_sb_bread(sb, end - 1, 0))) {
fs/ext4/resize.c
1640
clusters_per_group = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/resize.c
1647
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
1649
ext4_get_group_no_and_offset(sb, n_blocks_count - 1, &n_group, &last);
fs/ext4/resize.c
166
ext4_warning(sb, "Cannot read last block (%llu)",
fs/ext4/resize.c
1661
group_data[i].blocks_count = EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/resize.c
1662
overhead = ext4_group_overhead_blocks(sb, group + i);
fs/ext4/resize.c
1664
group_data[i].free_clusters_count = EXT4_CLUSTERS_PER_GROUP(sb);
fs/ext4/resize.c
1665
if (ext4_has_group_desc_csum(sb)) {
fs/ext4/resize.c
1668
if (!test_opt(sb, INIT_INODE_TABLE))
fs/ext4/resize.c
1674
if (last_group == n_group && ext4_has_group_desc_csum(sb))
fs/ext4/resize.c
169
ext4_warning(sb, "Block bitmap not in group (block %llu)",
fs/ext4/resize.c
1700
int ext4_group_add(struct super_block *sb, struct ext4_new_group_data *input)
fs/ext4/resize.c
1703
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1705
int reserved_gdb = ext4_bg_has_super(sb, input->group) ?
fs/ext4/resize.c
1712
gdb_off = input->group % EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1714
if (gdb_off == 0 && !ext4_has_feature_sparse_super(sb)) {
fs/ext4/resize.c
1715
ext4_warning(sb, "Can't resize non-sparse filesystem further");
fs/ext4/resize.c
172
ext4_warning(sb, "Inode bitmap not in group (block %llu)",
fs/ext4/resize.c
1721
ext4_warning(sb, "blocks_count overflow");
fs/ext4/resize.c
1725
if (le32_to_cpu(es->s_inodes_count) + EXT4_INODES_PER_GROUP(sb) <
fs/ext4/resize.c
1727
ext4_warning(sb, "inodes_count overflow");
fs/ext4/resize.c
1732
if (!ext4_has_feature_resize_inode(sb) ||
fs/ext4/resize.c
1734
ext4_warning(sb,
fs/ext4/resize.c
1738
inode = ext4_iget(sb, EXT4_RESIZE_INO, EXT4_IGET_SPECIAL);
fs/ext4/resize.c
1740
ext4_warning(sb, "Error opening resize inode");
fs/ext4/resize.c
1746
err = verify_group_input(sb, input);
fs/ext4/resize.c
1750
err = ext4_alloc_flex_bg_array(sb, input->group + 1);
fs/ext4/resize.c
1754
err = ext4_mb_alloc_groupinfo(sb, input->group + 1);
fs/ext4/resize.c
176
ext4_warning(sb, "Inode table not in group (blocks %llu-%llu)",
fs/ext4/resize.c
1761
err = ext4_flex_group_add(sb, inode, &flex_gd);
fs/ext4/resize.c
1770
static int ext4_group_extend_no_check(struct super_block *sb,
fs/ext4/resize.c
1773
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/resize.c
1780
handle = ext4_journal_start_sb(sb, EXT4_HT_RESIZE, 3);
fs/ext4/resize.c
1783
ext4_warning(sb, "error %d on journal start", err);
fs/ext4/resize.c
1787
BUFFER_TRACE(EXT4_SB(sb)->s_sbh, "get_write_access");
fs/ext4/resize.c
1788
err = ext4_journal_get_write_access(handle, sb, EXT4_SB(sb)->s_sbh,
fs/ext4/resize.c
179
ext4_warning(sb, "Block bitmap same as inode bitmap (%llu)",
fs/ext4/resize.c
1791
ext4_warning(sb, "error %d on journal write access", err);
fs/ext4/resize.c
1795
lock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
1798
ext4_superblock_csum_set(sb);
fs/ext4/resize.c
1799
unlock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
1803
err = ext4_group_add_blocks(handle, sb, o_blocks_count, add);
fs/ext4/resize.c
1806
ext4_handle_dirty_metadata(handle, NULL, EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
1815
if (test_opt(sb, DEBUG))
fs/ext4/resize.c
1818
update_backups(sb, ext4_group_first_block_no(sb, 0),
fs/ext4/resize.c
182
ext4_warning(sb, "Block bitmap (%llu) in inode table "
fs/ext4/resize.c
1834
int ext4_group_extend(struct super_block *sb, struct ext4_super_block *es,
fs/ext4/resize.c
1845
if (test_opt(sb, DEBUG))
fs/ext4/resize.c
1846
ext4_msg(sb, KERN_DEBUG,
fs/ext4/resize.c
1853
if (n_blocks_count > (sector_t)(~0ULL) >> (sb->s_blocksize_bits - 9)) {
fs/ext4/resize.c
1854
ext4_msg(sb, KERN_ERR,
fs/ext4/resize.c
1861
ext4_warning(sb, "can't shrink FS - resize aborted");
fs/ext4/resize.c
1866
ext4_get_group_no_and_offset(sb, o_blocks_count, &group, &last);
fs/ext4/resize.c
1869
ext4_warning(sb, "need to use ext2online to resize further");
fs/ext4/resize.c
187
ext4_warning(sb, "Inode bitmap (%llu) in inode table "
fs/ext4/resize.c
1873
add = EXT4_BLOCKS_PER_GROUP(sb) - last;
fs/ext4/resize.c
1876
ext4_warning(sb, "blocks_count overflow");
fs/ext4/resize.c
1884
ext4_warning(sb, "will only finish group (%llu blocks, %u new)",
fs/ext4/resize.c
1888
bh = ext4_sb_bread(sb, o_blocks_count + add - 1, 0);
fs/ext4/resize.c
1890
ext4_warning(sb, "can't read last block, resize aborted");
fs/ext4/resize.c
1895
return ext4_group_extend_no_check(sb, o_blocks_count, add);
fs/ext4/resize.c
1899
static int num_desc_blocks(struct super_block *sb, ext4_group_t groups)
fs/ext4/resize.c
1901
return (groups + EXT4_DESC_PER_BLOCK(sb) - 1) / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
1909
static int ext4_convert_meta_bg(struct super_block *sb, struct inode *inode)
fs/ext4/resize.c
1912
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
1919
ext4_msg(sb, KERN_INFO, "Converting file system to meta_bg");
fs/ext4/resize.c
192
ext4_warning(sb, "Block bitmap (%llu) in GDT table (%llu-%llu)",
fs/ext4/resize.c
1922
ext4_error(sb, "Unexpected non-zero "
fs/ext4/resize.c
1944
handle = ext4_journal_start_sb(sb, EXT4_HT_RESIZE, credits);
fs/ext4/resize.c
1949
err = ext4_journal_get_write_access(handle, sb, sbi->s_sbh,
fs/ext4/resize.c
1955
ext4_clear_feature_resize_inode(sb);
fs/ext4/resize.c
1956
ext4_set_feature_meta_bg(sb);
fs/ext4/resize.c
1958
cpu_to_le32(num_desc_blocks(sb, sbi->s_groups_count));
fs/ext4/resize.c
1959
ext4_superblock_csum_set(sb);
fs/ext4/resize.c
196
ext4_warning(sb, "Inode bitmap (%llu) in GDT table (%llu-%llu)",
fs/ext4/resize.c
1964
ext4_std_error(sb, err);
fs/ext4/resize.c
1978
ext4_std_error(sb, err);
fs/ext4/resize.c
1986
ext4_error(sb, "corrupted/inconsistent resize inode");
fs/ext4/resize.c
1996
int ext4_resize_fs(struct super_block *sb, ext4_fsblk_t n_blocks_count)
fs/ext4/resize.c
1999
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
201
ext4_warning(sb, "Inode table (%llu-%llu) overlaps GDT table "
fs/ext4/resize.c
2016
bh = ext4_sb_bread(sb, n_blocks_count - 1, 0);
fs/ext4/resize.c
2018
ext4_warning(sb, "can't read last block, resize aborted");
fs/ext4/resize.c
2030
if (ext4_has_feature_bigalloc(sb))
fs/ext4/resize.c
2031
n_blocks_count &= ~((1 << EXT4_CLUSTER_BITS(sb)) - 1);
fs/ext4/resize.c
2036
ext4_msg(sb, KERN_INFO, "resizing filesystem from %llu "
fs/ext4/resize.c
2041
ext4_warning(sb, "can't shrink FS - resize aborted");
fs/ext4/resize.c
2049
n_group = ext4_get_group_number(sb, n_blocks_count - 1);
fs/ext4/resize.c
2050
if (n_group >= (0xFFFFFFFFUL / EXT4_INODES_PER_GROUP(sb))) {
fs/ext4/resize.c
2051
ext4_warning(sb, "resize would cause inodes_count overflow");
fs/ext4/resize.c
2054
ext4_get_group_no_and_offset(sb, o_blocks_count - 1, &o_group, &offset);
fs/ext4/resize.c
2056
n_desc_blocks = num_desc_blocks(sb, n_group + 1);
fs/ext4/resize.c
2057
o_desc_blocks = num_desc_blocks(sb, sbi->s_groups_count);
fs/ext4/resize.c
2059
meta_bg = ext4_has_feature_meta_bg(sb);
fs/ext4/resize.c
2061
if (ext4_has_feature_resize_inode(sb)) {
fs/ext4/resize.c
2063
ext4_error(sb, "resize_inode and meta_bg enabled "
fs/ext4/resize.c
2072
n_group = n_desc_blocks * EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
2074
EXT4_BLOCKS_PER_GROUP(sb) +
fs/ext4/resize.c
2080
resize_inode = ext4_iget(sb, EXT4_RESIZE_INO,
fs/ext4/resize.c
2083
ext4_warning(sb, "Error opening resize inode");
fs/ext4/resize.c
2089
err = ext4_convert_meta_bg(sb, resize_inode);
fs/ext4/resize.c
2110
if ((ext4_group_first_block_no(sb, n_group) +
fs/ext4/resize.c
2111
ext4_group_overhead_blocks(sb, n_group) + 2 +
fs/ext4/resize.c
2113
n_blocks_count = ext4_group_first_block_no(sb, n_group);
fs/ext4/resize.c
2127
add = EXT4_C2B(sbi, EXT4_CLUSTERS_PER_GROUP(sb) - (offset + 1));
fs/ext4/resize.c
2129
err = ext4_group_extend_no_check(sb, o_blocks_count, add);
fs/ext4/resize.c
2137
err = ext4_alloc_flex_bg_array(sb, n_group + 1);
fs/ext4/resize.c
2141
err = ext4_mb_alloc_groupinfo(sb, n_group + 1);
fs/ext4/resize.c
2154
while (ext4_setup_next_flex_gd(sb, flex_gd, n_blocks_count)) {
fs/ext4/resize.c
2157
ext4_msg(sb, KERN_INFO,
fs/ext4/resize.c
2162
if (ext4_alloc_group_tables(sb, flex_gd, flexbg_size) != 0)
fs/ext4/resize.c
2164
err = ext4_flex_group_add(sb, resize_inode, flex_gd);
fs/ext4/resize.c
2187
ext4_warning(sb, "error (%d) occurred during "
fs/ext4/resize.c
2189
ext4_msg(sb, KERN_INFO, "resized filesystem to %llu",
fs/ext4/resize.c
303
static int ext4_alloc_group_tables(struct super_block *sb,
fs/ext4/resize.c
331
start_blk = ext4_group_first_block_no(sb, src_group);
fs/ext4/resize.c
334
overhead = ext4_group_overhead_blocks(sb, src_group);
fs/ext4/resize.c
341
overhead = ext4_group_overhead_blocks(sb, src_group);
fs/ext4/resize.c
353
group = ext4_get_group_number(sb, start_blk - 1);
fs/ext4/resize.c
364
group = ext4_get_group_number(sb, start_blk - 1);
fs/ext4/resize.c
372
unsigned int itb = EXT4_SB(sb)->s_itb_per_group;
fs/ext4/resize.c
378
group = ext4_get_group_number(sb, start_blk);
fs/ext4/resize.c
379
next_group_start = ext4_group_first_block_no(sb, group + 1);
fs/ext4/resize.c
391
start_blk += EXT4_SB(sb)->s_itb_per_group;
fs/ext4/resize.c
397
EXT4_NUM_B2C(EXT4_SB(sb),
fs/ext4/resize.c
401
if (test_opt(sb, DEBUG)) {
fs/ext4/resize.c
412
ext4_bg_has_super(sb, group + i) ? "normal" :
fs/ext4/resize.c
422
static struct buffer_head *bclean(handle_t *handle, struct super_block *sb,
fs/ext4/resize.c
428
bh = sb_getblk(sb, blk);
fs/ext4/resize.c
432
err = ext4_journal_get_write_access(handle, sb, bh, EXT4_JTR_NONE);
fs/ext4/resize.c
437
memset(bh->b_data, 0, sb->s_blocksize);
fs/ext4/resize.c
459
static int set_flexbg_block_bitmap(struct super_block *sb, handle_t *handle,
fs/ext4/resize.c
46
int ext4_resize_begin(struct super_block *sb)
fs/ext4/resize.c
463
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
475
group = ext4_get_group_number(sb, EXT4_C2B(sbi, first_cluster));
fs/ext4/resize.c
476
start = EXT4_B2C(sbi, ext4_group_first_block_no(sb, group));
fs/ext4/resize.c
479
count2 = EXT4_CLUSTERS_PER_GROUP(sb) - (first_cluster - start);
fs/ext4/resize.c
48
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
492
bh = sb_getblk(sb, flex_gd->groups[group].block_bitmap);
fs/ext4/resize.c
497
err = ext4_journal_get_write_access(handle, sb, bh,
fs/ext4/resize.c
530
static int setup_new_flex_group_blocks(struct super_block *sb,
fs/ext4/resize.c
533
int group_table_count[] = {1, 1, EXT4_SB(sb)->s_itb_per_group};
fs/ext4/resize.c
536
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/resize.c
550
meta_bg = ext4_has_feature_meta_bg(sb);
fs/ext4/resize.c
553
handle = ext4_journal_start_sb(sb, EXT4_HT_RESIZE, EXT4_MAX_TRANS_DATA);
fs/ext4/resize.c
562
gdblocks = ext4_bg_num_gdb(sb, group);
fs/ext4/resize.c
563
start = ext4_group_first_block_no(sb, group);
fs/ext4/resize.c
565
if (meta_bg == 0 && !ext4_bg_has_super(sb, group))
fs/ext4/resize.c
571
block = start + ext4_bg_has_super(sb, group);
fs/ext4/resize.c
581
gdb = sb_getblk(sb, block);
fs/ext4/resize.c
588
err = ext4_journal_get_write_access(handle, sb, gdb,
fs/ext4/resize.c
59
!ext4_has_feature_resize_inode(sb)) {
fs/ext4/resize.c
60
ext4_error(sb, "resize_inode disabled but reserved GDT blocks non-zero");
fs/ext4/resize.c
609
if (ext4_bg_has_super(sb, group)) {
fs/ext4/resize.c
610
err = sb_issue_zeroout(sb, gdblocks + start + 1,
fs/ext4/resize.c
625
err = sb_issue_zeroout(sb, block, sbi->s_itb_per_group,
fs/ext4/resize.c
640
bh = bclean(handle, sb, block);
fs/ext4/resize.c
645
overhead = ext4_group_overhead_blocks(sb, group);
fs/ext4/resize.c
653
sb->s_blocksize * 8, bh->b_data);
fs/ext4/resize.c
669
bh = bclean(handle, sb, block);
fs/ext4/resize.c
675
ext4_mark_bitmap_end(EXT4_INODES_PER_GROUP(sb),
fs/ext4/resize.c
676
sb->s_blocksize * 8, bh->b_data);
fs/ext4/resize.c
694
err = set_flexbg_block_bitmap(sb, handle,
fs/ext4/resize.c
707
err = set_flexbg_block_bitmap(sb, handle,
fs/ext4/resize.c
71
ext4_warning(sb, "won't resize using backup superblock at %llu",
fs/ext4/resize.c
732
unsigned int ext4_list_backups(struct super_block *sb, unsigned int *three,
fs/ext4/resize.c
735
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/resize.c
740
if (ext4_has_feature_sparse_super2(sb)) {
fs/ext4/resize.c
750
if (!ext4_has_feature_sparse_super(sb)) {
fs/ext4/resize.c
776
static int verify_reserved_gdb(struct super_block *sb,
fs/ext4/resize.c
788
while ((grp = ext4_list_backups(sb, &three, &five, &seven)) < end) {
fs/ext4/resize.c
790
grp * EXT4_BLOCKS_PER_GROUP(sb) + blk){
fs/ext4/resize.c
791
ext4_warning(sb, "reserved GDT %llu"
fs/ext4/resize.c
795
(ext4_fsblk_t)EXT4_BLOCKS_PER_GROUP(sb) +
fs/ext4/resize.c
799
if (++gdbackups > EXT4_ADDR_PER_BLOCK(sb))
fs/ext4/resize.c
81
ext4_warning(sb, "There are errors in the filesystem, "
fs/ext4/resize.c
822
struct super_block *sb = inode->i_sb;
fs/ext4/resize.c
823
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/resize.c
824
unsigned long gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
825
ext4_fsblk_t gdblock = EXT4_SB(sb)->s_sbh->b_blocknr + 1 + gdb_num;
fs/ext4/resize.c
834
if (test_opt(sb, DEBUG))
fs/ext4/resize.c
839
gdb_bh = ext4_sb_bread(sb, gdblock, 0);
fs/ext4/resize.c
843
gdbackups = verify_reserved_gdb(sb, group, gdb_bh);
fs/ext4/resize.c
850
dind = ext4_sb_bread(sb, le32_to_cpu(*data), 0);
fs/ext4/resize.c
858
if (le32_to_cpu(data[gdb_num % EXT4_ADDR_PER_BLOCK(sb)]) != gdblock) {
fs/ext4/resize.c
859
ext4_warning(sb, "new group %u GDT block %llu not reserved",
fs/ext4/resize.c
86
if (ext4_has_feature_sparse_super2(sb)) {
fs/ext4/resize.c
865
BUFFER_TRACE(EXT4_SB(sb)->s_sbh, "get_write_access");
fs/ext4/resize.c
866
err = ext4_journal_get_write_access(handle, sb, EXT4_SB(sb)->s_sbh,
fs/ext4/resize.c
87
ext4_msg(sb, KERN_ERR, "Online resizing not supported with sparse_super2");
fs/ext4/resize.c
872
err = ext4_journal_get_write_access(handle, sb, gdb_bh, EXT4_JTR_NONE);
fs/ext4/resize.c
877
err = ext4_journal_get_write_access(handle, sb, dind, EXT4_JTR_NONE);
fs/ext4/resize.c
879
ext4_std_error(sb, err);
fs/ext4/resize.c
892
ext4_warning(sb, "not enough memory for %lu groups",
fs/ext4/resize.c
906
data[gdb_num % EXT4_ADDR_PER_BLOCK(sb)] = 0;
fs/ext4/resize.c
909
ext4_std_error(sb, err);
fs/ext4/resize.c
912
inode->i_blocks -= (gdbackups + 1) * sb->s_blocksize >>
fs/ext4/resize.c
913
(9 - EXT4_SB(sb)->s_cluster_bits);
fs/ext4/resize.c
915
memset(gdb_bh->b_data, 0, sb->s_blocksize);
fs/ext4/resize.c
918
ext4_std_error(sb, err);
fs/ext4/resize.c
925
o_group_desc = rcu_dereference(EXT4_SB(sb)->s_group_desc);
fs/ext4/resize.c
927
EXT4_SB(sb)->s_gdb_count * sizeof(struct buffer_head *));
fs/ext4/resize.c
930
rcu_assign_pointer(EXT4_SB(sb)->s_group_desc, n_group_desc);
fs/ext4/resize.c
931
EXT4_SB(sb)->s_gdb_count++;
fs/ext4/resize.c
934
lock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
936
ext4_superblock_csum_set(sb);
fs/ext4/resize.c
937
unlock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
938
err = ext4_handle_dirty_metadata(handle, NULL, EXT4_SB(sb)->s_sbh);
fs/ext4/resize.c
940
ext4_std_error(sb, err);
fs/ext4/resize.c
961
static int add_new_gdb_meta_bg(struct super_block *sb,
fs/ext4/resize.c
966
unsigned long gdb_num = group / EXT4_DESC_PER_BLOCK(sb);
fs/ext4/resize.c
969
gdblock = ext4_group_first_block_no(sb, group) +
fs/ext4/resize.c
970
ext4_bg_has_super(sb, group);
fs/ext4/resize.c
971
gdb_bh = ext4_sb_bread(sb, gdblock, 0);
fs/ext4/resize.c
979
ext4_warning(sb, "not enough memory for %lu groups",
fs/ext4/resize.c
98
int ext4_resize_end(struct super_block *sb, bool update_backups)
fs/ext4/resize.c
985
o_group_desc = rcu_dereference(EXT4_SB(sb)->s_group_desc);
fs/ext4/resize.c
987
EXT4_SB(sb)->s_gdb_count * sizeof(struct buffer_head *));
fs/ext4/resize.c
992
err = ext4_journal_get_write_access(handle, sb, gdb_bh, EXT4_JTR_NONE);
fs/ext4/resize.c
999
rcu_assign_pointer(EXT4_SB(sb)->s_group_desc, n_group_desc);
fs/ext4/super.c
1001
static int ext4_warning_ratelimit(struct super_block *sb)
fs/ext4/super.c
1003
atomic_inc(&EXT4_SB(sb)->s_warning_count);
fs/ext4/super.c
1004
return ___ratelimit(&(EXT4_SB(sb)->s_warning_ratelimit_state),
fs/ext4/super.c
1008
void __ext4_warning(struct super_block *sb, const char *function,
fs/ext4/super.c
1014
if (!ext4_warning_ratelimit(sb))
fs/ext4/super.c
1021
sb->s_id, function, line, &vaf);
fs/ext4/super.c
1044
struct super_block *sb, ext4_group_t grp,
fs/ext4/super.c
1053
if (unlikely(ext4_emergency_state(sb)))
fs/ext4/super.c
1056
trace_ext4_error(sb, function, line);
fs/ext4/super.c
1057
if (ext4_error_ratelimit(sb)) {
fs/ext4/super.c
1062
sb->s_id, function, line, grp);
fs/ext4/super.c
1072
if (test_opt(sb, ERRORS_CONT)) {
fs/ext4/super.c
1073
if (test_opt(sb, WARN_ON_ERROR))
fs/ext4/super.c
1075
EXT4_SB(sb)->s_mount_state |= EXT4_ERROR_FS;
fs/ext4/super.c
1076
if (!bdev_read_only(sb->s_bdev)) {
fs/ext4/super.c
1077
save_error_info(sb, EFSCORRUPTED, ino, block, function,
fs/ext4/super.c
1079
schedule_work(&EXT4_SB(sb)->s_sb_upd_work);
fs/ext4/super.c
1083
ext4_unlock_group(sb, grp);
fs/ext4/super.c
1084
ext4_handle_error(sb, false, EFSCORRUPTED, ino, block, function, line);
fs/ext4/super.c
1096
ext4_lock_group(sb, grp);
fs/ext4/super.c
1100
void ext4_mark_group_bitmap_corrupted(struct super_block *sb,
fs/ext4/super.c
1104
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
1105
struct ext4_group_info *grp = ext4_get_group_info(sb, group);
fs/ext4/super.c
1106
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
1125
count = ext4_free_inodes_count(sb, gdp);
fs/ext4/super.c
1132
void ext4_update_dynamic_rev(struct super_block *sb)
fs/ext4/super.c
1134
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/super.c
1139
ext4_warning(sb,
fs/ext4/super.c
1162
static void dump_orphan_list(struct super_block *sb, struct ext4_sb_info *sbi)
fs/ext4/super.c
1166
ext4_msg(sb, KERN_ERR, "sb orphan head is %d",
fs/ext4/super.c
1181
static int ext4_quota_off(struct super_block *sb, int type);
fs/ext4/super.c
1183
static inline void ext4_quotas_off(struct super_block *sb, int type)
fs/ext4/super.c
1189
ext4_quota_off(sb, type);
fs/ext4/super.c
1196
static inline char *get_qf_name(struct super_block *sb,
fs/ext4/super.c
1201
lockdep_is_held(&sb->s_umount));
fs/ext4/super.c
1204
static inline void ext4_quotas_off(struct super_block *sb, int type)
fs/ext4/super.c
1276
static void ext4_put_super(struct super_block *sb)
fs/ext4/super.c
1278
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
1293
ext4_unregister_sysfs(sb);
fs/ext4/super.c
1296
ext4_msg(sb, KERN_INFO, "unmounting filesystem %pU.",
fs/ext4/super.c
1297
&sb->s_uuid);
fs/ext4/super.c
1299
ext4_unregister_li_request(sb);
fs/ext4/super.c
1300
ext4_quotas_off(sb, EXT4_MAXQUOTAS);
fs/ext4/super.c
1303
ext4_release_orphan_info(sb);
fs/ext4/super.c
1309
ext4_abort(sb, -err, "Couldn't clean up the journal");
fs/ext4/super.c
1316
ext4_release_system_zone(sb);
fs/ext4/super.c
1317
ext4_mb_release(sb);
fs/ext4/super.c
1318
ext4_ext_release(sb);
fs/ext4/super.c
1320
if (!ext4_emergency_state(sb) && !sb_rdonly(sb)) {
fs/ext4/super.c
1322
ext4_clear_feature_journal_needs_recovery(sb);
fs/ext4/super.c
1323
ext4_clear_feature_orphan_present(sb);
fs/ext4/super.c
1326
ext4_commit_super(sb);
fs/ext4/super.c
1337
kfree(get_qf_name(sb, sbi, i));
fs/ext4/super.c
1345
dump_orphan_list(sb, sbi);
fs/ext4/super.c
1348
sync_blockdev(sb->s_bdev);
fs/ext4/super.c
1349
invalidate_bdev(sb->s_bdev);
fs/ext4/super.c
1369
sb->s_fs_info = NULL;
fs/ext4/super.c
1380
utf8_unload(sb->s_encoding);
fs/ext4/super.c
1390
static struct inode *ext4_alloc_inode(struct super_block *sb)
fs/ext4/super.c
1394
ei = alloc_inode_sb(sb, ext4_inode_cachep, GFP_NOFS);
fs/ext4/super.c
144
#define IS_EXT2_SB(sb) ((sb)->s_type == &ext2_fs_type)
fs/ext4/super.c
146
#define IS_EXT2_SB(sb) (0)
fs/ext4/super.c
1471
static void ext4_shutdown(struct super_block *sb)
fs/ext4/super.c
1473
ext4_force_shutdown(sb, EXT4_GOING_FLAGS_NOLOGFLUSH);
fs/ext4/super.c
1558
static struct inode *ext4_nfs_get_inode(struct super_block *sb,
fs/ext4/super.c
1567
inode = ext4_iget(sb, ino, EXT4_IGET_HANDLE);
fs/ext4/super.c
1578
static struct dentry *ext4_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/ext4/super.c
1581
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/ext4/super.c
1585
static struct dentry *ext4_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/ext4/super.c
1588
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/ext4/super.c
160
#define IS_EXT3_SB(sb) ((sb)->s_type == &ext3_fs_type)
fs/ext4/super.c
1610
static int ext4_write_info(struct super_block *sb, int type);
fs/ext4/super.c
1611
static int ext4_quota_on(struct super_block *sb, int type, int format_id,
fs/ext4/super.c
1613
static ssize_t ext4_quota_read(struct super_block *sb, int type, char *data,
fs/ext4/super.c
1615
static ssize_t ext4_quota_write(struct super_block *sb, int type,
fs/ext4/super.c
1617
static int ext4_quota_enable(struct super_block *sb, int type, int format_id,
fs/ext4/super.c
230
static struct buffer_head *__ext4_sb_bread_gfp(struct super_block *sb,
fs/ext4/super.c
237
bh = sb_getblk_gfp(sb, block, gfp);
fs/ext4/super.c
2501
static int parse_apply_sb_mount_options(struct super_block *sb,
fs/ext4/super.c
2504
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
251
struct buffer_head *ext4_sb_bread(struct super_block *sb, sector_t block,
fs/ext4/super.c
2531
ret = ext4_check_opt_consistency(fc, sb);
fs/ext4/super.c
2534
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
254
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
2546
ext4_apply_options(fc, sb);
fs/ext4/super.c
2556
struct super_block *sb)
fs/ext4/super.c
2559
bool quota_feature = ext4_has_feature_quota(sb);
fs/ext4/super.c
2561
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
257
return __ext4_sb_bread_gfp(sb, block, op_flags, gfp);
fs/ext4/super.c
2575
set_opt(sb, QUOTA);
fs/ext4/super.c
2578
lockdep_is_held(&sb->s_umount));
fs/ext4/super.c
2593
struct super_block *sb)
fs/ext4/super.c
2597
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
2598
bool quota_feature = ext4_has_feature_quota(sb);
fs/ext4/super.c
2599
bool quota_loaded = sb_any_quota_loaded(sb);
fs/ext4/super.c
260
struct buffer_head *ext4_sb_bread_unmovable(struct super_block *sb,
fs/ext4/super.c
2609
!ext4_has_feature_project(sb)) {
fs/ext4/super.c
263
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
2633
strcmp(get_qf_name(sb, sbi, i),
fs/ext4/super.c
2657
usr_qf_name = (get_qf_name(sb, sbi, USRQUOTA) ||
fs/ext4/super.c
2659
grp_qf_name = (get_qf_name(sb, sbi, GRPQUOTA) ||
fs/ext4/super.c
266
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/ext4/super.c
2663
test_opt(sb, USRQUOTA));
fs/ext4/super.c
2666
test_opt(sb, GRPQUOTA));
fs/ext4/super.c
269
struct buffer_head *ext4_sb_bread_nofail(struct super_block *sb,
fs/ext4/super.c
2711
struct super_block *sb)
fs/ext4/super.c
2714
const struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
2719
if (!ext4_has_feature_encrypt(sb)) {
fs/ext4/super.c
272
gfp_t gfp = mapping_gfp_constraint(sb->s_bdev->bd_mapping,
fs/ext4/super.c
275
return __ext4_sb_bread_gfp(sb, block, 0, gfp);
fs/ext4/super.c
2751
struct super_block *sb)
fs/ext4/super.c
2755
fscrypt_is_dummy_policy_set(&EXT4_SB(sb)->s_dummy_enc_policy))
fs/ext4/super.c
2757
EXT4_SB(sb)->s_dummy_enc_policy = ctx->dummy_enc_policy;
fs/ext4/super.c
2759
ext4_msg(sb, KERN_WARNING, "Test dummy encryption mode enabled");
fs/ext4/super.c
2763
struct super_block *sb)
fs/ext4/super.c
2770
if ((ctx->opt_flags & MOPT_NO_EXT2) && IS_EXT2_SB(sb)) {
fs/ext4/super.c
2775
if ((ctx->opt_flags & MOPT_NO_EXT3) && IS_EXT3_SB(sb)) {
fs/ext4/super.c
278
void ext4_sb_breadahead_unmovable(struct super_block *sb, sector_t block)
fs/ext4/super.c
2789
err = ext4_check_test_dummy_encryption(fc, sb);
fs/ext4/super.c
280
struct buffer_head *bh = bdev_getblk(sb->s_bdev, block,
fs/ext4/super.c
2800
test_opt(sb, DATA_FLAGS)) {
fs/ext4/super.c
281
sb->s_blocksize, GFP_NOWAIT);
fs/ext4/super.c
2816
(test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA)) {
fs/ext4/super.c
2841
return ext4_check_quota_consistency(fc, sb);
fs/ext4/super.c
2844
static void ext4_apply_options(struct fs_context *fc, struct super_block *sb)
fs/ext4/super.c
2853
sb->s_flags &= ~ctx->mask_s_flags;
fs/ext4/super.c
2854
sb->s_flags |= ctx->vals_s_flags;
fs/ext4/super.c
2872
ext4_apply_quota_options(fc, sb);
fs/ext4/super.c
2873
ext4_apply_test_dummy_encryption(ctx, sb);
fs/ext4/super.c
290
static int ext4_verify_csum_type(struct super_block *sb,
fs/ext4/super.c
2905
struct super_block *sb)
fs/ext4/super.c
2908
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
293
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/super.c
2954
static int _ext4_show_options(struct seq_file *seq, struct super_block *sb,
fs/ext4/super.c
2957
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3005
if (test_opt(sb, ERRORS_RO) && def_errors != EXT4_ERRORS_RO)
fs/ext4/super.c
3007
if (test_opt(sb, ERRORS_CONT) && def_errors != EXT4_ERRORS_CONTINUE)
fs/ext4/super.c
3009
if (test_opt(sb, ERRORS_PANIC) && def_errors != EXT4_ERRORS_PANIC)
fs/ext4/super.c
3017
if (nodefs && sb->s_flags & SB_I_VERSION)
fs/ext4/super.c
3023
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA)
fs/ext4/super.c
3025
else if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_ORDERED_DATA)
fs/ext4/super.c
3027
else if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_WRITEBACK_DATA)
fs/ext4/super.c
3035
if (test_opt(sb, INIT_INODE_TABLE) && (nodefs ||
fs/ext4/super.c
3040
if (test_opt(sb, DATA_ERR_ABORT))
fs/ext4/super.c
3043
fscrypt_show_test_dummy_encryption(seq, sep, sb);
fs/ext4/super.c
3045
if (sb->s_flags & SB_INLINECRYPT)
fs/ext4/super.c
3048
if (test_opt(sb, DAX_ALWAYS)) {
fs/ext4/super.c
3049
if (IS_EXT2_SB(sb))
fs/ext4/super.c
3053
} else if (test_opt2(sb, DAX_NEVER)) {
fs/ext4/super.c
3055
} else if (test_opt2(sb, DAX_INODE)) {
fs/ext4/super.c
3060
!test_opt2(sb, MB_OPTIMIZE_SCAN)) {
fs/ext4/super.c
3063
test_opt2(sb, MB_OPTIMIZE_SCAN)) {
fs/ext4/super.c
3067
if (nodefs && !test_opt(sb, NO_PREFETCH_BLOCK_BITMAPS))
fs/ext4/super.c
3070
if (ext4_emergency_ro(sb))
fs/ext4/super.c
3073
if (ext4_forced_shutdown(sb))
fs/ext4/super.c
3076
ext4_show_quota_options(seq, sb);
fs/ext4/super.c
3087
struct super_block *sb = seq->private;
fs/ext4/super.c
309
static int ext4_superblock_csum_verify(struct super_block *sb,
fs/ext4/super.c
3090
seq_puts(seq, sb_rdonly(sb) ? "ro" : "rw");
fs/ext4/super.c
3091
rc = _ext4_show_options(seq, sb, 1);
fs/ext4/super.c
3096
static int ext4_setup_super(struct super_block *sb, struct ext4_super_block *es,
fs/ext4/super.c
3099
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3103
ext4_msg(sb, KERN_ERR, "revision level too high, "
fs/ext4/super.c
3111
ext4_msg(sb, KERN_WARNING, "warning: mounting unchecked fs, "
fs/ext4/super.c
3114
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
312
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/super.c
3120
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
3126
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
3136
ext4_set_feature_journal_needs_recovery(sb);
fs/ext4/super.c
3137
if (ext4_has_feature_orphan_file(sb))
fs/ext4/super.c
3138
ext4_set_feature_orphan_present(sb);
fs/ext4/super.c
3141
err = ext4_commit_super(sb);
fs/ext4/super.c
3143
if (test_opt(sb, DEBUG))
fs/ext4/super.c
3146
sb->s_blocksize,
fs/ext4/super.c
3148
EXT4_BLOCKS_PER_GROUP(sb),
fs/ext4/super.c
3149
EXT4_INODES_PER_GROUP(sb),
fs/ext4/super.c
3154
int ext4_alloc_flex_bg_array(struct super_block *sb, ext4_group_t ngroup)
fs/ext4/super.c
3156
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3170
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
318
void ext4_superblock_csum_set(struct super_block *sb)
fs/ext4/super.c
3182
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
320
struct ext4_super_block *es = EXT4_SB(sb)->s_es;
fs/ext4/super.c
3201
static int ext4_fill_flex_info(struct super_block *sb)
fs/ext4/super.c
3203
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3215
err = ext4_alloc_flex_bg_array(sb, sbi->s_groups_count);
fs/ext4/super.c
322
if (!ext4_has_feature_metadata_csum(sb))
fs/ext4/super.c
3220
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/super.c
3224
atomic_add(ext4_free_inodes_count(sb, gdp), &fg->free_inodes);
fs/ext4/super.c
3225
atomic64_add(ext4_free_group_clusters(sb, gdp),
fs/ext4/super.c
3227
atomic_add(ext4_used_dirs_count(sb, gdp), &fg->used_dirs);
fs/ext4/super.c
3235
static __le16 ext4_group_desc_csum(struct super_block *sb, __u32 block_group,
fs/ext4/super.c
3241
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3263
if (!ext4_has_feature_gdt_csum(sb))
fs/ext4/super.c
3271
if (ext4_has_feature_64bit(sb) && offset < sbi->s_desc_size)
fs/ext4/super.c
3279
int ext4_group_desc_csum_verify(struct super_block *sb, __u32 block_group,
fs/ext4/super.c
328
ext4_fsblk_t ext4_block_bitmap(struct super_block *sb,
fs/ext4/super.c
3282
if (ext4_has_group_desc_csum(sb) &&
fs/ext4/super.c
3283
(gdp->bg_checksum != ext4_group_desc_csum(sb, block_group, gdp)))
fs/ext4/super.c
3289
void ext4_group_desc_csum_set(struct super_block *sb, __u32 block_group,
fs/ext4/super.c
3292
if (!ext4_has_group_desc_csum(sb))
fs/ext4/super.c
3294
gdp->bg_checksum = ext4_group_desc_csum(sb, block_group, gdp);
fs/ext4/super.c
3298
static int ext4_check_descriptors(struct super_block *sb,
fs/ext4/super.c
3302
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3305
ext4_fsblk_t last_bg_block = sb_block + ext4_bg_num_gdb(sb, 0);
fs/ext4/super.c
3312
if (ext4_has_feature_flex_bg(sb))
fs/ext4/super.c
3318
struct ext4_group_desc *gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/super.c
332
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3324
(EXT4_BLOCKS_PER_GROUP(sb) - 1);
fs/ext4/super.c
3330
block_bitmap = ext4_block_bitmap(sb, gdp);
fs/ext4/super.c
3332
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3335
if (!sb_rdonly(sb))
fs/ext4/super.c
3340
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3343
if (!sb_rdonly(sb))
fs/ext4/super.c
3347
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3352
inode_bitmap = ext4_inode_bitmap(sb, gdp);
fs/ext4/super.c
3354
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3357
if (!sb_rdonly(sb))
fs/ext4/super.c
336
ext4_fsblk_t ext4_inode_bitmap(struct super_block *sb,
fs/ext4/super.c
3362
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3365
if (!sb_rdonly(sb))
fs/ext4/super.c
3369
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3374
inode_table = ext4_inode_table(sb, gdp);
fs/ext4/super.c
3376
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3379
if (!sb_rdonly(sb))
fs/ext4/super.c
3384
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3387
if (!sb_rdonly(sb))
fs/ext4/super.c
3392
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
3397
ext4_lock_group(sb, i);
fs/ext4/super.c
3398
if (!ext4_group_desc_csum_verify(sb, i, gdp)) {
fs/ext4/super.c
3399
ext4_msg(sb, KERN_ERR, "ext4_check_descriptors: "
fs/ext4/super.c
340
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3401
i, le16_to_cpu(ext4_group_desc_csum(sb, i,
fs/ext4/super.c
3403
if (!sb_rdonly(sb)) {
fs/ext4/super.c
3404
ext4_unlock_group(sb, i);
fs/ext4/super.c
3408
ext4_unlock_group(sb, i);
fs/ext4/super.c
3410
first_block += EXT4_BLOCKS_PER_GROUP(sb);
fs/ext4/super.c
344
ext4_fsblk_t ext4_inode_table(struct super_block *sb,
fs/ext4/super.c
348
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
352
__u32 ext4_free_group_clusters(struct super_block *sb,
fs/ext4/super.c
3541
static ext4_fsblk_t descriptor_loc(struct super_block *sb,
fs/ext4/super.c
3544
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
3550
if (!ext4_has_feature_meta_bg(sb) || nr < first_meta_bg)
fs/ext4/super.c
3553
if (ext4_bg_has_super(sb, bg))
fs/ext4/super.c
356
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3562
if (sb->s_blocksize == 1024 && nr == 0 &&
fs/ext4/super.c
3566
return (has_super + ext4_group_first_block_no(sb, bg));
fs/ext4/super.c
360
__u32 ext4_free_inodes_count(struct super_block *sb,
fs/ext4/super.c
3612
int ext4_feature_set_ok(struct super_block *sb, int readonly)
fs/ext4/super.c
3614
if (ext4_has_unknown_ext4_incompat_features(sb)) {
fs/ext4/super.c
3615
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
3618
(le32_to_cpu(EXT4_SB(sb)->s_es->s_feature_incompat) &
fs/ext4/super.c
3623
if (!IS_ENABLED(CONFIG_UNICODE) && ext4_has_feature_casefold(sb)) {
fs/ext4/super.c
3624
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
3633
if (ext4_has_feature_readonly(sb)) {
fs/ext4/super.c
3634
ext4_msg(sb, KERN_INFO, "filesystem is read-only");
fs/ext4/super.c
3635
sb->s_flags |= SB_RDONLY;
fs/ext4/super.c
364
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3640
if (ext4_has_unknown_ext4_ro_compat_features(sb)) {
fs/ext4/super.c
3641
ext4_msg(sb, KERN_ERR, "couldn't mount RDWR because of "
fs/ext4/super.c
3643
(le32_to_cpu(EXT4_SB(sb)->s_es->s_feature_ro_compat) &
fs/ext4/super.c
3647
if (ext4_has_feature_bigalloc(sb) && !ext4_has_feature_extents(sb)) {
fs/ext4/super.c
3648
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
3653
if (ext4_has_feature_bigalloc(sb) &&
fs/ext4/super.c
3654
le32_to_cpu(EXT4_SB(sb)->s_es->s_first_data_block)) {
fs/ext4/super.c
3655
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
3662
if (!readonly && (ext4_has_feature_quota(sb) ||
fs/ext4/super.c
3663
ext4_has_feature_project(sb))) {
fs/ext4/super.c
3664
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
368
__u32 ext4_used_dirs_count(struct super_block *sb,
fs/ext4/super.c
3681
struct super_block *sb = sbi->s_sb;
fs/ext4/super.c
3686
ext4_msg(sb, KERN_NOTICE, "error count since last fsck: %u",
fs/ext4/super.c
3690
sb->s_id,
fs/ext4/super.c
3705
sb->s_id,
fs/ext4/super.c
372
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3727
struct super_block *sb = elr->lr_super;
fs/ext4/super.c
3728
ext4_group_t ngroups = EXT4_SB(sb)->s_groups_count;
fs/ext4/super.c
3732
int nr = EXT4_SB(sb)->s_mb_prefetch;
fs/ext4/super.c
3736
elr->lr_next_group = ext4_mb_prefetch(sb, group, nr, &prefetch_ios);
fs/ext4/super.c
3737
ext4_mb_prefetch_fini(sb, elr->lr_next_group, nr);
fs/ext4/super.c
3738
trace_ext4_prefetch_bitmaps(sb, group, elr->lr_next_group, nr);
fs/ext4/super.c
3742
!ext4_emergency_state(sb) && !sb_rdonly(sb) &&
fs/ext4/super.c
3743
test_opt(sb, INIT_INODE_TABLE)) {
fs/ext4/super.c
3753
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
376
__u32 ext4_itable_unused_count(struct super_block *sb,
fs/ext4/super.c
3768
ret = ext4_init_inode_table(sb, group,
fs/ext4/super.c
3770
trace_ext4_lazy_itable_init(sb, group);
fs/ext4/super.c
3795
static void ext4_unregister_li_request(struct super_block *sb)
fs/ext4/super.c
380
(EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT ?
fs/ext4/super.c
3804
ext4_remove_li_request(EXT4_SB(sb)->s_li_request);
fs/ext4/super.c
384
void ext4_block_bitmap_set(struct super_block *sb,
fs/ext4/super.c
388
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
392
void ext4_inode_bitmap_set(struct super_block *sb,
fs/ext4/super.c
396
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
3965
static ext4_group_t ext4_has_uninit_itable(struct super_block *sb)
fs/ext4/super.c
3967
ext4_group_t group, ngroups = EXT4_SB(sb)->s_groups_count;
fs/ext4/super.c
3970
if (!ext4_has_group_desc_csum(sb))
fs/ext4/super.c
3974
gdp = ext4_get_group_desc(sb, group, NULL);
fs/ext4/super.c
400
void ext4_inode_table_set(struct super_block *sb,
fs/ext4/super.c
4003
static struct ext4_li_request *ext4_li_request_new(struct super_block *sb,
fs/ext4/super.c
4012
elr->lr_super = sb;
fs/ext4/super.c
4014
if (test_opt(sb, NO_PREFETCH_BLOCK_BITMAPS)) {
fs/ext4/super.c
4030
int ext4_register_li_request(struct super_block *sb,
fs/ext4/super.c
4033
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
404
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
4048
if (ext4_emergency_state(sb) || sb_rdonly(sb) ||
fs/ext4/super.c
4049
(test_opt(sb, NO_PREFETCH_BLOCK_BITMAPS) &&
fs/ext4/super.c
4050
(first_not_zeroed == ngroups || !test_opt(sb, INIT_INODE_TABLE))))
fs/ext4/super.c
4053
elr = ext4_li_request_new(sb, first_not_zeroed);
fs/ext4/super.c
408
void ext4_free_group_clusters_set(struct super_block *sb,
fs/ext4/super.c
4105
static int set_journal_csum_feature_set(struct super_block *sb)
fs/ext4/super.c
4109
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4111
if (ext4_has_feature_metadata_csum(sb)) {
fs/ext4/super.c
412
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
4125
if (test_opt(sb, JOURNAL_ASYNC_COMMIT)) {
fs/ext4/super.c
4130
} else if (test_opt(sb, JOURNAL_CHECKSUM)) {
fs/ext4/super.c
4159
static int count_overhead(struct super_block *sb, ext4_group_t grp,
fs/ext4/super.c
416
void ext4_free_inodes_set(struct super_block *sb,
fs/ext4/super.c
4162
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4165
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/super.c
4167
int has_super = ext4_bg_has_super(sb, grp);
fs/ext4/super.c
4169
if (!ext4_has_feature_bigalloc(sb))
fs/ext4/super.c
4170
return (has_super + ext4_bg_num_gdb(sb, grp) +
fs/ext4/super.c
4175
(grp * EXT4_BLOCKS_PER_GROUP(sb));
fs/ext4/super.c
4176
last_block = first_block + EXT4_BLOCKS_PER_GROUP(sb) - 1;
fs/ext4/super.c
4178
gdp = ext4_get_group_desc(sb, i, NULL);
fs/ext4/super.c
4179
b = ext4_block_bitmap(sb, gdp);
fs/ext4/super.c
4184
b = ext4_inode_bitmap(sb, gdp);
fs/ext4/super.c
4189
b = ext4_inode_table(sb, gdp);
fs/ext4/super.c
4199
if (ext4_bg_has_super(sb, grp)) {
fs/ext4/super.c
420
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
4203
j = ext4_bg_num_gdb(sb, grp);
fs/ext4/super.c
4204
if (s + j > EXT4_BLOCKS_PER_GROUP(sb)) {
fs/ext4/super.c
4205
ext4_error(sb, "Invalid number of block group "
fs/ext4/super.c
4207
j = EXT4_BLOCKS_PER_GROUP(sb) - s;
fs/ext4/super.c
4215
return EXT4_CLUSTERS_PER_GROUP(sb) -
fs/ext4/super.c
4216
ext4_count_free(buf, EXT4_CLUSTERS_PER_GROUP(sb) / 8);
fs/ext4/super.c
4222
int ext4_calculate_overhead(struct super_block *sb)
fs/ext4/super.c
4224
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4228
ext4_group_t i, ngroups = ext4_get_groups_count(sb);
fs/ext4/super.c
4230
char *buf = kvmalloc(sb->s_blocksize, GFP_NOFS | __GFP_ZERO);
fs/ext4/super.c
424
void ext4_used_dirs_set(struct super_block *sb,
fs/ext4/super.c
4252
blks = count_overhead(sb, i, buf);
fs/ext4/super.c
4255
memset(buf, 0, sb->s_blocksize);
fs/ext4/super.c
4265
else if (ext4_has_feature_journal(sb) && !sbi->s_journal && j_inum) {
fs/ext4/super.c
4267
j_inode = ext4_get_journal_inode(sb, j_inum);
fs/ext4/super.c
4269
j_blocks = j_inode->i_size >> sb->s_blocksize_bits;
fs/ext4/super.c
4273
ext4_msg(sb, KERN_ERR, "can't get journal size");
fs/ext4/super.c
428
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
4282
static void ext4_set_resv_clusters(struct super_block *sb)
fs/ext4/super.c
4285
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4293
if (!ext4_has_feature_extents(sb))
fs/ext4/super.c
4312
static const char *ext4_quota_mode(struct super_block *sb)
fs/ext4/super.c
4315
if (!ext4_quota_capable(sb))
fs/ext4/super.c
4318
if (EXT4_SB(sb)->s_journal && ext4_is_quota_journalled(sb))
fs/ext4/super.c
432
void ext4_itable_unused_set(struct super_block *sb,
fs/ext4/super.c
4327
static void ext4_setup_csum_trigger(struct super_block *sb,
fs/ext4/super.c
4335
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4337
sbi->s_journal_triggers[type].sb = sb;
fs/ext4/super.c
4351
static struct ext4_sb_info *ext4_alloc_sbi(struct super_block *sb)
fs/ext4/super.c
4359
sbi->s_daxdev = fs_dax_get_by_bdev(sb->s_bdev, &sbi->s_dax_part_off,
fs/ext4/super.c
436
if (EXT4_DESC_SIZE(sb) >= EXT4_MIN_DESC_SIZE_64BIT)
fs/ext4/super.c
4368
sb->s_fs_info = sbi;
fs/ext4/super.c
4369
sbi->s_sb = sb;
fs/ext4/super.c
4377
static void ext4_set_def_opts(struct super_block *sb,
fs/ext4/super.c
4384
set_opt(sb, INIT_INODE_TABLE);
fs/ext4/super.c
4386
set_opt(sb, DEBUG);
fs/ext4/super.c
4388
set_opt(sb, GRPID);
fs/ext4/super.c
4390
set_opt(sb, NO_UID32);
fs/ext4/super.c
4392
set_opt(sb, XATTR_USER);
fs/ext4/super.c
4394
set_opt(sb, POSIX_ACL);
fs/ext4/super.c
4396
if (ext4_has_feature_fast_commit(sb))
fs/ext4/super.c
4397
set_opt2(sb, JOURNAL_FAST_COMMIT);
fs/ext4/super.c
4399
if (ext4_has_feature_metadata_csum(sb))
fs/ext4/super.c
4400
set_opt(sb, JOURNAL_CHECKSUM);
fs/ext4/super.c
4403
set_opt(sb, JOURNAL_DATA);
fs/ext4/super.c
4405
set_opt(sb, ORDERED_DATA);
fs/ext4/super.c
4407
set_opt(sb, WRITEBACK_DATA);
fs/ext4/super.c
4410
set_opt(sb, ERRORS_PANIC);
fs/ext4/super.c
4412
set_opt(sb, ERRORS_CONT);
fs/ext4/super.c
4414
set_opt(sb, ERRORS_RO);
fs/ext4/super.c
4416
set_opt(sb, BLOCK_VALIDITY);
fs/ext4/super.c
4418
set_opt(sb, DISCARD);
fs/ext4/super.c
4421
set_opt(sb, BARRIER);
fs/ext4/super.c
4427
if (!IS_EXT3_SB(sb) && !IS_EXT2_SB(sb) &&
fs/ext4/super.c
4429
set_opt(sb, DELALLOC);
fs/ext4/super.c
4431
set_opt(sb, DIOREAD_NOLOCK);
fs/ext4/super.c
4434
static int ext4_handle_clustersize(struct super_block *sb)
fs/ext4/super.c
4436
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4442
if (ext4_has_feature_bigalloc(sb)) {
fs/ext4/super.c
4443
if (clustersize < sb->s_blocksize) {
fs/ext4/super.c
4444
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4446
"block size (%lu)", clustersize, sb->s_blocksize);
fs/ext4/super.c
4452
if (clustersize != sb->s_blocksize) {
fs/ext4/super.c
4453
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4455
"block size (%lu)", clustersize, sb->s_blocksize);
fs/ext4/super.c
4458
if (sbi->s_blocks_per_group > sb->s_blocksize * 8) {
fs/ext4/super.c
4459
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4467
if (sbi->s_clusters_per_group > sb->s_blocksize * 8) {
fs/ext4/super.c
4468
ext4_msg(sb, KERN_ERR, "#clusters per group too big: %lu",
fs/ext4/super.c
4473
(sbi->s_clusters_per_group * (clustersize / sb->s_blocksize))) {
fs/ext4/super.c
4474
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4479
sbi->s_cluster_ratio = clustersize / sb->s_blocksize;
fs/ext4/super.c
4483
set_opt2(sb, STD_GROUP_SIZE);
fs/ext4/super.c
4495
static void ext4_atomic_write_init(struct super_block *sb)
fs/ext4/super.c
4497
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4498
struct block_device *bdev = sb->s_bdev;
fs/ext4/super.c
4499
unsigned int clustersize = EXT4_CLUSTER_SIZE(sb);
fs/ext4/super.c
4504
if (!ext4_has_feature_extents(sb))
fs/ext4/super.c
4507
sbi->s_awu_min = max(sb->s_blocksize,
fs/ext4/super.c
4513
ext4_msg(sb, KERN_NOTICE, "Supports (experimental) DIO atomic writes awu_min: %u, awu_max: %u",
fs/ext4/super.c
4521
static void ext4_fast_commit_init(struct super_block *sb)
fs/ext4/super.c
4523
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4532
ext4_clear_mount_flag(sb, EXT4_MF_FC_INELIGIBLE);
fs/ext4/super.c
4545
static int ext4_inode_info_init(struct super_block *sb,
fs/ext4/super.c
4548
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4557
ext4_msg(sb, KERN_ERR, "invalid first ino: %u",
fs/ext4/super.c
4563
(sbi->s_inode_size > sb->s_blocksize)) {
fs/ext4/super.c
4564
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4567
ext4_msg(sb, KERN_ERR, "blocksize: %lu", sb->s_blocksize);
fs/ext4/super.c
4578
sb->s_time_gran = 1;
fs/ext4/super.c
4579
sb->s_time_max = EXT4_EXTRA_TIMESTAMP_MAX;
fs/ext4/super.c
4581
sb->s_time_gran = NSEC_PER_SEC;
fs/ext4/super.c
4582
sb->s_time_max = EXT4_NON_EXTRA_TIMESTAMP_MAX;
fs/ext4/super.c
4584
sb->s_time_min = EXT4_TIMESTAMP_MIN;
fs/ext4/super.c
4590
if (ext4_has_feature_extra_isize(sb)) {
fs/ext4/super.c
4596
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4605
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4618
static int ext4_encoding_init(struct super_block *sb, struct ext4_super_block *es)
fs/ext4/super.c
4624
if (!ext4_has_feature_casefold(sb) || sb->s_encoding)
fs/ext4/super.c
4629
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4636
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4646
ext4_msg(sb, KERN_INFO,"Using encoding defined by superblock: "
fs/ext4/super.c
4653
sb->s_encoding = encoding;
fs/ext4/super.c
4654
sb->s_encoding_flags = encoding_flags;
fs/ext4/super.c
4659
static inline int ext4_encoding_init(struct super_block *sb, struct ext4_super_block *es)
fs/ext4/super.c
4665
static int ext4_init_metadata_csum(struct super_block *sb, struct ext4_super_block *es)
fs/ext4/super.c
4667
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4670
if (ext4_has_feature_metadata_csum(sb) &&
fs/ext4/super.c
4671
ext4_has_feature_gdt_csum(sb))
fs/ext4/super.c
4672
ext4_warning(sb, "metadata_csum and uninit_bg are "
fs/ext4/super.c
4676
if (!ext4_verify_csum_type(sb, es)) {
fs/ext4/super.c
4677
ext4_msg(sb, KERN_ERR, "VFS: Found ext4 filesystem with "
fs/ext4/super.c
4681
ext4_setup_csum_trigger(sb, EXT4_JTR_ORPHAN_FILE,
fs/ext4/super.c
4685
if (!ext4_superblock_csum_verify(sb, es)) {
fs/ext4/super.c
4686
ext4_msg(sb, KERN_ERR, "VFS: Found ext4 filesystem with "
fs/ext4/super.c
4692
if (ext4_has_feature_csum_seed(sb))
fs/ext4/super.c
4694
else if (ext4_has_feature_metadata_csum(sb) ||
fs/ext4/super.c
4695
ext4_has_feature_ea_inode(sb))
fs/ext4/super.c
4701
static int ext4_check_feature_compatibility(struct super_block *sb,
fs/ext4/super.c
4705
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4708
(ext4_has_compat_features(sb) ||
fs/ext4/super.c
4709
ext4_has_ro_compat_features(sb) ||
fs/ext4/super.c
4710
ext4_has_incompat_features(sb)))
fs/ext4/super.c
4711
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
4716
set_opt2(sb, HURD_COMPAT);
fs/ext4/super.c
4717
if (ext4_has_feature_64bit(sb)) {
fs/ext4/super.c
4718
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
472
static void ext4_maybe_update_superblock(struct super_block *sb)
fs/ext4/super.c
4727
if (ext4_has_feature_ea_inode(sb)) {
fs/ext4/super.c
4728
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4734
if (IS_EXT2_SB(sb)) {
fs/ext4/super.c
4735
if (ext2_feature_set_ok(sb))
fs/ext4/super.c
4736
ext4_msg(sb, KERN_INFO, "mounting ext2 file system "
fs/ext4/super.c
474
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4743
if (silent && ext4_feature_set_ok(sb, sb_rdonly(sb)))
fs/ext4/super.c
4745
ext4_msg(sb, KERN_ERR, "couldn't mount as ext2 due "
fs/ext4/super.c
4751
if (IS_EXT3_SB(sb)) {
fs/ext4/super.c
4752
if (ext3_feature_set_ok(sb))
fs/ext4/super.c
4753
ext4_msg(sb, KERN_INFO, "mounting ext3 file system "
fs/ext4/super.c
4760
if (silent && ext4_feature_set_ok(sb, sb_rdonly(sb)))
fs/ext4/super.c
4762
ext4_msg(sb, KERN_ERR, "couldn't mount as ext3 due "
fs/ext4/super.c
4773
if (!ext4_feature_set_ok(sb, (sb_rdonly(sb))))
fs/ext4/super.c
4777
if (sb->s_blocksize == PAGE_SIZE)
fs/ext4/super.c
4780
ext4_msg(sb, KERN_ERR, "unsupported blocksize for DAX\n");
fs/ext4/super.c
4784
if (ext4_has_feature_inline_data(sb)) {
fs/ext4/super.c
4785
ext4_msg(sb, KERN_ERR, "Cannot use DAX on a filesystem"
fs/ext4/super.c
4790
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4796
if (ext4_has_feature_encrypt(sb) && es->s_encryption_level) {
fs/ext4/super.c
4797
ext4_msg(sb, KERN_ERR, "Unsupported encryption level %d",
fs/ext4/super.c
4805
static int ext4_check_geometry(struct super_block *sb,
fs/ext4/super.c
4808
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4812
if (le16_to_cpu(sbi->s_es->s_reserved_gdt_blocks) > (sb->s_blocksize / 4)) {
fs/ext4/super.c
4813
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
482
if (ext4_emergency_state(sb) || sb_rdonly(sb) ||
fs/ext4/super.c
4822
err = generic_check_addressable(sb->s_blocksize_bits,
fs/ext4/super.c
4825
ext4_msg(sb, KERN_ERR, "filesystem"
fs/ext4/super.c
483
!(sb->s_flags & SB_ACTIVE) || !journal ||
fs/ext4/super.c
4831
blocks_count = sb_bdev_nr_blocks(sb);
fs/ext4/super.c
4833
ext4_msg(sb, KERN_WARNING, "bad geometry: block count %llu "
fs/ext4/super.c
4844
ext4_msg(sb, KERN_WARNING, "bad geometry: first data "
fs/ext4/super.c
4852
ext4_msg(sb, KERN_WARNING, "bad geometry: first data "
fs/ext4/super.c
4859
EXT4_BLOCKS_PER_GROUP(sb) - 1);
fs/ext4/super.c
4860
do_div(blocks_count, EXT4_BLOCKS_PER_GROUP(sb));
fs/ext4/super.c
4861
if (blocks_count > ((uint64_t)1<<32) - EXT4_DESC_PER_BLOCK(sb)) {
fs/ext4/super.c
4862
ext4_msg(sb, KERN_WARNING, "groups count too large: %llu "
fs/ext4/super.c
4867
EXT4_BLOCKS_PER_GROUP(sb));
fs/ext4/super.c
4872
(EXT4_MAX_BLOCK_FILE_PHYS / EXT4_BLOCKS_PER_GROUP(sb)));
fs/ext4/super.c
4875
ext4_msg(sb, KERN_ERR, "inodes count not valid: %u vs %llu",
fs/ext4/super.c
4884
static int ext4_group_desc_init(struct super_block *sb,
fs/ext4/super.c
4889
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4894
db_count = (sbi->s_groups_count + EXT4_DESC_PER_BLOCK(sb) - 1) /
fs/ext4/super.c
4895
EXT4_DESC_PER_BLOCK(sb);
fs/ext4/super.c
4896
if (ext4_has_feature_meta_bg(sb)) {
fs/ext4/super.c
4898
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
4908
ext4_msg(sb, KERN_ERR, "not enough memory");
fs/ext4/super.c
4916
block = descriptor_loc(sb, logical_sb_block, i);
fs/ext4/super.c
4917
ext4_sb_breadahead_unmovable(sb, block);
fs/ext4/super.c
4923
block = descriptor_loc(sb, logical_sb_block, i);
fs/ext4/super.c
4924
bh = ext4_sb_bread_unmovable(sb, block);
fs/ext4/super.c
4926
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4936
if (!ext4_check_descriptors(sb, logical_sb_block, first_not_zeroed)) {
fs/ext4/super.c
4937
ext4_msg(sb, KERN_ERR, "group descriptors corrupted!");
fs/ext4/super.c
494
((part_stat_read(sb->s_bdev, sectors[STAT_WRITE]) -
fs/ext4/super.c
4944
static int ext4_load_and_init_journal(struct super_block *sb,
fs/ext4/super.c
4948
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
4951
err = ext4_load_journal(sb, es, ctx->journal_devnum);
fs/ext4/super.c
4955
if (ext4_has_feature_64bit(sb) &&
fs/ext4/super.c
4956
!jbd2_journal_set_features(EXT4_SB(sb)->s_journal, 0, 0,
fs/ext4/super.c
4958
ext4_msg(sb, KERN_ERR, "Failed to set 64-bit journal feature");
fs/ext4/super.c
4962
if (!set_journal_csum_feature_set(sb)) {
fs/ext4/super.c
4963
ext4_msg(sb, KERN_ERR, "Failed to set journal checksum "
fs/ext4/super.c
4968
if (test_opt2(sb, JOURNAL_FAST_COMMIT) &&
fs/ext4/super.c
4969
!jbd2_journal_set_features(EXT4_SB(sb)->s_journal, 0, 0,
fs/ext4/super.c
4971
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
4978
switch (test_opt(sb, DATA_FLAGS)) {
fs/ext4/super.c
4986
set_opt(sb, ORDERED_DATA);
fs/ext4/super.c
4989
set_opt(sb, JOURNAL_DATA);
fs/ext4/super.c
4998
ext4_msg(sb, KERN_ERR, "Journal does not support "
fs/ext4/super.c
5007
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_ORDERED_DATA &&
fs/ext4/super.c
5008
test_opt(sb, JOURNAL_ASYNC_COMMIT)) {
fs/ext4/super.c
5009
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
5028
static int ext4_check_journal_data_mode(struct super_block *sb)
fs/ext4/super.c
5030
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA) {
fs/ext4/super.c
5035
clear_opt(sb, DIOREAD_NOLOCK);
fs/ext4/super.c
5036
clear_opt2(sb, JOURNAL_FAST_COMMIT);
fs/ext4/super.c
5037
if (test_opt2(sb, EXPLICIT_DELALLOC)) {
fs/ext4/super.c
5038
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
5042
if (test_opt(sb, DAX_ALWAYS)) {
fs/ext4/super.c
5043
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
5047
if (ext4_has_feature_encrypt(sb)) {
fs/ext4/super.c
5048
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
5052
if (test_opt(sb, DELALLOC))
fs/ext4/super.c
5053
clear_opt(sb, DELALLOC);
fs/ext4/super.c
5055
sb->s_iflags |= SB_I_CGROUPWB;
fs/ext4/super.c
506
schedule_work(&EXT4_SB(sb)->s_sb_upd_work);
fs/ext4/super.c
5061
static const char *ext4_has_journal_option(struct super_block *sb)
fs/ext4/super.c
5063
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5065
if (test_opt(sb, JOURNAL_ASYNC_COMMIT))
fs/ext4/super.c
5067
if (test_opt2(sb, EXPLICIT_JOURNAL_CHECKSUM))
fs/ext4/super.c
5074
if (test_opt(sb, DATA_ERR_ABORT))
fs/ext4/super.c
5084
#define EXT4_MAX_PAGECACHE_ORDER(sb) \
fs/ext4/super.c
5085
umin(MAX_PAGECACHE_ORDER, (11 + (sb)->s_blocksize_bits - PAGE_SHIFT))
fs/ext4/super.c
5086
static void ext4_set_max_mapping_order(struct super_block *sb)
fs/ext4/super.c
5088
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5090
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA)
fs/ext4/super.c
5093
sbi->s_max_folio_order = EXT4_MAX_PAGECACHE_ORDER(sb);
fs/ext4/super.c
5096
static int ext4_check_large_folio(struct super_block *sb)
fs/ext4/super.c
5100
if (ext4_has_feature_encrypt(sb))
fs/ext4/super.c
5104
ext4_set_max_mapping_order(sb);
fs/ext4/super.c
5105
} else if (sb->s_blocksize > PAGE_SIZE) {
fs/ext4/super.c
5106
ext4_msg(sb, KERN_ERR, "bs(%lu) > ps(%lu) unsupported for %s",
fs/ext4/super.c
5107
sb->s_blocksize, PAGE_SIZE, err_str);
fs/ext4/super.c
511
struct super_block *sb = journal->j_private;
fs/ext4/super.c
5114
static int ext4_load_super(struct super_block *sb, ext4_fsblk_t *lsb,
fs/ext4/super.c
5117
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5125
blocksize = sb_min_blocksize(sb, EXT4_MIN_BLOCK_SIZE);
fs/ext4/super.c
5127
ext4_msg(sb, KERN_ERR, "unable to set blocksize");
fs/ext4/super.c
5142
bh = ext4_sb_bread_unmovable(sb, logical_sb_block);
fs/ext4/super.c
5144
ext4_msg(sb, KERN_ERR, "unable to read superblock");
fs/ext4/super.c
515
ext4_process_freed_data(sb, txn->t_tid);
fs/ext4/super.c
5153
sb->s_magic = le16_to_cpu(es->s_magic);
fs/ext4/super.c
5154
if (sb->s_magic != EXT4_SUPER_MAGIC) {
fs/ext4/super.c
5156
ext4_msg(sb, KERN_ERR, "VFS: Can't find ext4 filesystem");
fs/ext4/super.c
516
ext4_maybe_update_superblock(sb);
fs/ext4/super.c
5162
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5169
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5181
if (sb->s_blocksize == blocksize)
fs/ext4/super.c
5191
if (!sb_set_blocksize(sb, blocksize)) {
fs/ext4/super.c
5192
ext4_msg(sb, KERN_ERR, "bad block size %d",
fs/ext4/super.c
5200
bh = ext4_sb_bread_unmovable(sb, logical_sb_block);
fs/ext4/super.c
5202
ext4_msg(sb, KERN_ERR, "Can't read superblock on 2nd try");
fs/ext4/super.c
5210
ext4_msg(sb, KERN_ERR, "Magic mismatch, very weird!");
fs/ext4/super.c
5224
static int ext4_hash_info_init(struct super_block *sb)
fs/ext4/super.c
5226
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5233
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5237
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5245
if (ext4_has_feature_dir_index(sb)) {
fs/ext4/super.c
5251
if (!sb_rdonly(sb))
fs/ext4/super.c
5256
if (!sb_rdonly(sb))
fs/ext4/super.c
5265
static int ext4_block_group_meta_init(struct super_block *sb, int silent)
fs/ext4/super.c
5267
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5271
has_huge_files = ext4_has_feature_huge_file(sb);
fs/ext4/super.c
5272
sbi->s_bitmap_maxbytes = ext4_max_bitmap_size(sb->s_blocksize_bits,
fs/ext4/super.c
5274
sb->s_maxbytes = ext4_max_size(sb->s_blocksize_bits, has_huge_files);
fs/ext4/super.c
5277
if (ext4_has_feature_64bit(sb)) {
fs/ext4/super.c
5281
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5292
sbi->s_inodes_per_block = sb->s_blocksize / EXT4_INODE_SIZE(sb);
fs/ext4/super.c
5295
ext4_msg(sb, KERN_ERR, "VFS: Can't find ext4 filesystem");
fs/ext4/super.c
5299
sbi->s_inodes_per_group > sb->s_blocksize * 8) {
fs/ext4/super.c
5300
ext4_msg(sb, KERN_ERR, "invalid inodes per group: %lu\n",
fs/ext4/super.c
5306
sbi->s_desc_per_block = sb->s_blocksize / EXT4_DESC_SIZE(sb);
fs/ext4/super.c
5308
sbi->s_addr_per_block_bits = ilog2(EXT4_ADDR_PER_BLOCK(sb));
fs/ext4/super.c
5309
sbi->s_desc_per_block_bits = ilog2(EXT4_DESC_PER_BLOCK(sb));
fs/ext4/super.c
5319
static bool ext4_is_stripe_incompatible(struct super_block *sb, unsigned long stripe)
fs/ext4/super.c
5321
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5326
static int __ext4_fill_super(struct fs_context *fc, struct super_block *sb)
fs/ext4/super.c
5329
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5344
part_stat_read(sb->s_bdev, sectors[STAT_WRITE]);
fs/ext4/super.c
5346
err = ext4_load_super(sb, &logical_sb_block, silent);
fs/ext4/super.c
5353
err = ext4_init_metadata_csum(sb, es);
fs/ext4/super.c
5357
ext4_set_def_opts(sb, es);
fs/ext4/super.c
5373
err = ext4_inode_info_init(sb, es);
fs/ext4/super.c
5377
err = parse_apply_sb_mount_options(sb, ctx);
fs/ext4/super.c
5384
err = ext4_check_opt_consistency(fc, sb);
fs/ext4/super.c
5388
ext4_apply_options(fc, sb);
fs/ext4/super.c
5390
err = ext4_check_large_folio(sb);
fs/ext4/super.c
5394
err = ext4_encoding_init(sb, es);
fs/ext4/super.c
5398
err = ext4_check_journal_data_mode(sb);
fs/ext4/super.c
5402
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/ext4/super.c
5403
(test_opt(sb, POSIX_ACL) ? SB_POSIXACL : 0);
fs/ext4/super.c
5406
sb->s_iflags |= SB_I_ALLOW_HSM;
fs/ext4/super.c
5408
err = ext4_check_feature_compatibility(sb, es, silent);
fs/ext4/super.c
5412
err = ext4_block_group_meta_init(sb, silent);
fs/ext4/super.c
5416
err = ext4_hash_info_init(sb);
fs/ext4/super.c
5420
err = ext4_handle_clustersize(sb);
fs/ext4/super.c
5424
err = ext4_check_geometry(sb, es);
fs/ext4/super.c
5433
err = ext4_group_desc_init(sb, es, logical_sb_block, &first_not_zeroed);
fs/ext4/super.c
5442
if (ext4_is_stripe_incompatible(sb, sbi->s_stripe)) {
fs/ext4/super.c
5443
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
5454
sb->s_op = &ext4_sops;
fs/ext4/super.c
5455
sb->s_export_op = &ext4_export_ops;
fs/ext4/super.c
5456
sb->s_xattr = ext4_xattr_handlers;
fs/ext4/super.c
5458
sb->s_cop = &ext4_cryptops;
fs/ext4/super.c
5461
sb->s_vop = &ext4_verityops;
fs/ext4/super.c
5464
sb->dq_op = &ext4_quota_operations;
fs/ext4/super.c
5465
if (ext4_has_feature_quota(sb))
fs/ext4/super.c
5466
sb->s_qcop = &dquot_quotactl_sysfile_ops;
fs/ext4/super.c
5468
sb->s_qcop = &ext4_qctl_operations;
fs/ext4/super.c
5469
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP | QTYPE_MASK_PRJ;
fs/ext4/super.c
5471
super_set_uuid(sb, es->s_uuid, sizeof(es->s_uuid));
fs/ext4/super.c
5472
super_set_sysfs_name_bdev(sb);
fs/ext4/super.c
5479
ext4_atomic_write_init(sb);
fs/ext4/super.c
5480
ext4_fast_commit_init(sb);
fs/ext4/super.c
5482
sb->s_root = NULL;
fs/ext4/super.c
5485
ext4_has_feature_orphan_present(sb) ||
fs/ext4/super.c
5486
ext4_has_feature_journal_needs_recovery(sb));
fs/ext4/super.c
5488
if (ext4_has_feature_mmp(sb) && !sb_rdonly(sb)) {
fs/ext4/super.c
5489
err = ext4_multi_mount_protect(sb, le64_to_cpu(es->s_mmp_block));
fs/ext4/super.c
5499
if (!test_opt(sb, NOLOAD) && ext4_has_feature_journal(sb)) {
fs/ext4/super.c
5500
err = ext4_load_and_init_journal(sb, es, ctx);
fs/ext4/super.c
5503
if (bdev_read_only(sb->s_bdev))
fs/ext4/super.c
5505
} else if (test_opt(sb, NOLOAD) && !sb_rdonly(sb) &&
fs/ext4/super.c
5506
ext4_has_feature_journal_needs_recovery(sb)) {
fs/ext4/super.c
5507
ext4_msg(sb, KERN_ERR, "required journal recovery "
fs/ext4/super.c
5514
journal_option = ext4_has_journal_option(sb);
fs/ext4/super.c
5516
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5523
clear_opt(sb, JOURNAL_CHECKSUM);
fs/ext4/super.c
5524
clear_opt(sb, DATA_FLAGS);
fs/ext4/super.c
5525
clear_opt2(sb, JOURNAL_FAST_COMMIT);
fs/ext4/super.c
5530
if (!test_opt(sb, NO_MBCACHE)) {
fs/ext4/super.c
5533
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5539
if (ext4_has_feature_ea_inode(sb)) {
fs/ext4/super.c
5542
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5563
if (!ext4_has_feature_bigalloc(sb))
fs/ext4/super.c
5566
err = ext4_calculate_overhead(sb);
fs/ext4/super.c
5575
EXT4_SB(sb)->rsv_conversion_wq =
fs/ext4/super.c
5577
if (!EXT4_SB(sb)->rsv_conversion_wq) {
fs/ext4/super.c
5588
root = ext4_iget(sb, EXT4_ROOT_INO, EXT4_IGET_SPECIAL);
fs/ext4/super.c
5590
ext4_msg(sb, KERN_ERR, "get root inode failed");
fs/ext4/super.c
5596
ext4_msg(sb, KERN_ERR, "corrupt root inode, run e2fsck");
fs/ext4/super.c
5602
generic_set_sb_d_ops(sb);
fs/ext4/super.c
5603
sb->s_root = d_make_root(root);
fs/ext4/super.c
5604
if (!sb->s_root) {
fs/ext4/super.c
5605
ext4_msg(sb, KERN_ERR, "get root dentry failed");
fs/ext4/super.c
5610
err = ext4_setup_super(sb, es, sb_rdonly(sb));
fs/ext4/super.c
5612
sb->s_flags |= SB_RDONLY;
fs/ext4/super.c
5616
ext4_set_resv_clusters(sb);
fs/ext4/super.c
5618
if (test_opt(sb, BLOCK_VALIDITY)) {
fs/ext4/super.c
5619
err = ext4_setup_system_zone(sb);
fs/ext4/super.c
5621
ext4_msg(sb, KERN_ERR, "failed to initialize system "
fs/ext4/super.c
5626
ext4_fc_replay_cleanup(sb);
fs/ext4/super.c
5628
ext4_ext_init(sb);
fs/ext4/super.c
5637
set_opt2(sb, MB_OPTIMIZE_SCAN);
fs/ext4/super.c
5639
clear_opt2(sb, MB_OPTIMIZE_SCAN);
fs/ext4/super.c
5646
err = ext4_mb_init(sb);
fs/ext4/super.c
5648
ext4_msg(sb, KERN_ERR, "failed to initialize mballoc (%d)",
fs/ext4/super.c
5661
if (ext4_has_feature_flex_bg(sb))
fs/ext4/super.c
5662
if (!ext4_fill_flex_info(sb)) {
fs/ext4/super.c
5663
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5670
err = ext4_register_li_request(sb, first_not_zeroed);
fs/ext4/super.c
5674
err = ext4_init_orphan_info(sb);
fs/ext4/super.c
5679
if (ext4_has_feature_quota(sb) && !sb_rdonly(sb)) {
fs/ext4/super.c
5680
err = ext4_enable_quotas(sb);
fs/ext4/super.c
5690
errseq_check_and_advance(&sb->s_bdev->bd_mapping->wb_err,
fs/ext4/super.c
5692
EXT4_SB(sb)->s_mount_state |= EXT4_ORPHAN_FS;
fs/ext4/super.c
5693
ext4_orphan_cleanup(sb, es);
fs/ext4/super.c
5694
EXT4_SB(sb)->s_mount_state &= ~EXT4_ORPHAN_FS;
fs/ext4/super.c
5702
ext4_superblock_csum_set(sb);
fs/ext4/super.c
5704
ext4_msg(sb, KERN_INFO, "recovery complete");
fs/ext4/super.c
5705
err = ext4_mark_recovery_complete(sb, es);
fs/ext4/super.c
5710
if (test_opt(sb, DISCARD) && !bdev_max_discard_sectors(sb->s_bdev)) {
fs/ext4/super.c
5711
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
5713
clear_opt(sb, DISCARD);
fs/ext4/super.c
5731
err = ext4_register_sysfs(sb);
fs/ext4/super.c
5738
ext4_quotas_off(sb, EXT4_MAXQUOTAS);
fs/ext4/super.c
5740
ext4_release_orphan_info(sb);
fs/ext4/super.c
5742
ext4_unregister_li_request(sb);
fs/ext4/super.c
5744
ext4_mb_release(sb);
fs/ext4/super.c
5748
ext4_ext_release(sb);
fs/ext4/super.c
5749
ext4_release_system_zone(sb);
fs/ext4/super.c
5751
dput(sb->s_root);
fs/ext4/super.c
5752
sb->s_root = NULL;
fs/ext4/super.c
5754
ext4_msg(sb, KERN_ERR, "mount failed");
fs/ext4/super.c
5755
if (EXT4_SB(sb)->rsv_conversion_wq)
fs/ext4/super.c
5756
destroy_workqueue(EXT4_SB(sb)->rsv_conversion_wq);
fs/ext4/super.c
5777
utf8_unload(sb->s_encoding);
fs/ext4/super.c
5782
kfree(get_qf_name(sb, sbi, i));
fs/ext4/super.c
5791
invalidate_bdev(sb->s_bdev);
fs/ext4/super.c
5792
sb->s_fs_info = NULL;
fs/ext4/super.c
5796
static int ext4_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ext4/super.c
5803
sbi = ext4_alloc_sbi(sb);
fs/ext4/super.c
5810
strreplace(sb->s_id, '/', '!');
fs/ext4/super.c
5816
ret = __ext4_fill_super(fc, sb);
fs/ext4/super.c
5821
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA)
fs/ext4/super.c
5823
else if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_ORDERED_DATA)
fs/ext4/super.c
5831
ext4_msg(sb, KERN_INFO, "mounted filesystem %pU %s with%s. "
fs/ext4/super.c
5832
"Quota mode: %s.", &sb->s_uuid,
fs/ext4/super.c
5833
sb_rdonly(sb) ? "ro" : "r/w", descr,
fs/ext4/super.c
5834
ext4_quota_mode(sb));
fs/ext4/super.c
5837
ext4_update_overhead(sb, false);
fs/ext4/super.c
5856
static void ext4_init_journal_params(struct super_block *sb, journal_t *journal)
fs/ext4/super.c
5858
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
5863
ext4_fc_init(sb, journal);
fs/ext4/super.c
5866
if (test_opt(sb, BARRIER))
fs/ext4/super.c
5878
static struct inode *ext4_get_journal_inode(struct super_block *sb,
fs/ext4/super.c
5888
journal_inode = ext4_iget(sb, journal_inum, EXT4_IGET_SPECIAL);
fs/ext4/super.c
5890
ext4_msg(sb, KERN_ERR, "no journal found");
fs/ext4/super.c
5896
ext4_msg(sb, KERN_ERR, "journal inode is deleted");
fs/ext4/super.c
5900
ext4_msg(sb, KERN_ERR, "invalid journal inode");
fs/ext4/super.c
5932
static journal_t *ext4_open_inode_journal(struct super_block *sb,
fs/ext4/super.c
5938
journal_inode = ext4_get_journal_inode(sb, journal_inum);
fs/ext4/super.c
5944
ext4_msg(sb, KERN_ERR, "Could not load journal inode");
fs/ext4/super.c
5948
journal->j_private = sb;
fs/ext4/super.c
5950
ext4_init_journal_params(sb, journal);
fs/ext4/super.c
5954
static struct file *ext4_get_journal_blkdev(struct super_block *sb,
fs/ext4/super.c
5969
sb, &fs_holder_ops);
fs/ext4/super.c
5971
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5978
blocksize = sb->s_blocksize;
fs/ext4/super.c
5981
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
5992
ext4_msg(sb, KERN_ERR, "couldn't read superblock of "
fs/ext4/super.c
6002
ext4_msg(sb, KERN_ERR, "external journal has bad superblock");
fs/ext4/super.c
6010
ext4_msg(sb, KERN_ERR, "external journal has corrupt superblock");
fs/ext4/super.c
6015
if (memcmp(EXT4_SB(sb)->s_es->s_journal_uuid, es->s_uuid, 16)) {
fs/ext4/super.c
6016
ext4_msg(sb, KERN_ERR, "journal UUID does not match");
fs/ext4/super.c
6033
static journal_t *ext4_open_dev_journal(struct super_block *sb,
fs/ext4/super.c
6042
bdev_file = ext4_get_journal_blkdev(sb, j_dev, &j_start, &j_len);
fs/ext4/super.c
6046
journal = jbd2_journal_init_dev(file_bdev(bdev_file), sb->s_bdev, j_start,
fs/ext4/super.c
6047
j_len, sb->s_blocksize);
fs/ext4/super.c
6049
ext4_msg(sb, KERN_ERR, "failed to create device journal");
fs/ext4/super.c
6054
ext4_msg(sb, KERN_ERR, "External journal has more than one "
fs/ext4/super.c
6060
journal->j_private = sb;
fs/ext4/super.c
6061
EXT4_SB(sb)->s_journal_bdev_file = bdev_file;
fs/ext4/super.c
6062
ext4_init_journal_params(sb, journal);
fs/ext4/super.c
6066
ext4_journal_destroy(EXT4_SB(sb), journal);
fs/ext4/super.c
6072
static int ext4_load_journal(struct super_block *sb,
fs/ext4/super.c
6083
if (WARN_ON_ONCE(!ext4_has_feature_journal(sb)))
fs/ext4/super.c
6088
ext4_msg(sb, KERN_INFO, "external journal device major/minor "
fs/ext4/super.c
6095
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
6101
journal = ext4_open_inode_journal(sb, journal_inum);
fs/ext4/super.c
6105
journal = ext4_open_dev_journal(sb, journal_dev);
fs/ext4/super.c
6111
really_read_only = bdev_read_only(sb->s_bdev) | journal_dev_ro;
fs/ext4/super.c
6113
if (journal_dev_ro && !sb_rdonly(sb)) {
fs/ext4/super.c
6114
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
6125
if (ext4_has_feature_journal_needs_recovery(sb)) {
fs/ext4/super.c
6126
if (sb_rdonly(sb)) {
fs/ext4/super.c
6127
ext4_msg(sb, KERN_INFO, "INFO: recovery "
fs/ext4/super.c
6130
ext4_msg(sb, KERN_ERR, "write access "
fs/ext4/super.c
6136
ext4_msg(sb, KERN_INFO, "write access will "
fs/ext4/super.c
6142
ext4_msg(sb, KERN_INFO, "barriers disabled");
fs/ext4/super.c
6144
if (!ext4_has_feature_journal_needs_recovery(sb))
fs/ext4/super.c
6163
es->s_state |= cpu_to_le16(EXT4_SB(sb)->s_mount_state &
fs/ext4/super.c
6170
err2 = ext4_commit_super(sb);
fs/ext4/super.c
6176
ext4_msg(sb, KERN_ERR, "error loading journal");
fs/ext4/super.c
6180
EXT4_SB(sb)->s_journal = journal;
fs/ext4/super.c
6181
err = ext4_clear_journal_err(sb, es);
fs/ext4/super.c
6183
ext4_journal_destroy(EXT4_SB(sb), journal);
fs/ext4/super.c
6190
ext4_commit_super(sb);
fs/ext4/super.c
6195
ext4_commit_super(sb);
fs/ext4/super.c
6201
ext4_journal_destroy(EXT4_SB(sb), journal);
fs/ext4/super.c
6206
static void ext4_update_super(struct super_block *sb)
fs/ext4/super.c
6208
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
6223
if (!sb_rdonly(sb))
fs/ext4/super.c
6227
((part_stat_read(sb->s_bdev, sectors[STAT_WRITE]) -
fs/ext4/super.c
6277
ext4_superblock_csum_set(sb);
fs/ext4/super.c
6281
static int ext4_commit_super(struct super_block *sb)
fs/ext4/super.c
6283
struct buffer_head *sbh = EXT4_SB(sb)->s_sbh;
fs/ext4/super.c
6288
ext4_update_super(sb);
fs/ext4/super.c
6306
ext4_msg(sb, KERN_ERR, "previous I/O error to "
fs/ext4/super.c
6316
(test_opt(sb, BARRIER) ? REQ_FUA : 0), sbh);
fs/ext4/super.c
6319
ext4_msg(sb, KERN_ERR, "I/O error while writing "
fs/ext4/super.c
6333
static int ext4_mark_recovery_complete(struct super_block *sb,
fs/ext4/super.c
6337
journal_t *journal = EXT4_SB(sb)->s_journal;
fs/ext4/super.c
6339
if (!ext4_has_feature_journal(sb)) {
fs/ext4/super.c
6341
ext4_error(sb, "Journal got removed while the fs was "
fs/ext4/super.c
6352
if (sb_rdonly(sb) && (ext4_has_feature_journal_needs_recovery(sb) ||
fs/ext4/super.c
6353
ext4_has_feature_orphan_present(sb))) {
fs/ext4/super.c
6354
if (!ext4_orphan_file_empty(sb)) {
fs/ext4/super.c
6355
ext4_error(sb, "Orphan file not empty on read-only fs.");
fs/ext4/super.c
6359
ext4_clear_feature_journal_needs_recovery(sb);
fs/ext4/super.c
6360
ext4_clear_feature_orphan_present(sb);
fs/ext4/super.c
6361
ext4_commit_super(sb);
fs/ext4/super.c
6373
static int ext4_clear_journal_err(struct super_block *sb,
fs/ext4/super.c
6380
if (!ext4_has_feature_journal(sb)) {
fs/ext4/super.c
6381
ext4_error(sb, "Journal got removed while the fs was mounted!");
fs/ext4/super.c
6385
journal = EXT4_SB(sb)->s_journal;
fs/ext4/super.c
6396
errstr = ext4_decode_error(sb, j_errno, nbuf);
fs/ext4/super.c
6397
ext4_warning(sb, "Filesystem error recorded "
fs/ext4/super.c
6400
EXT4_SB(sb)->s_mount_state |= EXT4_ERROR_FS;
fs/ext4/super.c
6402
j_errno = ext4_commit_super(sb);
fs/ext4/super.c
6405
ext4_warning(sb, "Marked fs in need of filesystem check.");
fs/ext4/super.c
641
static void save_error_info(struct super_block *sb, int error,
fs/ext4/super.c
6417
int ext4_force_commit(struct super_block *sb)
fs/ext4/super.c
6419
return ext4_journal_force_commit(EXT4_SB(sb)->s_journal);
fs/ext4/super.c
6422
static int ext4_sync_fs(struct super_block *sb, int wait)
fs/ext4/super.c
6427
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
6429
ret = ext4_emergency_state(sb);
fs/ext4/super.c
6433
trace_ext4_sync_fs(sb, wait);
fs/ext4/super.c
6439
dquot_writeback_dquots(sb, -1);
fs/ext4/super.c
645
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
6456
} else if (wait && test_opt(sb, BARRIER))
fs/ext4/super.c
6460
err = blkdev_issue_flush(sb->s_bdev);
fs/ext4/super.c
6476
static int ext4_freeze(struct super_block *sb)
fs/ext4/super.c
6479
journal_t *journal = EXT4_SB(sb)->s_journal;
fs/ext4/super.c
6494
ext4_clear_feature_journal_needs_recovery(sb);
fs/ext4/super.c
6495
if (ext4_orphan_file_empty(sb))
fs/ext4/super.c
6496
ext4_clear_feature_orphan_present(sb);
fs/ext4/super.c
6499
error = ext4_commit_super(sb);
fs/ext4/super.c
6511
static int ext4_unfreeze(struct super_block *sb)
fs/ext4/super.c
6513
if (ext4_emergency_state(sb))
fs/ext4/super.c
6516
if (EXT4_SB(sb)->s_journal) {
fs/ext4/super.c
6518
ext4_set_feature_journal_needs_recovery(sb);
fs/ext4/super.c
6519
if (ext4_has_feature_orphan_file(sb))
fs/ext4/super.c
6520
ext4_set_feature_orphan_present(sb);
fs/ext4/super.c
6523
ext4_commit_super(sb);
fs/ext4/super.c
6543
static int __ext4_remount(struct fs_context *fc, struct super_block *sb)
fs/ext4/super.c
6547
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
6561
old_sb_flags = sb->s_flags;
fs/ext4/super.c
6573
char *qf_name = get_qf_name(sb, sbi, i);
fs/ext4/super.c
6594
ext4_is_stripe_incompatible(sb, ctx->s_stripe)) {
fs/ext4/super.c
6595
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
6609
alloc_ctx = ext4_writepages_down_write(sb);
fs/ext4/super.c
6610
ext4_apply_options(fc, sb);
fs/ext4/super.c
6611
ext4_writepages_up_write(sb, alloc_ctx);
fs/ext4/super.c
6614
test_opt(sb, JOURNAL_CHECKSUM)) {
fs/ext4/super.c
6615
ext4_msg(sb, KERN_ERR, "changing journal_checksum "
fs/ext4/super.c
6620
if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_JOURNAL_DATA) {
fs/ext4/super.c
6621
if (test_opt2(sb, EXPLICIT_DELALLOC)) {
fs/ext4/super.c
6622
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
6627
if (test_opt(sb, DIOREAD_NOLOCK)) {
fs/ext4/super.c
6628
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
6633
} else if (test_opt(sb, DATA_FLAGS) == EXT4_MOUNT_ORDERED_DATA) {
fs/ext4/super.c
6634
if (test_opt(sb, JOURNAL_ASYNC_COMMIT)) {
fs/ext4/super.c
6635
ext4_msg(sb, KERN_ERR, "can't mount with "
fs/ext4/super.c
6643
ext4_msg(sb, KERN_ERR, "can't enable nombcache during remount");
fs/ext4/super.c
6649
!test_opt(sb, DELALLOC)) {
fs/ext4/super.c
6650
ext4_msg(sb, KERN_ERR, "can't disable delalloc during remount");
fs/ext4/super.c
6655
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/ext4/super.c
6656
(test_opt(sb, POSIX_ACL) ? SB_POSIXACL : 0);
fs/ext4/super.c
6661
ext4_init_journal_params(sb, sbi->s_journal);
fs/ext4/super.c
6668
if ((bool)(fc->sb_flags & SB_RDONLY) != sb_rdonly(sb)) {
fs/ext4/super.c
6669
if (ext4_emergency_state(sb)) {
fs/ext4/super.c
6675
err = sync_filesystem(sb);
fs/ext4/super.c
6678
err = dquot_suspend(sb, -1);
fs/ext4/super.c
6686
sb->s_flags |= SB_RDONLY;
fs/ext4/super.c
6702
ext4_mark_recovery_complete(sb, es);
fs/ext4/super.c
6706
if (ext4_has_feature_readonly(sb) ||
fs/ext4/super.c
6707
!ext4_feature_set_ok(sb, 0)) {
fs/ext4/super.c
6717
ext4_get_group_desc(sb, g, NULL);
fs/ext4/super.c
6719
if (!ext4_group_desc_csum_verify(sb, g, gdp)) {
fs/ext4/super.c
6720
ext4_msg(sb, KERN_ERR,
fs/ext4/super.c
6722
g, le16_to_cpu(ext4_group_desc_csum(sb, g, gdp)),
fs/ext4/super.c
6734
if (es->s_last_orphan || !ext4_orphan_file_empty(sb)) {
fs/ext4/super.c
6735
ext4_msg(sb, KERN_WARNING, "Couldn't "
fs/ext4/super.c
6750
err = ext4_clear_journal_err(sb, es);
fs/ext4/super.c
6757
err = ext4_setup_super(sb, es, 0);
fs/ext4/super.c
6761
sb->s_flags &= ~SB_RDONLY;
fs/ext4/super.c
6762
if (ext4_has_feature_mmp(sb)) {
fs/ext4/super.c
6763
err = ext4_multi_mount_protect(sb,
fs/ext4/super.c
6779
if (test_opt(sb, BLOCK_VALIDITY) && !sbi->s_system_blks) {
fs/ext4/super.c
6780
err = ext4_setup_system_zone(sb);
fs/ext4/super.c
6786
err = ext4_commit_super(sb);
fs/ext4/super.c
6793
if (sb_any_quota_suspended(sb))
fs/ext4/super.c
6794
dquot_resume(sb, -1);
fs/ext4/super.c
6795
else if (ext4_has_feature_quota(sb)) {
fs/ext4/super.c
6796
err = ext4_enable_quotas(sb);
fs/ext4/super.c
6805
if (!test_opt(sb, BLOCK_VALIDITY) && sbi->s_system_blks)
fs/ext4/super.c
6806
ext4_release_system_zone(sb);
fs/ext4/super.c
6812
if (sb_rdonly(sb) || !test_opt(sb, INIT_INODE_TABLE))
fs/ext4/super.c
6813
ext4_unregister_li_request(sb);
fs/ext4/super.c
6816
first_not_zeroed = ext4_has_uninit_itable(sb);
fs/ext4/super.c
6817
ext4_register_li_request(sb, first_not_zeroed);
fs/ext4/super.c
6820
if (!ext4_has_feature_mmp(sb) || sb_rdonly(sb))
fs/ext4/super.c
6828
if (test_opt2(sb, ABORT))
fs/ext4/super.c
6829
ext4_abort(sb, ESHUTDOWN, "Abort forced by user");
fs/ext4/super.c
6838
if (sb_rdonly(sb) && !(old_sb_flags & SB_RDONLY) &&
fs/ext4/super.c
6839
sb_any_quota_suspended(sb))
fs/ext4/super.c
6840
dquot_resume(sb, -1);
fs/ext4/super.c
6842
alloc_ctx = ext4_writepages_down_write(sb);
fs/ext4/super.c
6843
sb->s_flags = old_sb_flags;
fs/ext4/super.c
6851
ext4_writepages_up_write(sb, alloc_ctx);
fs/ext4/super.c
6853
if (!test_opt(sb, BLOCK_VALIDITY) && sbi->s_system_blks)
fs/ext4/super.c
6854
ext4_release_system_zone(sb);
fs/ext4/super.c
6858
to_free[i] = get_qf_name(sb, sbi, i);
fs/ext4/super.c
6865
if (!ext4_has_feature_mmp(sb) || sb_rdonly(sb))
fs/ext4/super.c
6872
struct super_block *sb = fc->root->d_sb;
fs/ext4/super.c
6874
bool old_ro = sb_rdonly(sb);
fs/ext4/super.c
6876
fc->s_fs_info = EXT4_SB(sb);
fs/ext4/super.c
6878
ret = ext4_check_opt_consistency(fc, sb);
fs/ext4/super.c
6882
ret = __ext4_remount(fc, sb);
fs/ext4/super.c
6886
ext4_msg(sb, KERN_INFO, "re-mounted %pU%s.",
fs/ext4/super.c
6887
&sb->s_uuid,
fs/ext4/super.c
6888
(old_ro != sb_rdonly(sb)) ? (sb_rdonly(sb) ? " ro" : " r/w") : "");
fs/ext4/super.c
6894
static int ext4_statfs_project(struct super_block *sb,
fs/ext4/super.c
690
static void ext4_handle_error(struct super_block *sb, bool force_ro, int error,
fs/ext4/super.c
6903
dquot = dqget(sb, qid);
fs/ext4/super.c
6910
limit >>= sb->s_blocksize_bits;
fs/ext4/super.c
6916
dquot->dq_dqb.dqb_rsvspace) >> sb->s_blocksize_bits;
fs/ext4/super.c
694
journal_t *journal = EXT4_SB(sb)->s_journal;
fs/ext4/super.c
6945
struct super_block *sb = dentry->d_sb;
fs/ext4/super.c
6946
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
695
bool continue_fs = !force_ro && test_opt(sb, ERRORS_CONT);
fs/ext4/super.c
6952
if (!test_opt(sb, MINIX_DF))
fs/ext4/super.c
6956
buf->f_bsize = sb->s_blocksize;
fs/ext4/super.c
697
EXT4_SB(sb)->s_mount_state |= EXT4_ERROR_FS;
fs/ext4/super.c
6973
sb_has_quota_limits_enabled(sb, PRJQUOTA))
fs/ext4/super.c
6974
ext4_statfs_project(sb, EXT4_I(dentry->d_inode)->i_projid, buf);
fs/ext4/super.c
698
if (test_opt(sb, WARN_ON_ERROR))
fs/ext4/super.c
701
if (!continue_fs && !ext4_emergency_ro(sb) && journal)
fs/ext4/super.c
704
if (!bdev_read_only(sb->s_bdev)) {
fs/ext4/super.c
705
save_error_info(sb, error, ino, block, func, line);
fs/ext4/super.c
7075
struct super_block *sb = dquot->dq_sb;
fs/ext4/super.c
7077
if (ext4_is_quota_journalled(sb)) {
fs/ext4/super.c
7085
static int ext4_write_info(struct super_block *sb, int type)
fs/ext4/super.c
7091
handle = ext4_journal_start_sb(sb, EXT4_HT_QUOTA, 2);
fs/ext4/super.c
7094
ret = dquot_commit_info(sb, type);
fs/ext4/super.c
71
static void ext4_update_super(struct super_block *sb);
fs/ext4/super.c
7118
static int ext4_quota_on(struct super_block *sb, int type, int format_id,
fs/ext4/super.c
7123
if (!test_opt(sb, QUOTA))
fs/ext4/super.c
7127
if (path->dentry->d_sb != sb)
fs/ext4/super.c
7135
if (EXT4_SB(sb)->s_qf_names[type]) {
fs/ext4/super.c
7137
if (path->dentry->d_parent != sb->s_root)
fs/ext4/super.c
7138
ext4_msg(sb, KERN_WARNING,
fs/ext4/super.c
7141
sb_dqopt(sb)->flags |= DQUOT_NOLIST_DIRTY;
fs/ext4/super.c
7147
sb_dqopt(sb)->flags &= ~DQUOT_NOLIST_DIRTY;
fs/ext4/super.c
7151
err = dquot_quota_on(sb, type, format_id, path);
fs/ext4/super.c
716
!ext4_test_mount_flag(sb, EXT4_MF_JOURNAL_DESTROY))
fs/ext4/super.c
717
schedule_work(&EXT4_SB(sb)->s_sb_upd_work);
fs/ext4/super.c
7173
dquot_quota_off(sb, type);
fs/ext4/super.c
719
ext4_commit_super(sb);
fs/ext4/super.c
7195
static int ext4_quota_enable(struct super_block *sb, int type, int format_id,
fs/ext4/super.c
72
static int ext4_commit_super(struct super_block *sb);
fs/ext4/super.c
7201
le32_to_cpu(EXT4_SB(sb)->s_es->s_usr_quota_inum),
fs/ext4/super.c
7202
le32_to_cpu(EXT4_SB(sb)->s_es->s_grp_quota_inum),
fs/ext4/super.c
7203
le32_to_cpu(EXT4_SB(sb)->s_es->s_prj_quota_inum)
fs/ext4/super.c
7206
BUG_ON(!ext4_has_feature_quota(sb));
fs/ext4/super.c
7212
ext4_error(sb, "Bad quota inum: %lu, type: %d",
fs/ext4/super.c
7217
qf_inode = ext4_iget(sb, qf_inums[type], EXT4_IGET_SPECIAL);
fs/ext4/super.c
7219
ext4_error(sb, "Bad quota inode: %lu, type: %d",
fs/ext4/super.c
7236
int ext4_enable_quotas(struct super_block *sb)
fs/ext4/super.c
7240
le32_to_cpu(EXT4_SB(sb)->s_es->s_usr_quota_inum),
fs/ext4/super.c
7241
le32_to_cpu(EXT4_SB(sb)->s_es->s_grp_quota_inum),
fs/ext4/super.c
7242
le32_to_cpu(EXT4_SB(sb)->s_es->s_prj_quota_inum)
fs/ext4/super.c
7245
test_opt(sb, USRQUOTA),
fs/ext4/super.c
7246
test_opt(sb, GRPQUOTA),
fs/ext4/super.c
7247
test_opt(sb, PRJQUOTA),
fs/ext4/super.c
7250
sb_dqopt(sb)->flags |= DQUOT_QUOTA_SYS_FILE | DQUOT_NOLIST_DIRTY;
fs/ext4/super.c
7253
err = ext4_quota_enable(sb, type, QFMT_VFS_V1,
fs/ext4/super.c
7257
ext4_warning(sb,
fs/ext4/super.c
7263
ext4_quotas_off(sb, type);
fs/ext4/super.c
727
if (test_opt(sb, ERRORS_PANIC) && !system_going_down()) {
fs/ext4/super.c
7271
static int ext4_quota_off(struct super_block *sb, int type)
fs/ext4/super.c
7273
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext4/super.c
7279
if (test_opt(sb, DELALLOC))
fs/ext4/super.c
7280
sync_filesystem(sb);
fs/ext4/super.c
7285
err = dquot_quota_off(sb, type);
fs/ext4/super.c
7286
if (err || ext4_has_feature_quota(sb))
fs/ext4/super.c
729
sb->s_id);
fs/ext4/super.c
7293
if (sb_rdonly(sb))
fs/ext4/super.c
73
static int ext4_mark_recovery_complete(struct super_block *sb,
fs/ext4/super.c
7319
return dquot_quota_off(sb, type);
fs/ext4/super.c
732
if (ext4_emergency_ro(sb) || continue_fs)
fs/ext4/super.c
7326
static ssize_t ext4_quota_read(struct super_block *sb, int type, char *data,
fs/ext4/super.c
7329
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext4/super.c
7330
ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/super.c
7331
int offset = off & (sb->s_blocksize - 1);
fs/ext4/super.c
7343
tocopy = min_t(unsigned long, sb->s_blocksize - offset, toread);
fs/ext4/super.c
735
ext4_msg(sb, KERN_CRIT, "Remounting filesystem read-only");
fs/ext4/super.c
7362
static ssize_t ext4_quota_write(struct super_block *sb, int type,
fs/ext4/super.c
7365
struct inode *inode = sb_dqopt(sb)->files[type];
fs/ext4/super.c
7366
ext4_lblk_t blk = off >> EXT4_BLOCK_SIZE_BITS(sb);
fs/ext4/super.c
7367
int err = 0, err2 = 0, offset = off & (sb->s_blocksize - 1);
fs/ext4/super.c
7373
ext4_msg(sb, KERN_WARNING, "Quota write (off=%llu, len=%llu)"
fs/ext4/super.c
7382
if (sb->s_blocksize - offset < len) {
fs/ext4/super.c
7383
ext4_msg(sb, KERN_WARNING, "Quota write (off=%llu, len=%llu)"
fs/ext4/super.c
7400
err = ext4_journal_get_write_access(handle, sb, bh, EXT4_JTR_NONE);
fs/ext4/super.c
742
set_bit(EXT4_FLAGS_EMERGENCY_RO, &EXT4_SB(sb)->s_ext4_flags);
fs/ext4/super.c
7437
static inline int ext2_feature_set_ok(struct super_block *sb)
fs/ext4/super.c
7439
if (ext4_has_unknown_ext2_incompat_features(sb))
fs/ext4/super.c
7441
if (sb_rdonly(sb))
fs/ext4/super.c
7443
if (ext4_has_unknown_ext2_ro_compat_features(sb))
fs/ext4/super.c
7450
static inline int ext2_feature_set_ok(struct super_block *sb) { return 0; }
fs/ext4/super.c
7466
static inline int ext3_feature_set_ok(struct super_block *sb)
fs/ext4/super.c
7468
if (ext4_has_unknown_ext3_incompat_features(sb))
fs/ext4/super.c
7470
if (!ext4_has_feature_journal(sb))
fs/ext4/super.c
7472
if (sb_rdonly(sb))
fs/ext4/super.c
7474
if (ext4_has_unknown_ext3_ro_compat_features(sb))
fs/ext4/super.c
7479
static void ext4_kill_sb(struct super_block *sb)
fs/ext4/super.c
7481
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/super.c
7484
kill_block_super(sb);
fs/ext4/super.c
75
static int ext4_clear_journal_err(struct super_block *sb,
fs/ext4/super.c
77
static int ext4_sync_fs(struct super_block *sb, int wait);
fs/ext4/super.c
79
static int ext4_unfreeze(struct super_block *sb);
fs/ext4/super.c
80
static int ext4_freeze(struct super_block *sb);
fs/ext4/super.c
804
#define ext4_error_ratelimit(sb) \
fs/ext4/super.c
805
___ratelimit(&(EXT4_SB(sb)->s_err_ratelimit_state), \
fs/ext4/super.c
808
void __ext4_error(struct super_block *sb, const char *function,
fs/ext4/super.c
81
static inline int ext2_feature_set_ok(struct super_block *sb);
fs/ext4/super.c
815
if (unlikely(ext4_emergency_state(sb)))
fs/ext4/super.c
818
trace_ext4_error(sb, function, line);
fs/ext4/super.c
819
if (ext4_error_ratelimit(sb)) {
fs/ext4/super.c
82
static inline int ext3_feature_set_ok(struct super_block *sb);
fs/ext4/super.c
825
sb->s_id, function, line, current->comm, &vaf);
fs/ext4/super.c
828
fserror_report_metadata(sb, error ? -abs(error) : -EFSCORRUPTED,
fs/ext4/super.c
83
static void ext4_unregister_li_request(struct super_block *sb);
fs/ext4/super.c
831
ext4_handle_error(sb, force_ro, error, 0, block, function, line);
fs/ext4/super.c
85
static struct inode *ext4_get_journal_inode(struct super_block *sb,
fs/ext4/super.c
89
struct super_block *sb);
fs/ext4/super.c
90
static void ext4_apply_options(struct fs_context *fc, struct super_block *sb);
fs/ext4/super.c
909
const char *ext4_decode_error(struct super_block *sb, int errno,
fs/ext4/super.c
928
if (!sb || (EXT4_SB(sb)->s_journal &&
fs/ext4/super.c
929
EXT4_SB(sb)->s_journal->j_flags & JBD2_ABORT))
fs/ext4/super.c
952
void __ext4_std_error(struct super_block *sb, const char *function,
fs/ext4/super.c
958
if (unlikely(ext4_emergency_state(sb)))
fs/ext4/super.c
96
static void ext4_kill_sb(struct super_block *sb);
fs/ext4/super.c
964
if (errno == -EROFS && journal_current_handle() == NULL && sb_rdonly(sb))
fs/ext4/super.c
967
if (ext4_error_ratelimit(sb)) {
fs/ext4/super.c
968
errstr = ext4_decode_error(sb, errno, nbuf);
fs/ext4/super.c
970
sb->s_id, function, line, errstr);
fs/ext4/super.c
972
fserror_report_metadata(sb, errno ? -abs(errno) : -EFSCORRUPTED,
fs/ext4/super.c
975
ext4_handle_error(sb, false, -errno, 0, 0, function, line);
fs/ext4/super.c
978
void __ext4_msg(struct super_block *sb,
fs/ext4/super.c
984
if (sb) {
fs/ext4/super.c
985
atomic_inc(&EXT4_SB(sb)->s_msg_count);
fs/ext4/super.c
986
if (!___ratelimit(&(EXT4_SB(sb)->s_msg_ratelimit_state),
fs/ext4/super.c
994
if (sb)
fs/ext4/super.c
995
printk("%sEXT4-fs (%s): %pV\n", prefix, sb->s_id, &vaf);
fs/ext4/sysfs.c
610
int ext4_register_sysfs(struct super_block *sb)
fs/ext4/sysfs.c
612
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/sysfs.c
618
"%s", sb->s_id);
fs/ext4/sysfs.c
627
sbi->s_proc = proc_mkdir(sb->s_id, ext4_proc_root);
fs/ext4/sysfs.c
630
ext4_seq_options_show, sb);
fs/ext4/sysfs.c
633
sb);
fs/ext4/sysfs.c
635
ext4_fc_info_show, sb);
fs/ext4/sysfs.c
637
&ext4_mb_seq_groups_ops, sb);
fs/ext4/sysfs.c
639
ext4_seq_mb_stats_show, sb);
fs/ext4/sysfs.c
641
&ext4_mb_seq_structs_summary_ops, sb);
fs/ext4/sysfs.c
646
void ext4_unregister_sysfs(struct super_block *sb)
fs/ext4/sysfs.c
648
struct ext4_sb_info *sbi = EXT4_SB(sb);
fs/ext4/sysfs.c
651
remove_proc_subtree(sb->s_id, ext4_proc_root);
fs/ext4/sysfs.c
68
struct super_block *sb = sbi->s_buddy_cache->i_sb;
fs/ext4/sysfs.c
71
(part_stat_read(sb->s_bdev, sectors[STAT_WRITE]) -
fs/ext4/sysfs.c
77
struct super_block *sb = sbi->s_buddy_cache->i_sb;
fs/ext4/sysfs.c
81
((part_stat_read(sb->s_bdev, sectors[STAT_WRITE]) -
fs/ext4/sysfs.c
82
EXT4_SB(sb)->s_sectors_written_start) >> 1)));
fs/ext4/xattr.c
1862
struct super_block *sb = inode->i_sb;
fs/ext4/xattr.c
1870
bs->bh = ext4_sb_bread(sb, EXT4_I(inode)->i_file_acl, REQ_PRIO);
fs/ext4/xattr.c
1901
struct super_block *sb = inode->i_sb;
fs/ext4/xattr.c
1931
error = ext4_journal_get_write_access(handle, sb, bs->bh,
fs/ext4/xattr.c
2018
s->base = kzalloc(sb->s_blocksize, GFP_NOFS);
fs/ext4/xattr.c
2027
s->end = s->base + sb->s_blocksize;
fs/ext4/xattr.c
2059
EXT4_C2B(EXT4_SB(sb), 1));
fs/ext4/xattr.c
2064
handle, sb, new_bh,
fs/ext4/xattr.c
2084
EXT4_C2B(EXT4_SB(sb),
fs/ext4/xattr.c
2121
goal = ext4_group_first_block_no(sb,
fs/ext4/xattr.c
2131
new_bh = sb_getblk(sb, block);
fs/ext4/xattr.c
2156
error = ext4_journal_get_create_access(handle, sb,
fs/ext4/xattr.c
2215
dquot_free_block(inode, EXT4_C2B(EXT4_SB(sb), 1));
fs/ext4/xattr.c
2539
struct super_block *sb = inode->i_sb;
fs/ext4/xattr.c
2565
ext4_should_retry_alloc(sb, &retries))
fs/ext4/xattr.c
839
struct super_block *sb)
fs/ext4/xattr.c
841
if (ext4_has_feature_xattr(sb))
fs/ext4/xattr.c
844
BUFFER_TRACE(EXT4_SB(sb)->s_sbh, "get_write_access");
fs/ext4/xattr.c
845
if (ext4_journal_get_write_access(handle, sb, EXT4_SB(sb)->s_sbh,
fs/ext4/xattr.c
847
lock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/xattr.c
848
ext4_set_feature_xattr(sb);
fs/ext4/xattr.c
849
ext4_superblock_csum_set(sb);
fs/ext4/xattr.c
850
unlock_buffer(EXT4_SB(sb)->s_sbh);
fs/ext4/xattr.c
851
ext4_handle_dirty_metadata(handle, NULL, EXT4_SB(sb)->s_sbh);
fs/ext4/xattr.c
907
struct super_block *sb = inode->i_sb;
fs/ext4/xattr.c
908
size_t cluster_size = 1 << (EXT4_SB(sb)->s_cluster_bits +
fs/ext4/xattr.c
939
int __ext4_xattr_set_credits(struct super_block *sb, struct inode *inode,
fs/ext4/xattr.c
962
credits += EXT4_MAXQUOTAS_TRANS_BLOCKS(sb);
fs/ext4/xattr.c
972
if (!ext4_has_feature_ea_inode(sb))
fs/ext4/xattr.c
979
blocks = (value_len + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/ext4/xattr.c
997
blocks = XATTR_SIZE_MAX >> sb->s_blocksize_bits;
fs/ext4/xattr.h
185
extern int __ext4_xattr_set_credits(struct super_block *sb, struct inode *inode,
fs/f2fs/checkpoint.c
1284
trace_f2fs_sync_dirty_inodes_enter(sbi->sb, is_dir,
fs/f2fs/checkpoint.c
1289
trace_f2fs_sync_dirty_inodes_exit(sbi->sb, is_dir,
fs/f2fs/checkpoint.c
1300
trace_f2fs_sync_dirty_inodes_exit(sbi->sb, is_dir,
fs/f2fs/checkpoint.c
1441
f2fs_do_quota_sync(sbi->sb, -1);
fs/f2fs/checkpoint.c
1442
} else if (down_read_trylock(&sbi->sb->s_umount)) {
fs/f2fs/checkpoint.c
1443
f2fs_do_quota_sync(sbi->sb, -1);
fs/f2fs/checkpoint.c
1444
up_read(&sbi->sb->s_umount);
fs/f2fs/checkpoint.c
1648
return get_sectors_written(sbi->sb->s_bdev);
fs/f2fs/checkpoint.c
1861
if (f2fs_readonly(sbi->sb) || f2fs_hw_is_readonly(sbi))
fs/f2fs/checkpoint.c
1883
trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, CP_PHASE_START_BLOCK_OPS);
fs/f2fs/checkpoint.c
1891
trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, CP_PHASE_FINISH_BLOCK_OPS);
fs/f2fs/checkpoint.c
1959
trace_f2fs_write_checkpoint(sbi->sb, cpc->reason, CP_PHASE_FINISH_CHECKPOINT);
fs/f2fs/checkpoint.c
2145
dev_t dev = sbi->sb->s_bdev->bd_dev;
fs/f2fs/checkpoint.c
892
inode = f2fs_iget_retry(sbi->sb, ino);
fs/f2fs/compress.c
2039
inode = f2fs_iget(sbi->sb, F2FS_COMPRESS_INO(sbi));
fs/f2fs/compress.c
2062
dev_t dev = sbi->sb->s_bdev->bd_dev;
fs/f2fs/data.c
423
struct block_device *bdev = sbi->sb->s_bdev;
fs/f2fs/data.c
4355
if (f2fs_readonly(sbi->sb))
fs/f2fs/data.c
554
trace_f2fs_submit_read_bio(sbi->sb, type, bio);
fs/f2fs/data.c
564
trace_f2fs_submit_write_bio(sbi->sb, type, bio);
fs/f2fs/data.c
577
trace_f2fs_prepare_read_bio(io->sbi->sb, fio->type, io->bio);
fs/f2fs/data.c
580
trace_f2fs_prepare_write_bio(io->sbi->sb, fio->type, io->bio);
fs/f2fs/data.c
984
struct block_device *bdev = sbi->sb->s_bdev;
fs/f2fs/debug.c
311
si->base_mem += sizeof(struct f2fs_sb_info) + sbi->sb->s_blocksize;
fs/f2fs/debug.c
452
sbi->sb->s_bdev, i++,
fs/f2fs/debug.c
453
f2fs_readonly(sbi->sb) ? "RO" : "RW",
fs/f2fs/dir.c
30
return !sb_no_casefold_compat_fallback(sbi->sb);
fs/f2fs/dir.c
66
struct super_block *sb = dir->i_sb;
fs/f2fs/dir.c
73
GFP_NOFS, false, F2FS_SB(sb));
fs/f2fs/dir.c
77
len = utf8_casefold(sb->s_encoding, fname->usr_fname,
fs/f2fs/dir.c
81
if (sb_has_strict_encoding(sb))
fs/f2fs/f2fs.h
1743
struct super_block *sb; /* pointer to VFS super block */
fs/f2fs/f2fs.h
2178
static inline struct f2fs_sb_info *F2FS_SB(struct super_block *sb)
fs/f2fs/f2fs.h
2180
return sb->s_fs_info;
fs/f2fs/f2fs.h
2293
static inline unsigned long f2fs_qf_ino(struct super_block *sb, int type)
fs/f2fs/f2fs.h
2296
return le32_to_cpu(F2FS_SB(sb)->raw_super->qf_ino[type]);
fs/f2fs/f2fs.h
3665
static inline bool f2fs_readonly(struct super_block *sb)
fs/f2fs/f2fs.h
3667
return sb_rdonly(sb);
fs/f2fs/f2fs.h
3804
struct inode *f2fs_iget(struct super_block *sb, unsigned long ino);
fs/f2fs/f2fs.h
3805
struct inode *f2fs_iget_retry(struct super_block *sb, unsigned long ino);
fs/f2fs/f2fs.h
3901
int f2fs_do_quota_sync(struct super_block *sb, int type);
fs/f2fs/f2fs.h
3903
void f2fs_quota_off_umount(struct super_block *sb);
fs/f2fs/f2fs.h
3908
int f2fs_sync_fs(struct super_block *sb, int sync);
fs/f2fs/f2fs.h
4863
return f2fs_bdev_support_discard(sbi->sb->s_bdev);
fs/f2fs/f2fs.h
4874
unsigned int discard_granularity = bdev_discard_granularity(sbi->sb->s_bdev);
fs/f2fs/f2fs.h
4894
return bdev_read_only(sbi->sb->s_bdev);
fs/f2fs/f2fs.h
5022
if (f2fs_qf_ino(sbi->sb, i) == ino)
fs/f2fs/f2fs.h
5078
return f2fs_sb_has_readonly(sbi) || f2fs_readonly(sbi->sb);
fs/f2fs/file.c
2439
struct super_block *sb = sbi->sb;
fs/f2fs/file.c
2444
ret = bdev_freeze(sb->s_bdev);
fs/f2fs/file.c
2448
bdev_thaw(sb->s_bdev);
fs/f2fs/file.c
2452
ret = f2fs_sync_fs(sb, 1);
fs/f2fs/file.c
2472
ret = f2fs_sync_fs(sb, 1);
fs/f2fs/file.c
2489
down_write(&sbi->sb->s_umount);
fs/f2fs/file.c
2498
up_write(&sbi->sb->s_umount);
fs/f2fs/file.c
2772
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
2812
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
2869
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
2881
ret = f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/file.c
3064
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
3291
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
3360
struct super_block *sb = sbi->sb;
fs/f2fs/file.c
3363
transfer_to[PRJQUOTA] = dqget(sb, make_kqid_projid(kprojid));
fs/f2fs/file.c
3523
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
3664
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
3861
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
4056
if (f2fs_readonly(sbi->sb))
fs/f2fs/file.c
4222
sbi->sb->s_maxbytes : inode->i_size;
fs/f2fs/gc.c
1552
struct super_block *sb = sbi->sb;
fs/f2fs/gc.c
1611
inode = f2fs_iget(sb, dni.ino);
fs/f2fs/gc.c
173
trace_f2fs_background_gc(sbi->sb, wait_ms,
fs/f2fs/gc.c
188
sb_end_write(sbi->sb);
fs/f2fs/gc.c
1922
trace_f2fs_gc_begin(sbi->sb, gc_type, gc_control->no_bg_gc,
fs/f2fs/gc.c
1935
if (unlikely(!(sbi->sb->s_flags & SB_ACTIVE))) {
fs/f2fs/gc.c
197
dev_t dev = sbi->sb->s_bdev->bd_dev;
fs/f2fs/gc.c
2045
trace_f2fs_gc_end(sbi->sb, ret, total_freed, total_sec_freed,
fs/f2fs/gc.c
2345
err = freeze_super(sbi->sb, FREEZE_HOLDER_KERNEL, NULL);
fs/f2fs/gc.c
2349
if (f2fs_readonly(sbi->sb)) {
fs/f2fs/gc.c
2350
err = thaw_super(sbi->sb, FREEZE_HOLDER_KERNEL, NULL);
fs/f2fs/gc.c
2407
thaw_super(sbi->sb, FREEZE_HOLDER_KERNEL, NULL);
fs/f2fs/gc.c
67
if (f2fs_readonly(sbi->sb)) {
fs/f2fs/gc.c
74
if (sbi->sb->s_writers.frozen >= SB_FREEZE_WRITE) {
fs/f2fs/gc.c
84
if (!sb_start_write_trylock(sbi->sb)) {
fs/f2fs/gc.c
972
trace_f2fs_get_victim(sbi->sb, type, gc_type, &p,
fs/f2fs/inline.c
225
if (f2fs_hw_is_readonly(sbi) || f2fs_readonly(sbi->sb))
fs/f2fs/inode.c
31
if (f2fs_readonly(F2FS_I_SB(inode)->sb))
fs/f2fs/inode.c
568
struct inode *f2fs_iget(struct super_block *sb, unsigned long ino)
fs/f2fs/inode.c
570
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/inode.c
574
inode = iget_locked(sb, ino);
fs/f2fs/inode.c
658
struct inode *f2fs_iget_retry(struct super_block *sb, unsigned long ino)
fs/f2fs/inode.c
662
inode = f2fs_iget(sb, ino);
fs/f2fs/iostat.c
35
struct super_block *sb = seq->private;
fs/f2fs/iostat.c
36
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/namei.c
1098
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
1262
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
391
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
441
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
607
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
699
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
751
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/namei.c
806
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/node.c
122
if (!sbi->sb->s_bdi->wb.dirty_exceeded)
fs/f2fs/node.c
1646
inode = ilookup(sbi->sb, ino);
fs/f2fs/node.c
2030
inode = find_inode_nowait(sbi->sb, ino, f2fs_match_ino, NULL);
fs/f2fs/node.c
81
if (sbi->sb->s_bdi->wb.dirty_exceeded)
fs/f2fs/recovery.c
566
inode = f2fs_iget_retry(sbi->sb, ino);
fs/f2fs/recovery.c
81
inode = f2fs_iget_retry(sbi->sb, ino);
fs/f2fs/recovery.c
881
unsigned long s_flags = sbi->sb->s_flags;
fs/f2fs/recovery.c
911
f2fs_bug_on(sbi, sbi->sb->s_flags & SB_ACTIVE);
fs/f2fs/recovery.c
953
sbi->sb->s_flags = s_flags; /* Restore SB_RDONLY status */
fs/f2fs/segment.c
1935
if (f2fs_readonly(sbi->sb))
fs/f2fs/segment.c
1945
sb_start_intwrite(sbi->sb);
fs/f2fs/segment.c
1961
sb_end_intwrite(sbi->sb);
fs/f2fs/segment.c
2279
dev_t dev = sbi->sb->s_bdev->bd_dev;
fs/f2fs/segment.c
545
f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/segment.c
5561
if (!f2fs_sb_has_blkzoned(sbi) || f2fs_readonly(sbi->sb) ||
fs/f2fs/segment.c
566
return __submit_flush_wait(sbi, sbi->sb->s_bdev);
fs/f2fs/segment.c
685
dev_t dev = sbi->sb->s_bdev->bd_dev;
fs/f2fs/segment.h
995
if (sbi->sb->s_bdi->wb.dirty_exceeded)
fs/f2fs/super.c
1242
struct super_block *sb)
fs/f2fs/super.c
1244
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1248
bool quota_turnon = sb_any_quota_loaded(sb);
fs/f2fs/super.c
1344
if (f2fs_readonly(sbi->sb))
fs/f2fs/super.c
1360
struct super_block *sb)
fs/f2fs/super.c
1363
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1410
struct super_block *sb)
fs/f2fs/super.c
1414
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1480
struct super_block *sb)
fs/f2fs/super.c
1483
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1486
if (ctx_test_opt(ctx, F2FS_MOUNT_NORECOVERY) && !f2fs_readonly(sb))
fs/f2fs/super.c
1528
err = f2fs_check_test_dummy_encryption(fc, sb);
fs/f2fs/super.c
1532
err = f2fs_check_compression(fc, sb);
fs/f2fs/super.c
1536
err = f2fs_check_quota_consistency(fc, sb);
fs/f2fs/super.c
1597
if (f2fs_sb_has_readonly(sbi) && !f2fs_readonly(sbi->sb)) {
fs/f2fs/super.c
1605
struct super_block *sb)
fs/f2fs/super.c
1609
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1641
struct super_block *sb)
fs/f2fs/super.c
1644
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1655
struct super_block *sb)
fs/f2fs/super.c
1659
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1708
static void f2fs_apply_options(struct fs_context *fc, struct super_block *sb)
fs/f2fs/super.c
1711
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1761
f2fs_apply_compression(fc, sb);
fs/f2fs/super.c
1762
f2fs_apply_test_dummy_encryption(fc, sb);
fs/f2fs/super.c
1763
f2fs_apply_quota_options(fc, sb);
fs/f2fs/super.c
1793
static struct inode *f2fs_alloc_inode(struct super_block *sb)
fs/f2fs/super.c
1797
if (time_to_inject(F2FS_SB(sb), FAULT_SLAB_ALLOC))
fs/f2fs/super.c
1800
fi = alloc_inode_sb(sb, f2fs_inode_cachep, GFP_F2FS_ZERO);
fs/f2fs/super.c
1821
fi->i_dir_level = F2FS_SB(sb)->dir_level;
fs/f2fs/super.c
1976
static void f2fs_put_super(struct super_block *sb)
fs/f2fs/super.c
1978
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
1986
f2fs_quota_off_umount(sb);
fs/f2fs/super.c
2089
utf8_unload(sb->s_encoding);
fs/f2fs/super.c
2093
int f2fs_sync_fs(struct super_block *sb, int sync)
fs/f2fs/super.c
2095
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2103
trace_f2fs_sync_fs(sb, sync);
fs/f2fs/super.c
2116
static int f2fs_freeze(struct super_block *sb)
fs/f2fs/super.c
2118
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2120
if (f2fs_readonly(sb))
fs/f2fs/super.c
2143
static int f2fs_unfreeze(struct super_block *sb)
fs/f2fs/super.c
2145
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2157
clear_sbi_flag(F2FS_SB(sb), SBI_IS_FREEZING);
fs/f2fs/super.c
2162
static int f2fs_statfs_project(struct super_block *sb,
fs/f2fs/super.c
2171
dquot = dqget(sb, qid);
fs/f2fs/super.c
2178
limit >>= sb->s_blocksize_bits;
fs/f2fs/super.c
2184
dquot->dq_dqb.dqb_rsvspace) >> sb->s_blocksize_bits;
fs/f2fs/super.c
2214
struct super_block *sb = dentry->d_sb;
fs/f2fs/super.c
2215
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2216
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/f2fs/super.c
2263
sb_has_quota_limits_enabled(sb, PRJQUOTA)) {
fs/f2fs/super.c
2264
f2fs_statfs_project(sb, F2FS_I(d_inode(dentry))->i_projid, buf);
fs/f2fs/super.c
2271
struct super_block *sb)
fs/f2fs/super.c
2274
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2309
struct super_block *sb)
fs/f2fs/super.c
2311
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2479
f2fs_show_quota_options(seq, sbi->sb);
fs/f2fs/super.c
2481
fscrypt_show_test_dummy_encryption(seq, ',', sbi->sb);
fs/f2fs/super.c
2483
if (sbi->sb->s_flags & SB_INLINECRYPT)
fs/f2fs/super.c
2506
f2fs_show_compress_options(seq, sbi->sb);
fs/f2fs/super.c
2603
static int f2fs_enable_quotas(struct super_block *sb);
fs/f2fs/super.c
2608
unsigned int s_flags = sbi->sb->s_flags;
fs/f2fs/super.c
2620
sbi->sb->s_flags |= SB_ACTIVE;
fs/f2fs/super.c
2651
ret = sync_filesystem(sbi->sb);
fs/f2fs/super.c
2680
sbi->sb->s_flags = s_flags; /* Restore SB_RDONLY status */
fs/f2fs/super.c
2707
sync_inodes_sb(sbi->sb);
fs/f2fs/super.c
2747
ret = f2fs_sync_fs(sbi->sb, 1);
fs/f2fs/super.c
2762
static int __f2fs_remount(struct fs_context *fc, struct super_block *sb)
fs/f2fs/super.c
2764
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
2790
old_sb_flags = sb->s_flags;
fs/f2fs/super.c
2823
err = f2fs_check_opt_consistency(fc, sb);
fs/f2fs/super.c
2827
f2fs_apply_options(fc, sb);
fs/f2fs/super.c
2840
if (f2fs_readonly(sb) && (flags & SB_RDONLY))
fs/f2fs/super.c
2849
if (!f2fs_readonly(sb) && (flags & SB_RDONLY)) {
fs/f2fs/super.c
2850
err = dquot_suspend(sb, -1);
fs/f2fs/super.c
2853
} else if (f2fs_readonly(sb) && !(flags & SB_RDONLY)) {
fs/f2fs/super.c
2855
sb->s_flags &= ~SB_RDONLY;
fs/f2fs/super.c
2856
if (sb_any_quota_suspended(sb)) {
fs/f2fs/super.c
2857
dquot_resume(sb, -1);
fs/f2fs/super.c
2859
err = f2fs_enable_quotas(sb);
fs/f2fs/super.c
2929
sync_inodes_sb(sb);
fs/f2fs/super.c
2933
f2fs_sync_fs(sb, 1);
fs/f2fs/super.c
3008
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/f2fs/super.c
3012
fc->sb_flags = (flags & ~SB_LAZYTIME) | (sb->s_flags & SB_LAZYTIME);
fs/f2fs/super.c
3055
sb->s_flags = old_sb_flags;
fs/f2fs/super.c
3061
static void f2fs_shutdown(struct super_block *sb)
fs/f2fs/super.c
3063
f2fs_do_shutdown(F2FS_SB(sb), F2FS_GOING_DOWN_NOSYNC, false, false);
fs/f2fs/super.c
3082
bool readonly = f2fs_readonly(sbi->sb);
fs/f2fs/super.c
3092
sbi->sb->s_flags &= ~SB_RDONLY;
fs/f2fs/super.c
3107
f2fs_quota_off_umount(sbi->sb);
fs/f2fs/super.c
3111
sbi->sb->s_flags |= SB_RDONLY;
fs/f2fs/super.c
3116
static ssize_t f2fs_quota_read(struct super_block *sb, int type, char *data,
fs/f2fs/super.c
3119
struct inode *inode = sb_dqopt(sb)->files[type];
fs/f2fs/super.c
3143
set_sbi_flag(F2FS_SB(sb), SBI_QUOTA_NEED_REPAIR);
fs/f2fs/super.c
3160
f2fs_bug_on(F2FS_SB(sb), !folio_test_uptodate(folio));
fs/f2fs/super.c
3173
static ssize_t f2fs_quota_write(struct super_block *sb, int type,
fs/f2fs/super.c
3176
struct inode *inode = sb_dqopt(sb)->files[type];
fs/f2fs/super.c
3179
int offset = off & (sb->s_blocksize - 1);
fs/f2fs/super.c
3187
tocopy = min_t(unsigned long, sb->s_blocksize - offset,
fs/f2fs/super.c
3197
set_sbi_flag(F2FS_SB(sb), SBI_QUOTA_NEED_REPAIR);
fs/f2fs/super.c
3244
return dquot_quota_on_mount(sbi->sb, F2FS_OPTION(sbi).s_qf_names[type],
fs/f2fs/super.c
3254
err = f2fs_enable_quotas(sbi->sb);
fs/f2fs/super.c
3276
static int f2fs_quota_enable(struct super_block *sb, int type, int format_id,
fs/f2fs/super.c
3283
qf_inode = f2fs_iget(sb, qf_inum);
fs/f2fs/super.c
3285
f2fs_err(F2FS_SB(sb), "Bad quota inode %u:%lu", type, qf_inum);
fs/f2fs/super.c
3304
static int f2fs_enable_quotas(struct super_block *sb)
fs/f2fs/super.c
3306
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3320
sb_dqopt(sb)->flags |= DQUOT_QUOTA_SYS_FILE;
fs/f2fs/super.c
3323
qf_inum = f2fs_qf_ino(sb, type);
fs/f2fs/super.c
3325
err = f2fs_quota_enable(sb, type, QFMT_VFS_V1,
fs/f2fs/super.c
3332
dquot_quota_off(sb, type);
fs/f2fs/super.c
3343
struct quota_info *dqopt = sb_dqopt(sbi->sb);
fs/f2fs/super.c
3347
ret = dquot_writeback_dquots(sbi->sb, type);
fs/f2fs/super.c
3368
int f2fs_do_quota_sync(struct super_block *sb, int type)
fs/f2fs/super.c
3370
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3371
struct quota_info *dqopt = sb_dqopt(sb);
fs/f2fs/super.c
3385
if (!sb_has_quota_active(sb, cnt))
fs/f2fs/super.c
3417
static int f2fs_quota_sync(struct super_block *sb, int type)
fs/f2fs/super.c
3421
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
3422
ret = f2fs_do_quota_sync(sb, type);
fs/f2fs/super.c
3423
F2FS_SB(sb)->umount_lock_holder = NULL;
fs/f2fs/super.c
3427
static int f2fs_quota_on(struct super_block *sb, int type, int format_id,
fs/f2fs/super.c
3434
if (f2fs_sb_has_quota_ino(F2FS_SB(sb))) {
fs/f2fs/super.c
3435
f2fs_err(F2FS_SB(sb), "quota sysfile already exists");
fs/f2fs/super.c
3439
if (path->dentry->d_sb != sb)
fs/f2fs/super.c
3442
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
3444
err = f2fs_do_quota_sync(sb, type);
fs/f2fs/super.c
3458
err = dquot_quota_on(sb, type, format_id, path);
fs/f2fs/super.c
3468
F2FS_SB(sb)->umount_lock_holder = NULL;
fs/f2fs/super.c
3472
static int __f2fs_quota_off(struct super_block *sb, int type)
fs/f2fs/super.c
3474
struct inode *inode = sb_dqopt(sb)->files[type];
fs/f2fs/super.c
3478
return dquot_quota_off(sb, type);
fs/f2fs/super.c
3480
err = f2fs_do_quota_sync(sb, type);
fs/f2fs/super.c
3484
err = dquot_quota_off(sb, type);
fs/f2fs/super.c
3485
if (err || f2fs_sb_has_quota_ino(F2FS_SB(sb)))
fs/f2fs/super.c
3498
static int f2fs_quota_off(struct super_block *sb, int type)
fs/f2fs/super.c
3500
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3503
F2FS_SB(sb)->umount_lock_holder = current;
fs/f2fs/super.c
3505
err = __f2fs_quota_off(sb, type);
fs/f2fs/super.c
3515
F2FS_SB(sb)->umount_lock_holder = NULL;
fs/f2fs/super.c
3520
void f2fs_quota_off_umount(struct super_block *sb)
fs/f2fs/super.c
3526
err = __f2fs_quota_off(sb, type);
fs/f2fs/super.c
3528
int ret = dquot_quota_off(sb, type);
fs/f2fs/super.c
3530
f2fs_err(F2FS_SB(sb), "Fail to turn off disk quota (type: %d, err: %d, ret:%d), Please run fsck to fix it.",
fs/f2fs/super.c
3532
set_sbi_flag(F2FS_SB(sb), SBI_QUOTA_NEED_REPAIR);
fs/f2fs/super.c
3540
sync_filesystem(sb);
fs/f2fs/super.c
3543
static void f2fs_truncate_quota_inode_pages(struct super_block *sb)
fs/f2fs/super.c
3545
struct quota_info *dqopt = sb_dqopt(sb);
fs/f2fs/super.c
3593
struct super_block *sb = dquot->dq_sb;
fs/f2fs/super.c
3594
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3604
static int f2fs_dquot_commit_info(struct super_block *sb, int type)
fs/f2fs/super.c
3606
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3607
int ret = dquot_commit_info(sb, type);
fs/f2fs/super.c
3649
int f2fs_do_quota_sync(struct super_block *sb, int type)
fs/f2fs/super.c
3654
void f2fs_quota_off_umount(struct super_block *sb)
fs/f2fs/super.c
3708
static const union fscrypt_policy *f2fs_get_dummy_policy(struct super_block *sb)
fs/f2fs/super.c
3710
return F2FS_OPTION(F2FS_SB(sb)).dummy_enc_policy.policy;
fs/f2fs/super.c
3713
static bool f2fs_has_stable_inodes(struct super_block *sb)
fs/f2fs/super.c
3718
static struct block_device **f2fs_get_devices(struct super_block *sb,
fs/f2fs/super.c
3721
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3754
static struct inode *f2fs_nfs_get_inode(struct super_block *sb,
fs/f2fs/super.c
3757
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
3768
inode = f2fs_iget(sb, ino);
fs/f2fs/super.c
3779
static struct dentry *f2fs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/f2fs/super.c
3782
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/f2fs/super.c
3786
static struct dentry *f2fs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/f2fs/super.c
3789
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/f2fs/super.c
3857
bio = bio_alloc(sbi->sb->s_bdev, 1, opf, GFP_NOFS);
fs/f2fs/super.c
3876
struct super_block *sb = sbi->sb;
fs/f2fs/super.c
3946
if (f2fs_readonly(sb) || f2fs_hw_is_readonly(sbi)) {
fs/f2fs/super.c
4081
!bdev_is_zoned(sbi->sb->s_bdev)) {
fs/f2fs/super.c
448
KERN_SOH_ASCII, level, sbi->sb->s_id, &vaf);
fs/f2fs/super.c
4498
struct super_block *sb = sbi->sb;
fs/f2fs/super.c
4509
folio = read_mapping_folio(sb->s_bdev->bd_mapping, block, NULL);
fs/f2fs/super.c
455
KERN_SOH_ASCII, level, sbi->sb->s_id, &vaf);
fs/f2fs/super.c
4553
if ((recover && f2fs_readonly(sbi->sb)) ||
fs/f2fs/super.c
4568
folio = read_mapping_folio(sbi->sb->s_bdev->bd_mapping, index, NULL);
fs/f2fs/super.c
4580
folio = read_mapping_folio(sbi->sb->s_bdev->bd_mapping, index, NULL);
fs/f2fs/super.c
4655
struct super_block *sb = sbi->sb;
fs/f2fs/super.c
4682
sb->s_id);
fs/f2fs/super.c
4694
if (continue_fs || f2fs_readonly(sb) || shutdown) {
fs/f2fs/super.c
473
f2fs_sb_read_encoding(const struct f2fs_super_block *sb)
fs/f2fs/super.c
475
__u16 magic = le16_to_cpu(sb->s_encoding);
fs/f2fs/super.c
4754
blk_mode_t mode = sb_open_mode(sbi->sb->s_flags);
fs/f2fs/super.c
4759
if (!bdev_is_zoned(sbi->sb->s_bdev))
fs/f2fs/super.c
4775
logical_blksize = bdev_logical_block_size(sbi->sb->s_bdev);
fs/f2fs/super.c
4794
FDEV(0).bdev_file = sbi->sb->s_bdev_file;
fs/f2fs/super.c
4817
FDEV(i).path, mode, sbi->sb, NULL);
fs/f2fs/super.c
4860
if (f2fs_sb_has_casefold(sbi) && !sbi->sb->s_encoding) {
fs/f2fs/super.c
4892
sbi->sb->s_encoding = encoding;
fs/f2fs/super.c
4893
sbi->sb->s_encoding_flags = encoding_flags;
fs/f2fs/super.c
4919
static int f2fs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/f2fs/super.c
4945
sbi->sb = sb;
fs/f2fs/super.c
4966
if (unlikely(!sb_set_blocksize(sb, F2FS_BLKSIZE))) {
fs/f2fs/super.c
4976
sb->s_fs_info = sbi;
fs/f2fs/super.c
4990
err = f2fs_check_opt_consistency(fc, sb);
fs/f2fs/super.c
4994
f2fs_apply_options(fc, sb);
fs/f2fs/super.c
5000
sb->s_maxbytes = max_file_blocks(NULL) <<
fs/f2fs/super.c
5002
sb->s_max_links = F2FS_LINK_MAX;
fs/f2fs/super.c
5009
sb->dq_op = &f2fs_quota_operations;
fs/f2fs/super.c
5010
sb->s_qcop = &f2fs_quotactl_ops;
fs/f2fs/super.c
5011
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP | QTYPE_MASK_PRJ;
fs/f2fs/super.c
5015
if (f2fs_qf_ino(sbi->sb, i))
fs/f2fs/super.c
5021
sb->s_op = &f2fs_sops;
fs/f2fs/super.c
5023
sb->s_cop = &f2fs_cryptops;
fs/f2fs/super.c
5026
sb->s_vop = &f2fs_verityops;
fs/f2fs/super.c
5028
sb->s_xattr = f2fs_xattr_handlers;
fs/f2fs/super.c
5029
sb->s_export_op = &f2fs_export_ops;
fs/f2fs/super.c
5030
sb->s_magic = F2FS_SUPER_MAGIC;
fs/f2fs/super.c
5031
sb->s_time_gran = 1;
fs/f2fs/super.c
5032
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/f2fs/super.c
5035
sb->s_flags |= SB_INLINECRYPT;
fs/f2fs/super.c
5038
sb->s_flags |= SB_LAZYTIME;
fs/f2fs/super.c
5040
sb->s_flags &= ~SB_LAZYTIME;
fs/f2fs/super.c
5042
super_set_uuid(sb, (void *) raw_super->uuid, sizeof(raw_super->uuid));
fs/f2fs/super.c
5043
super_set_sysfs_name_bdev(sb);
fs/f2fs/super.c
5044
sb->s_iflags |= SB_I_CGROUPWB;
fs/f2fs/super.c
5071
sbi->meta_inode = f2fs_iget(sb, F2FS_META_INO(sbi));
fs/f2fs/super.c
5128
if (!f2fs_readonly(sb) && !test_opt(sbi, DISABLE_CHECKPOINT) &&
fs/f2fs/super.c
5176
sbi->node_inode = f2fs_iget(sb, F2FS_NODE_INO(sbi));
fs/f2fs/super.c
5184
root = f2fs_iget(sb, F2FS_ROOT_INO(sbi));
fs/f2fs/super.c
5197
generic_set_sb_d_ops(sb);
fs/f2fs/super.c
5198
sb->s_root = d_make_root(root); /* allocate root dentry */
fs/f2fs/super.c
5199
if (!sb->s_root) {
fs/f2fs/super.c
5215
if (f2fs_sb_has_quota_ino(sbi) && !f2fs_readonly(sb)) {
fs/f2fs/super.c
5216
err = f2fs_enable_quotas(sb);
fs/f2fs/super.c
5274
if (!f2fs_readonly(sb)) {
fs/f2fs/super.c
5320
test_opt(sbi, GC_MERGE)) && !f2fs_readonly(sb)) {
fs/f2fs/super.c
5349
sync_filesystem(sbi->sb);
fs/f2fs/super.c
5354
f2fs_truncate_quota_inode_pages(sb);
fs/f2fs/super.c
5355
if (f2fs_sb_has_quota_ino(sbi) && !f2fs_readonly(sb))
fs/f2fs/super.c
5356
f2fs_quota_off_umount(sbi->sb);
fs/f2fs/super.c
5366
evict_inodes(sb);
fs/f2fs/super.c
5371
dput(sb->s_root);
fs/f2fs/super.c
5372
sb->s_root = NULL;
fs/f2fs/super.c
5409
utf8_unload(sb->s_encoding);
fs/f2fs/super.c
5410
sb->s_encoding = NULL;
fs/f2fs/super.c
5423
sb->s_fs_info = NULL;
fs/f2fs/super.c
5428
shrink_dcache_sb(sb);
fs/f2fs/super.c
5441
struct super_block *sb = fc->root->d_sb;
fs/f2fs/super.c
5443
return __f2fs_remount(fc, sb);
fs/f2fs/super.c
5467
static void kill_f2fs_super(struct super_block *sb)
fs/f2fs/super.c
5469
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/super.c
5471
if (sb->s_root) {
fs/f2fs/super.c
5496
if (is_sbi_flag_set(sbi, SBI_IS_RECOVERED) && f2fs_readonly(sb))
fs/f2fs/super.c
5497
sb->s_flags &= ~SB_RDONLY;
fs/f2fs/super.c
5499
kill_block_super(sb);
fs/f2fs/super.c
5504
sb->s_fs_info = NULL;
fs/f2fs/sysfs.c
1001
if (!down_read_trylock(&sbi->sb->s_umount))
fs/f2fs/sysfs.c
1006
up_read(&sbi->sb->s_umount);
fs/f2fs/sysfs.c
1737
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1738
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1764
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1765
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1789
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1790
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1811
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1812
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1849
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1850
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1896
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1897
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
1960
struct super_block *sb = seq->private;
fs/f2fs/sysfs.c
1961
struct f2fs_sb_info *sbi = F2FS_SB(sb);
fs/f2fs/sysfs.c
2020
struct super_block *sb = sbi->sb;
fs/f2fs/sysfs.c
2026
"%s", sb->s_id);
fs/f2fs/sysfs.c
2045
sbi->s_proc = proc_mkdir(sb->s_id, f2fs_proc_root);
fs/f2fs/sysfs.c
2052
segment_info_seq_show, sb);
fs/f2fs/sysfs.c
2054
segment_bits_seq_show, sb);
fs/f2fs/sysfs.c
2057
iostat_info_seq_show, sb);
fs/f2fs/sysfs.c
2060
victim_bits_seq_show, sb);
fs/f2fs/sysfs.c
2062
discard_plist_seq_show, sb);
fs/f2fs/sysfs.c
2064
disk_map_seq_show, sb);
fs/f2fs/sysfs.c
2066
donation_list_seq_show, sb);
fs/f2fs/sysfs.c
2069
inject_stats_seq_show, sb);
fs/f2fs/sysfs.c
2086
remove_proc_subtree(sbi->sb->s_id, f2fs_proc_root);
fs/f2fs/sysfs.c
272
struct super_block *sb = sbi->sb;
fs/f2fs/sysfs.c
276
(sb->s_encoding->version >> 16) & 0xff,
fs/f2fs/sysfs.c
277
(sb->s_encoding->version >> 8) & 0xff,
fs/f2fs/sysfs.c
278
sb->s_encoding->version & 0xff);
fs/f2fs/sysfs.c
299
if (sb_no_casefold_compat_fallback(sbi->sb))
fs/fat/cache.c
222
struct super_block *sb = inode->i_sb;
fs/fat/cache.c
223
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/cache.c
224
const int limit = sb->s_maxbytes >> sbi->cluster_bits;
fs/fat/cache.c
234
fat_fs_error_ratelimit(sb,
fs/fat/cache.c
254
fat_fs_error_ratelimit(sb,
fs/fat/cache.c
265
fat_fs_error_ratelimit(sb,
fs/fat/cache.c
288
struct super_block *sb = inode->i_sb;
fs/fat/cache.c
298
fat_fs_error(sb, "%s: request beyond EOF (i_pos %lld)",
fs/fat/cache.c
309
struct super_block *sb = inode->i_sb;
fs/fat/cache.c
310
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/cache.c
313
cluster = sector >> (sbi->cluster_bits - sb->s_blocksize_bits);
fs/fat/cache.c
331
struct super_block *sb = inode->i_sb;
fs/fat/cache.c
332
const unsigned long blocksize = sb->s_blocksize;
fs/fat/cache.c
333
const unsigned char blocksize_bits = sb->s_blocksize_bits;
fs/fat/dir.c
1002
sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de);
fs/fat/dir.c
1011
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
102
*bh = sb_bread(sb, phys);
fs/fat/dir.c
1024
endp = (struct msdos_dir_entry *)(bh->b_data + sb->s_blocksize);
fs/fat/dir.c
104
fat_msg_ratelimit(sb, KERN_ERR,
fs/fat/dir.c
1046
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
107
*pos = (iblock + 1) << sb->s_blocksize_bits;
fs/fat/dir.c
1081
fat_msg(sb, KERN_WARNING,
fs/fat/dir.c
1099
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
1100
sector_t last_blknr = blknr + MSDOS_SB(sb)->sec_per_clus;
fs/fat/dir.c
1107
bhs[n] = sb_getblk(sb, blknr);
fs/fat/dir.c
111
offset = *pos & (sb->s_blocksize - 1);
fs/fat/dir.c
1114
memset(bhs[n]->b_data, 0, sb->s_blocksize);
fs/fat/dir.c
1150
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
1151
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
1164
bhs[0] = sb_getblk(sb, blknr);
fs/fat/dir.c
1195
memset(de + 2, 0, sb->s_blocksize - 2 * sizeof(*de));
fs/fat/dir.c
1217
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
1218
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
1247
bhs[n] = sb_getblk(sb, blknr);
fs/fat/dir.c
1254
copy = min(size, sb->s_blocksize);
fs/fat/dir.c
1270
memset(bhs[n]->b_data + copy, 0, sb->s_blocksize - copy);
fs/fat/dir.c
1298
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
1299
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
1337
fat_msg(sb, KERN_ERR, "Corrupted directory (i_pos %lld)",
fs/fat/dir.c
1354
int offset = pos & (sb->s_blocksize - 1);
fs/fat/dir.c
1359
int copy = umin(sb->s_blocksize - offset, size);
fs/fat/dir.c
1370
int copy = umin(sb->s_blocksize - offset, size);
fs/fat/dir.c
1402
fat_fs_error(sb, "Odd directory size");
fs/fat/dir.c
1412
sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de);
fs/fat/dir.c
143
static int uni16_to_x8(struct super_block *sb, unsigned char *ascii,
fs/fat/dir.c
146
int uni_xlate = MSDOS_SB(sb)->options.unicode_xlate;
fs/fat/dir.c
175
fat_msg(sb, KERN_WARNING,
fs/fat/dir.c
183
static inline int fat_uni_to_x8(struct super_block *sb, const wchar_t *uni,
fs/fat/dir.c
186
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
191
return uni16_to_x8(sb, buf, uni, size, sbi->nls_io);
fs/fat/dir.c
363
static int fat_parse_short(struct super_block *sb,
fs/fat/dir.c
367
const struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
43
static inline loff_t fat_make_i_pos(struct super_block *sb,
fs/fat/dir.c
466
name_len = fat_uni_to_x8(sb, uni_name, name,
fs/fat/dir.c
47
return ((loff_t)bh->b_blocknr << MSDOS_SB(sb)->dir_per_block_bits)
fs/fat/dir.c
480
struct super_block *sb = inode->i_sb;
fs/fat/dir.c
481
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
521
len = fat_parse_short(sb, de, bufname, 0);
fs/fat/dir.c
534
len = fat_uni_to_x8(sb, unicode, longname, size);
fs/fat/dir.c
54
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
546
sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de);
fs/fat/dir.c
55
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
571
struct super_block *sb = inode->i_sb;
fs/fat/dir.c
572
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/dir.c
641
int len = fat_uni_to_x8(sb, unicode, longname, size);
fs/fat/dir.c
649
short_len = fat_parse_short(sb, de, bufname,
fs/fat/dir.c
66
bh = sb_find_get_block(sb, phys);
fs/fat/dir.c
664
short_len = fat_parse_short(sb, de, bufname, sbi->options.dotsOK);
fs/fat/dir.c
684
loff_t i_pos = fat_make_i_pos(sb, bh, de);
fs/fat/dir.c
685
struct inode *tmp = fat_iget(sb, i_pos);
fs/fat/dir.c
69
sb_breadahead(sb, phys + sec);
fs/fat/dir.c
690
inum = iunique(sb, MSDOS_ROOT_INO);
fs/fat/dir.c
87
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
95
iblock = *pos >> sb->s_blocksize_bits;
fs/fat/dir.c
969
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
978
sinfo->i_pos = fat_make_i_pos(sb, sinfo->bh, sinfo->de);
fs/fat/dir.c
993
struct super_block *sb = dir->i_sb;
fs/fat/dir.c
999
if (fat_get_start(MSDOS_SB(sb), sinfo->de) == i_logstart) {
fs/fat/fat.h
144
static inline struct msdos_sb_info *MSDOS_SB(struct super_block *sb)
fs/fat/fat.h
146
return sb->s_fs_info;
fs/fat/fat.h
169
static inline u32 max_fat(struct super_block *sb)
fs/fat/fat.h
171
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fat.h
388
extern void fat_ent_access_init(struct super_block *sb);
fs/fat/fat.h
396
extern int fat_count_free_clusters(struct super_block *sb);
fs/fat/fat.h
417
extern struct inode *fat_iget(struct super_block *sb, loff_t i_pos);
fs/fat/fat.h
418
extern struct inode *fat_build_inode(struct super_block *sb,
fs/fat/fat.h
421
extern int fat_fill_super(struct super_block *sb, struct fs_context *fc,
fs/fat/fat.h
425
extern int fat_flush_inodes(struct super_block *sb, struct inode *i1,
fs/fat/fat.h
444
void __fat_fs_error(struct super_block *sb, int report, const char *fmt, ...);
fs/fat/fat.h
445
#define fat_fs_error(sb, fmt, args...) \
fs/fat/fat.h
446
__fat_fs_error(sb, 1, fmt , ## args)
fs/fat/fat.h
447
#define fat_fs_error_ratelimit(sb, fmt, args...) \
fs/fat/fat.h
448
__fat_fs_error(sb, __ratelimit(&MSDOS_SB(sb)->ratelimit), fmt , ## args)
fs/fat/fat.h
451
#define fat_msg(sb, level, fmt, args...) \
fs/fat/fat.h
454
_fat_msg(sb, level, fmt, ##args); \
fs/fat/fat.h
457
void _fat_msg(struct super_block *sb, const char *level, const char *fmt, ...);
fs/fat/fat.h
458
#define fat_msg_ratelimit(sb, level, fmt, args...) \
fs/fat/fat.h
460
if (__ratelimit(&MSDOS_SB(sb)->ratelimit)) \
fs/fat/fat.h
461
fat_msg(sb, level, fmt, ## args); \
fs/fat/fat.h
463
extern int fat_clusters_flush(struct super_block *sb);
fs/fat/fatent.c
102
static int fat_ent_bread(struct super_block *sb, struct fat_entry *fatent,
fs/fat/fatent.c
105
const struct fatent_operations *ops = MSDOS_SB(sb)->fatent_ops;
fs/fat/fatent.c
107
WARN_ON(blocknr < MSDOS_SB(sb)->fat_start);
fs/fat/fatent.c
108
fatent->fat_inode = MSDOS_SB(sb)->fat_inode;
fs/fat/fatent.c
109
fatent->bhs[0] = sb_bread(sb, blocknr);
fs/fat/fatent.c
111
fat_msg_ratelimit(sb, KERN_ERR, "FAT read failed (blocknr %llu)",
fs/fat/fatent.c
23
static void fat12_ent_blocknr(struct super_block *sb, int entry,
fs/fat/fatent.c
26
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
289
void fat_ent_access_init(struct super_block *sb)
fs/fat/fatent.c
29
*offset = bytes & (sb->s_blocksize - 1);
fs/fat/fatent.c
291
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
30
*blocknr = sbi->fat_start + (bytes >> sb->s_blocksize_bits);
fs/fat/fatent.c
305
fat_fs_error(sb, "invalid FAT variant, %u bits", sbi->fat_bits);
fs/fat/fatent.c
309
static void mark_fsinfo_dirty(struct super_block *sb)
fs/fat/fatent.c
311
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
313
if (sb_rdonly(sb) || !is_fat32(sbi))
fs/fat/fatent.c
319
static inline int fat_ent_update_ptr(struct super_block *sb,
fs/fat/fatent.c
323
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
33
static void fat_ent_blocknr(struct super_block *sb, int entry,
fs/fat/fatent.c
331
if ((offset + 1) < sb->s_blocksize) {
fs/fat/fatent.c
351
struct super_block *sb = inode->i_sb;
fs/fat/fatent.c
359
fat_fs_error_ratelimit(sb, "invalid access to FAT (entry 0x%08x)", entry);
fs/fat/fatent.c
36
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
364
ops->ent_blocknr(sb, entry, &offset, &blocknr);
fs/fat/fatent.c
366
if (!fat_ent_update_ptr(sb, fatent, offset, blocknr)) {
fs/fat/fatent.c
368
err = ops->ent_bread(sb, fatent, offset, blocknr);
fs/fat/fatent.c
376
static int fat_mirror_bhs(struct super_block *sb, struct buffer_head **bhs,
fs/fat/fatent.c
379
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
388
c_bh = sb_getblk(sb, backup_fat + bhs[n]->b_blocknr);
fs/fat/fatent.c
39
*offset = bytes & (sb->s_blocksize - 1);
fs/fat/fatent.c
395
memcpy(c_bh->b_data, bhs[n]->b_data, sb->s_blocksize);
fs/fat/fatent.c
399
if (sb->s_flags & SB_SYNCHRONOUS)
fs/fat/fatent.c
40
*blocknr = sbi->fat_start + (bytes >> sb->s_blocksize_bits);
fs/fat/fatent.c
413
struct super_block *sb = inode->i_sb;
fs/fat/fatent.c
414
const struct fatent_operations *ops = MSDOS_SB(sb)->fatent_ops;
fs/fat/fatent.c
423
return fat_mirror_bhs(sb, fatent->bhs, fatent->nr_bhs);
fs/fat/fatent.c
436
static inline int fat_ent_read_block(struct super_block *sb,
fs/fat/fatent.c
439
const struct fatent_operations *ops = MSDOS_SB(sb)->fatent_ops;
fs/fat/fatent.c
444
ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr);
fs/fat/fatent.c
445
return ops->ent_bread(sb, fatent, offset, blocknr);
fs/fat/fatent.c
468
struct super_block *sb = inode->i_sb;
fs/fat/fatent.c
469
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
493
err = fat_ent_read_block(sb, &fatent);
fs/fat/fatent.c
537
mark_fsinfo_dirty(sb);
fs/fat/fatent.c
543
err = fat_mirror_bhs(sb, bhs, nr_bhs);
fs/fat/fatent.c
556
struct super_block *sb = inode->i_sb;
fs/fat/fatent.c
557
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
573
fat_fs_error(sb, "%s: deleting FAT entry beyond EOF",
fs/fat/fatent.c
588
sb_issue_discard(sb,
fs/fat/fatent.c
604
if (sb->s_flags & SB_SYNCHRONOUS) {
fs/fat/fatent.c
609
err = fat_mirror_bhs(sb, bhs, nr_bhs);
fs/fat/fatent.c
619
if (sb->s_flags & SB_SYNCHRONOUS) {
fs/fat/fatent.c
624
err = fat_mirror_bhs(sb, bhs, nr_bhs);
fs/fat/fatent.c
631
mark_fsinfo_dirty(sb);
fs/fat/fatent.c
647
static void fat_ra_init(struct super_block *sb, struct fatent_ra *ra,
fs/fat/fatent.c
650
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
659
unsigned long ra_pages = sb->s_bdi->ra_pages;
fs/fat/fatent.c
665
if (ra_pages > sb->s_bdi->io_pages)
fs/fat/fatent.c
666
ra_pages = rounddown(ra_pages, sb->s_bdi->io_pages);
fs/fat/fatent.c
667
reada_blocks = ra_pages << (PAGE_SHIFT - sb->s_blocksize_bits + 1);
fs/fat/fatent.c
670
ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr);
fs/fat/fatent.c
671
ops->ent_blocknr(sb, ent_limit - 1, &offset, &block_end);
fs/fat/fatent.c
683
static void fat_ent_reada(struct super_block *sb, struct fatent_ra *ra,
fs/fat/fatent.c
69
static int fat12_ent_bread(struct super_block *sb, struct fat_entry *fatent,
fs/fat/fatent.c
690
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
696
ops->ent_blocknr(sb, fatent->entry, &offset, &blocknr);
fs/fat/fatent.c
705
sb_breadahead(sb, ra->ra_next + diff);
fs/fat/fatent.c
716
int fat_count_free_clusters(struct super_block *sb)
fs/fat/fatent.c
718
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
731
fat_ra_init(sb, &fatent_ra, &fatent, sbi->max_cluster);
fs/fat/fatent.c
734
fat_ent_reada(sb, &fatent_ra, &fatent);
fs/fat/fatent.c
736
err = fat_ent_read_block(sb, &fatent);
fs/fat/fatent.c
74
WARN_ON(blocknr < MSDOS_SB(sb)->fat_start);
fs/fat/fatent.c
748
mark_fsinfo_dirty(sb);
fs/fat/fatent.c
75
fatent->fat_inode = MSDOS_SB(sb)->fat_inode;
fs/fat/fatent.c
755
static int fat_trim_clusters(struct super_block *sb, u32 clus, u32 nr_clus)
fs/fat/fatent.c
757
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
758
return sb_issue_discard(sb, fat_clus_to_blknr(sbi, clus),
fs/fat/fatent.c
764
struct super_block *sb = inode->i_sb;
fs/fat/fatent.c
765
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/fatent.c
77
bhs[0] = sb_bread(sb, blocknr);
fs/fat/fatent.c
791
fat_ra_init(sb, &fatent_ra, &fatent, ent_end + 1);
fs/fat/fatent.c
794
fat_ent_reada(sb, &fatent_ra, &fatent);
fs/fat/fatent.c
796
err = fat_ent_read_block(sb, &fatent);
fs/fat/fatent.c
806
err = fat_trim_clusters(sb, clus, free);
fs/fat/fatent.c
81
if ((offset + 1) < sb->s_blocksize)
fs/fat/fatent.c
833
err = fat_trim_clusters(sb, clus, free);
fs/fat/fatent.c
86
bhs[1] = sb_bread(sb, blocknr);
fs/fat/fatent.c
97
fat_msg_ratelimit(sb, KERN_ERR, "FAT read failed (blocknr %llu)",
fs/fat/file.c
129
struct super_block *sb = inode->i_sb;
fs/fat/file.c
137
if (!bdev_max_discard_sectors(sb->s_bdev))
fs/fat/file.c
144
range.minlen = max(range.minlen, bdev_discard_granularity(sb->s_bdev));
fs/fat/file.c
268
struct super_block *sb = inode->i_sb;
fs/fat/file.c
269
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/file.c
313
struct super_block *sb = inode->i_sb;
fs/fat/file.c
359
fat_fs_error(sb,
fs/fat/file.c
374
inode->i_blocks = skip << (MSDOS_SB(sb)->cluster_bits - 9);
fs/fat/inode.c
123
struct super_block *sb = inode->i_sb;
fs/fat/inode.c
124
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
133
map_bh(bh_result, sb, phys);
fs/fat/inode.c
1353
static unsigned long calc_fat_clusters(struct super_block *sb)
fs/fat/inode.c
1355
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
1359
unsigned long ent_per_sec = sb->s_blocksize * 8 / sbi->fat_bits;
fs/fat/inode.c
1363
return sbi->fat_length * sb->s_blocksize * 8 / sbi->fat_bits;
fs/fat/inode.c
1391
static int fat_read_bpb(struct super_block *sb, struct fat_boot_sector *b,
fs/fat/inode.c
140
if (iblock != MSDOS_I(inode)->mmu_private >> sb->s_blocksize_bits) {
fs/fat/inode.c
141
fat_fs_error(sb, "corrupted file size (i_pos %lld, %lld)",
fs/fat/inode.c
1419
fat_msg(sb, KERN_ERR,
fs/fat/inode.c
1425
fat_msg(sb, KERN_ERR, "bogus number of FAT structure");
fs/fat/inode.c
1436
fat_msg(sb, KERN_ERR, "invalid media value (0x%02x)",
fs/fat/inode.c
1445
fat_msg(sb, KERN_ERR, "bogus logical sector size %u",
fs/fat/inode.c
1452
fat_msg(sb, KERN_ERR, "bogus sectors per cluster %u",
fs/fat/inode.c
1459
fat_msg(sb, KERN_ERR, "bogus number of FAT sectors");
fs/fat/inode.c
146
last_block = inode->i_blocks >> (sb->s_blocksize_bits - 9);
fs/fat/inode.c
1469
static int fat_read_static_bpb(struct super_block *sb,
fs/fat/inode.c
1474
sector_t bd_sects = bdev_nr_sectors(sb->s_bdev);
fs/fat/inode.c
1482
fat_msg(sb, KERN_ERR,
fs/fat/inode.c
1493
fat_msg(sb, KERN_ERR,
fs/fat/inode.c
1507
fat_msg(sb, KERN_WARNING,
fs/fat/inode.c
1514
fat_msg(sb, KERN_INFO,
fs/fat/inode.c
1535
int fat_fill_super(struct super_block *sb, struct fs_context *fc,
fs/fat/inode.c
1560
sb->s_fs_info = sbi;
fs/fat/inode.c
1562
sb->s_flags |= SB_NODIRATIME;
fs/fat/inode.c
1563
sb->s_magic = MSDOS_SUPER_MAGIC;
fs/fat/inode.c
1564
sb->s_op = &fat_sops;
fs/fat/inode.c
1565
sb->s_export_op = &fat_export_ops;
fs/fat/inode.c
1570
sb->s_time_gran = 1;
fs/fat/inode.c
1577
fat_msg(sb, KERN_WARNING, "utf8 is not a recommended IO charset"
fs/fat/inode.c
1588
sb->s_flags |= SB_RDONLY;
fs/fat/inode.c
1589
sb->s_export_op = &fat_export_ops_nostale;
fs/fat/inode.c
1597
setup(sb); /* flavour-specific stuff that needs options */
fs/fat/inode.c
1600
if (!sb_min_blocksize(sb, 512)) {
fs/fat/inode.c
1601
fat_msg(sb, KERN_ERR, "unable to set blocksize");
fs/fat/inode.c
1605
bh = sb_bread(sb, 0);
fs/fat/inode.c
1607
fat_msg(sb, KERN_ERR, "unable to read boot sector");
fs/fat/inode.c
1611
error = fat_read_bpb(sb, (struct fat_boot_sector *)bh->b_data, silent,
fs/fat/inode.c
1614
error = fat_read_static_bpb(sb,
fs/fat/inode.c
1627
if (logical_sector_size < sb->s_blocksize) {
fs/fat/inode.c
1628
fat_msg(sb, KERN_ERR, "logical sector size too small for device"
fs/fat/inode.c
163
MSDOS_I(inode)->mmu_private += *max_blocks << sb->s_blocksize_bits;
fs/fat/inode.c
1633
if (logical_sector_size > sb->s_blocksize) {
fs/fat/inode.c
1636
if (!sb_set_blocksize(sb, logical_sector_size)) {
fs/fat/inode.c
1637
fat_msg(sb, KERN_ERR, "unable to set blocksize %u",
fs/fat/inode.c
1643
bh_resize = sb_bread(sb, 0);
fs/fat/inode.c
1645
fat_msg(sb, KERN_ERR, "unable to read boot sector"
fs/fat/inode.c
1647
sb->s_blocksize);
fs/fat/inode.c
1654
sbi->cluster_size = sb->s_blocksize * sbi->sec_per_clus;
fs/fat/inode.c
1664
sb->s_maxbytes = 0xffffffff;
fs/fat/inode.c
1666
sb->s_time_min = ts.tv_sec;
fs/fat/inode.c
1670
sb->s_time_max = ts.tv_sec;
fs/fat/inode.c
1686
fsinfo_bh = sb_bread(sb, sbi->fsinfo_sector);
fs/fat/inode.c
1688
fat_msg(sb, KERN_ERR, "bread failed, FSINFO block"
fs/fat/inode.c
169
fat_fs_error(sb,
fs/fat/inode.c
1695
fat_msg(sb, KERN_WARNING, "Invalid FSINFO signature: "
fs/fat/inode.c
1717
super_set_uuid(sb, (void *) &vol_id_le, sizeof(vol_id_le));
fs/fat/inode.c
1719
sbi->dir_per_block = sb->s_blocksize / sizeof(struct msdos_dir_entry);
fs/fat/inode.c
1726
fat_msg(sb, KERN_ERR, "bogus number of directory entries"
fs/fat/inode.c
1732
* sizeof(struct msdos_dir_entry) / sb->s_blocksize;
fs/fat/inode.c
1750
fat_clusters = calc_fat_clusters(sb);
fs/fat/inode.c
1752
if (total_clusters > max_fat(sb)) {
fs/fat/inode.c
1754
fat_msg(sb, KERN_ERR, "count of clusters too big (%u)",
fs/fat/inode.c
1769
fat_hash_init(sb);
fs/fat/inode.c
1770
dir_hash_init(sb);
fs/fat/inode.c
1771
fat_ent_access_init(sb);
fs/fat/inode.c
178
map_bh(bh_result, sb, phys);
fs/fat/inode.c
1789
fat_msg(sb, KERN_ERR, "codepage %s not found", buf);
fs/fat/inode.c
1797
fat_msg(sb, KERN_ERR, "IO charset %s not found",
fs/fat/inode.c
1804
fat_inode = new_inode(sb);
fs/fat/inode.c
1809
fsinfo_inode = new_inode(sb);
fs/fat/inode.c
1816
root_inode = new_inode(sb);
fs/fat/inode.c
1829
sb->s_root = d_make_root(root_inode);
fs/fat/inode.c
1830
if (!sb->s_root) {
fs/fat/inode.c
1831
fat_msg(sb, KERN_ERR, "get root inode failed");
fs/fat/inode.c
1835
if (sbi->options.discard && !bdev_max_discard_sectors(sb->s_bdev))
fs/fat/inode.c
1836
fat_msg(sb, KERN_WARNING,
fs/fat/inode.c
1839
fat_set_state(sb, 1, 0);
fs/fat/inode.c
1845
fat_msg(sb, KERN_INFO, "Can't find a valid FAT filesystem");
fs/fat/inode.c
1853
sb->s_fs_info = NULL;
fs/fat/inode.c
186
struct super_block *sb = inode->i_sb;
fs/fat/inode.c
1889
int fat_flush_inodes(struct super_block *sb, struct inode *i1, struct inode *i2)
fs/fat/inode.c
1892
if (!MSDOS_SB(sb)->options.flush)
fs/fat/inode.c
1899
ret = sync_blockdev_nowait(sb->s_bdev);
fs/fat/inode.c
193
bh_result->b_size = max_blocks << sb->s_blocksize_bits;
fs/fat/inode.c
294
struct super_block *sb = inode->i_sb;
fs/fat/inode.c
307
map_bh(bh_result, sb, bmap);
fs/fat/inode.c
311
bh_result->b_size = max_blocks << sb->s_blocksize_bits;
fs/fat/inode.c
377
static void fat_hash_init(struct super_block *sb)
fs/fat/inode.c
379
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
392
static void dir_hash_init(struct super_block *sb)
fs/fat/inode.c
394
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
447
struct inode *fat_iget(struct super_block *sb, loff_t i_pos)
fs/fat/inode.c
449
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
456
BUG_ON(i->vfs_inode.i_sb != sb);
fs/fat/inode.c
496
struct super_block *sb = dir->i_sb;
fs/fat/inode.c
500
fat_fs_error(sb, "corrupted directory (invalid entries)");
fs/fat/inode.c
504
MSDOS_I(dir)->i_start == MSDOS_SB(sb)->root_cluster) {
fs/fat/inode.c
506
fat_fs_error(sb, "corrupted directory (invalid i_start)");
fs/fat/inode.c
594
struct inode *fat_build_inode(struct super_block *sb,
fs/fat/inode.c
600
fat_lock_build_inode(MSDOS_SB(sb));
fs/fat/inode.c
601
inode = fat_iget(sb, i_pos);
fs/fat/inode.c
604
inode = new_inode(sb);
fs/fat/inode.c
609
inode->i_ino = iunique(sb, MSDOS_ROOT_INO);
fs/fat/inode.c
620
fat_unlock_build_inode(MSDOS_SB(sb));
fs/fat/inode.c
669
static void fat_set_state(struct super_block *sb,
fs/fat/inode.c
674
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
677
if (sb_rdonly(sb) && !force)
fs/fat/inode.c
684
fat_msg(sb, KERN_WARNING, "Volume was not properly "
fs/fat/inode.c
690
bh = sb_bread(sb, 0);
fs/fat/inode.c
692
fat_msg(sb, KERN_ERR, "unable to read boot sector "
fs/fat/inode.c
734
static void fat_put_super(struct super_block *sb)
fs/fat/inode.c
736
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
738
fat_set_state(sb, 0, 0);
fs/fat/inode.c
748
static struct inode *fat_alloc_inode(struct super_block *sb)
fs/fat/inode.c
751
ei = alloc_inode_sb(sb, fat_inode_cachep, GFP_NOFS);
fs/fat/inode.c
811
struct super_block *sb = fc->root->d_sb;
fs/fat/inode.c
812
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
815
sync_filesystem(sb);
fs/fat/inode.c
819
if (new_rdonly != sb_rdonly(sb)) {
fs/fat/inode.c
821
fat_set_state(sb, 0, 0);
fs/fat/inode.c
823
fat_set_state(sb, 1, 1);
fs/fat/inode.c
831
struct super_block *sb = dentry->d_sb;
fs/fat/inode.c
832
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
833
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/fat/inode.c
856
struct super_block *sb = inode->i_sb;
fs/fat/inode.c
857
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/inode.c
874
bh = sb_bread(sb, blocknr);
fs/fat/inode.c
876
fat_msg(sb, KERN_ERR, "unable to read inode block "
fs/fat/inode.c
919
struct super_block *sb = inode->i_sb;
fs/fat/inode.c
921
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/inode.c
922
err = fat_clusters_flush(sb);
fs/fat/inode.c
923
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/misc.c
109
struct super_block *sb = inode->i_sb;
fs/fat/misc.c
110
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/misc.c
162
sb, "clusters badly computed (%d != %llu)", new_fclus,
fs/fat/misc.c
21
void __fat_fs_error(struct super_block *sb, int report, const char *fmt, ...)
fs/fat/misc.c
23
struct fat_mount_options *opts = &MSDOS_SB(sb)->options;
fs/fat/misc.c
31
fat_msg(sb, KERN_ERR, "error, %pV", &vaf);
fs/fat/misc.c
36
panic("FAT-fs (%s): fs panic from previous error\n", sb->s_id);
fs/fat/misc.c
37
else if (opts->errors == FAT_ERRORS_RO && !sb_rdonly(sb)) {
fs/fat/misc.c
38
sb->s_flags |= SB_RDONLY;
fs/fat/misc.c
39
fat_msg(sb, KERN_ERR, "Filesystem has been set read-only");
fs/fat/misc.c
54
void _fat_msg(struct super_block *sb, const char *level, const char *fmt, ...)
fs/fat/misc.c
62
_printk(FAT_PRINTK_PREFIX "%pV\n", level, sb->s_id, &vaf);
fs/fat/misc.c
68
int fat_clusters_flush(struct super_block *sb)
fs/fat/misc.c
70
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/misc.c
77
bh = sb_bread(sb, sbi->fsinfo_sector);
fs/fat/misc.c
79
fat_msg(sb, KERN_ERR, "bread failed in fat_clusters_flush");
fs/fat/misc.c
86
fat_msg(sb, KERN_ERR, "Invalid FSINFO signature: "
fs/fat/namei_msdos.c
204
struct super_block *sb = dir->i_sb;
fs/fat/namei_msdos.c
209
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
216
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_msdos.c
222
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
267
struct super_block *sb = dir->i_sb;
fs/fat/namei_msdos.c
274
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
277
msdos_name, &MSDOS_SB(sb)->options);
fs/fat/namei_msdos.c
292
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_msdos.c
303
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
305
err = fat_flush_inodes(sb, dir, inode);
fs/fat/namei_msdos.c
312
struct super_block *sb = dir->i_sb;
fs/fat/namei_msdos.c
317
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
331
fat_fs_error(sb, "parent dir link count too low (%u)",
fs/fat/namei_msdos.c
338
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
340
err = fat_flush_inodes(sb, dir, inode);
fs/fat/namei_msdos.c
349
struct super_block *sb = dir->i_sb;
fs/fat/namei_msdos.c
356
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
359
msdos_name, &MSDOS_SB(sb)->options);
fs/fat/namei_msdos.c
381
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_msdos.c
394
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
395
fat_flush_inodes(sb, dir, inode);
fs/fat/namei_msdos.c
401
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
409
struct super_block *sb = inode->i_sb;
fs/fat/namei_msdos.c
413
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
424
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
426
err = fat_flush_inodes(sb, dir, inode);
fs/fat/namei_msdos.c
604
struct super_block *sb = old_dir->i_sb;
fs/fat/namei_msdos.c
611
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
630
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_msdos.c
632
err = fat_flush_inodes(sb, old_dir, new_dir);
fs/fat/namei_msdos.c
648
static void setup(struct super_block *sb)
fs/fat/namei_msdos.c
650
MSDOS_SB(sb)->dir_ops = &msdos_dir_inode_operations;
fs/fat/namei_msdos.c
651
set_default_d_op(sb, &msdos_dentry_operations);
fs/fat/namei_msdos.c
652
sb->s_flags |= SB_NOATIME;
fs/fat/namei_msdos.c
655
static int msdos_fill_super(struct super_block *sb, struct fs_context *fc)
fs/fat/namei_msdos.c
657
return fat_fill_super(sb, fc, setup);
fs/fat/namei_vfat.c
1010
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
1071
struct super_block *sb = old_dir->i_sb;
fs/fat/namei_vfat.c
1077
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
1134
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
1191
static void setup(struct super_block *sb)
fs/fat/namei_vfat.c
1193
MSDOS_SB(sb)->dir_ops = &vfat_dir_inode_operations;
fs/fat/namei_vfat.c
1194
if (MSDOS_SB(sb)->options.name_check != 's')
fs/fat/namei_vfat.c
1195
set_default_d_op(sb, &vfat_ci_dentry_ops);
fs/fat/namei_vfat.c
1197
set_default_d_op(sb, &vfat_dentry_ops);
fs/fat/namei_vfat.c
1200
static int vfat_fill_super(struct super_block *sb, struct fs_context *fc)
fs/fat/namei_vfat.c
1202
return fat_fill_super(sb, fc, setup);
fs/fat/namei_vfat.c
702
struct super_block *sb = dir->i_sb;
fs/fat/namei_vfat.c
708
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
719
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_vfat.c
742
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
748
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
753
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
760
struct super_block *sb = dir->i_sb;
fs/fat/namei_vfat.c
766
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
774
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_vfat.c
784
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
791
struct super_block *sb = dir->i_sb;
fs/fat/namei_vfat.c
795
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
810
fat_fs_error(sb, "parent dir link count too low (%u)",
fs/fat/namei_vfat.c
819
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
827
struct super_block *sb = dir->i_sb;
fs/fat/namei_vfat.c
831
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
845
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
853
struct super_block *sb = dir->i_sb;
fs/fat/namei_vfat.c
859
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
873
inode = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/namei_vfat.c
885
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
891
mutex_unlock(&MSDOS_SB(sb)->s_lock);
fs/fat/namei_vfat.c
944
struct super_block *sb = old_dir->i_sb;
fs/fat/namei_vfat.c
949
mutex_lock(&MSDOS_SB(sb)->s_lock);
fs/fat/nfs.c
148
static struct dentry *fat_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/fat/nfs.c
151
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/fat/nfs.c
155
static struct dentry *fat_fh_to_dentry_nostale(struct super_block *sb,
fs/fat/nfs.c
177
inode = __fat_nfs_get_inode(sb, 0, fid->i_gen, i_pos);
fs/fat/nfs.c
186
static struct dentry *fat_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/fat/nfs.c
189
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/fat/nfs.c
193
static struct dentry *fat_fh_to_parent_nostale(struct super_block *sb,
fs/fat/nfs.c
208
inode = __fat_nfs_get_inode(sb, 0, fid->parent_i_gen, i_pos);
fs/fat/nfs.c
220
struct inode *fat_rebuild_parent(struct super_block *sb, int parent_logstart)
fs/fat/nfs.c
227
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/nfs.c
229
struct buffer_head *parent_bh = sb_bread(sb, blknr);
fs/fat/nfs.c
23
static struct inode *fat_dget(struct super_block *sb, int i_logstart)
fs/fat/nfs.c
231
fat_msg(sb, KERN_ERR,
fs/fat/nfs.c
240
dummy_grand_parent = fat_dget(sb, search_clus);
fs/fat/nfs.c
242
dummy_grand_parent = new_inode(sb);
fs/fat/nfs.c
248
dummy_grand_parent->i_ino = iunique(sb, MSDOS_ROOT_INO);
fs/fat/nfs.c
25
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/nfs.c
254
parent = fat_build_inode(sb, sinfo.de, sinfo.i_pos);
fs/fat/nfs.c
270
struct super_block *sb = child_dir->d_sb;
fs/fat/nfs.c
274
struct msdos_sb_info *sbi = MSDOS_SB(sb);
fs/fat/nfs.c
278
parent_inode = fat_dget(sb, parent_logstart);
fs/fat/nfs.c
280
parent_inode = fat_rebuild_parent(sb, parent_logstart);
fs/fat/nfs.c
33
BUG_ON(i->vfs_inode.i_sb != sb);
fs/fat/nfs.c
44
static struct inode *fat_ilookup(struct super_block *sb, u64 ino, loff_t i_pos)
fs/fat/nfs.c
46
if (MSDOS_SB(sb)->options.nfs == FAT_NFS_NOSTALE_RO)
fs/fat/nfs.c
47
return fat_iget(sb, i_pos);
fs/fat/nfs.c
52
return ilookup(sb, ino);
fs/fat/nfs.c
56
static struct inode *__fat_nfs_get_inode(struct super_block *sb,
fs/fat/nfs.c
59
struct inode *inode = fat_ilookup(sb, ino, i_pos);
fs/fat/nfs.c
65
if (inode == NULL && MSDOS_SB(sb)->options.nfs == FAT_NFS_NOSTALE_RO) {
fs/fat/nfs.c
70
fat_get_blknr_offset(MSDOS_SB(sb), i_pos, &blocknr, &offset);
fs/fat/nfs.c
71
bh = sb_bread(sb, blocknr);
fs/fat/nfs.c
73
fat_msg(sb, KERN_ERR,
fs/fat/nfs.c
85
inode = fat_build_inode(sb, &de[offset], i_pos);
fs/fat/nfs.c
92
static struct inode *fat_nfs_get_inode(struct super_block *sb,
fs/fat/nfs.c
96
return __fat_nfs_get_inode(sb, ino, generation, 0);
fs/freevxfs/vxfs_bmap.c
45
struct super_block *sb = ip->i_sb;
fs/freevxfs/vxfs_bmap.c
47
struct vxfs_sb_info *sbi = VXFS_SBI(sb);
fs/freevxfs/vxfs_bmap.c
48
unsigned long bsize = sb->s_blocksize;
fs/freevxfs/vxfs_bmap.c
52
if (indsize > sb->s_blocksize)
fs/freevxfs/vxfs_bmap.c
67
buf = sb_bread(sb,
fs/freevxfs/vxfs_super.c
101
static struct inode *vxfs_alloc_inode(struct super_block *sb)
fs/freevxfs/vxfs_super.c
105
vi = alloc_inode_sb(sb, vxfs_inode_cachep, GFP_KERNEL);
fs/fs-writeback.c
1209
void cgroup_writeback_umount(struct super_block *sb)
fs/fs-writeback.c
1212
if (!(sb->s_bdi->capabilities & BDI_CAP_WRITEBACK))
fs/fs-writeback.c
1386
struct super_block *sb = inode->i_sb;
fs/fs-writeback.c
1390
spin_lock_irqsave(&sb->s_inode_wblist_lock, flags);
fs/fs-writeback.c
1392
list_add_tail(&inode->i_wb_list, &sb->s_inodes_wb);
fs/fs-writeback.c
1395
spin_unlock_irqrestore(&sb->s_inode_wblist_lock, flags);
fs/fs-writeback.c
1404
struct super_block *sb = inode->i_sb;
fs/fs-writeback.c
1408
spin_lock_irqsave(&sb->s_inode_wblist_lock, flags);
fs/fs-writeback.c
1413
spin_unlock_irqrestore(&sb->s_inode_wblist_lock, flags);
fs/fs-writeback.c
1502
struct super_block *sb = NULL;
fs/fs-writeback.c
1518
if (sb && sb != inode->i_sb)
fs/fs-writeback.c
1520
sb = inode->i_sb;
fs/fs-writeback.c
1536
sb = wb_inode(tmp.prev)->i_sb;
fs/fs-writeback.c
1539
if (inode->i_sb == sb)
fs/fs-writeback.c
1918
static long writeback_chunk_size(struct super_block *sb,
fs/fs-writeback.c
1942
return round_down(pages + sb->s_min_writeback_pages,
fs/fs-writeback.c
1943
sb->s_min_writeback_pages);
fs/fs-writeback.c
1955
static long writeback_sb_inodes(struct super_block *sb,
fs/fs-writeback.c
1984
if (inode->i_sb != sb) {
fs/fs-writeback.c
1985
if (work->sb) {
fs/fs-writeback.c
2119
struct super_block *sb = inode->i_sb;
fs/fs-writeback.c
2121
if (!super_trylock_shared(sb)) {
fs/fs-writeback.c
2130
wrote += writeback_sb_inodes(sb, wb, work);
fs/fs-writeback.c
2131
up_read(&sb->s_umount);
fs/fs-writeback.c
2238
if (work->sb)
fs/fs-writeback.c
2239
progress = writeback_sb_inodes(work->sb, wb, work);
fs/fs-writeback.c
2588
struct super_block *sb = inode->i_sb;
fs/fs-writeback.c
2620
if (sb->s_op->dirty_inode) {
fs/fs-writeback.c
2621
sb->s_op->dirty_inode(inode,
fs/fs-writeback.c
2741
static void wait_sb_inodes(struct super_block *sb)
fs/fs-writeback.c
2749
WARN_ON(!rwsem_is_locked(&sb->s_umount));
fs/fs-writeback.c
2751
mutex_lock(&sb->s_sync_lock);
fs/fs-writeback.c
2763
spin_lock_irq(&sb->s_inode_wblist_lock);
fs/fs-writeback.c
2764
list_splice_init(&sb->s_inodes_wb, &sync_list);
fs/fs-writeback.c
2784
list_move_tail(&inode->i_wb_list, &sb->s_inodes_wb);
fs/fs-writeback.c
2794
spin_unlock_irq(&sb->s_inode_wblist_lock);
fs/fs-writeback.c
2800
spin_lock_irq(&sb->s_inode_wblist_lock);
fs/fs-writeback.c
2819
spin_lock_irq(&sb->s_inode_wblist_lock);
fs/fs-writeback.c
2821
spin_unlock_irq(&sb->s_inode_wblist_lock);
fs/fs-writeback.c
2823
mutex_unlock(&sb->s_sync_lock);
fs/fs-writeback.c
2826
static void __writeback_inodes_sb_nr(struct super_block *sb, unsigned long nr,
fs/fs-writeback.c
2829
struct backing_dev_info *bdi = sb->s_bdi;
fs/fs-writeback.c
2832
.sb = sb,
fs/fs-writeback.c
2842
WARN_ON(!rwsem_is_locked(&sb->s_umount));
fs/fs-writeback.c
2844
bdi_split_work_to_wbs(sb->s_bdi, &work, skip_if_busy);
fs/fs-writeback.c
2858
void writeback_inodes_sb_nr(struct super_block *sb,
fs/fs-writeback.c
2862
__writeback_inodes_sb_nr(sb, nr, reason, false);
fs/fs-writeback.c
2875
void writeback_inodes_sb(struct super_block *sb, enum wb_reason reason)
fs/fs-writeback.c
2877
writeback_inodes_sb_nr(sb, get_nr_dirty_pages(), reason);
fs/fs-writeback.c
2888
void try_to_writeback_inodes_sb(struct super_block *sb, enum wb_reason reason)
fs/fs-writeback.c
2890
if (!down_read_trylock(&sb->s_umount))
fs/fs-writeback.c
2893
__writeback_inodes_sb_nr(sb, get_nr_dirty_pages(), reason, true);
fs/fs-writeback.c
2894
up_read(&sb->s_umount);
fs/fs-writeback.c
2905
void sync_inodes_sb(struct super_block *sb)
fs/fs-writeback.c
2907
struct backing_dev_info *bdi = sb->s_bdi;
fs/fs-writeback.c
2910
.sb = sb,
fs/fs-writeback.c
2932
if (sb->s_iflags & SB_I_NO_DATA_INTEGRITY) {
fs/fs-writeback.c
2937
WARN_ON(!rwsem_is_locked(&sb->s_umount));
fs/fs-writeback.c
2945
wait_sb_inodes(sb);
fs/fs-writeback.c
40
struct super_block *sb;
fs/fs_context.c
353
struct super_block *sb = fc->root->d_sb;
fs/fs_context.c
356
deactivate_locked_super(sb);
fs/fs_context.c
489
struct super_block *sb;
fs/fs_context.c
492
sb = fc->root->d_sb;
fs/fs_context.c
495
deactivate_super(sb);
fs/fserror.c
105
fserror_pending_dec(sb);
fs/fserror.c
133
void fserror_report(struct super_block *sb, struct inode *inode,
fs/fserror.c
140
WARN_ON_ONCE(inode && inode->i_sb != sb);
fs/fserror.c
145
event = fserror_alloc_event(sb, gfp);
fs/fserror.c
15
void fserror_mount(struct super_block *sb)
fs/fserror.c
180
sb->s_id, inode->i_ino, type, pos, len, error);
fs/fserror.c
184
sb->s_id, type, error);
fs/fserror.c
21
refcount_set(&sb->s_pending_errors, 1);
fs/fserror.c
24
void fserror_unmount(struct super_block *sb)
fs/fserror.c
31
if (!refcount_dec_and_test(&sb->s_pending_errors))
fs/fserror.c
32
wait_var_event(&sb->s_pending_errors,
fs/fserror.c
33
refcount_read(&sb->s_pending_errors) < 1);
fs/fserror.c
36
static inline void fserror_pending_dec(struct super_block *sb)
fs/fserror.c
38
if (refcount_dec_and_test(&sb->s_pending_errors))
fs/fserror.c
39
wake_up_var(&sb->s_pending_errors);
fs/fserror.c
44
fserror_pending_dec(event->sb);
fs/fserror.c
52
struct super_block *sb = event->sb;
fs/fserror.c
54
if (sb->s_flags & SB_ACTIVE) {
fs/fserror.c
59
.sb = event->sb,
fs/fserror.c
62
if (sb->s_op->report_error)
fs/fserror.c
63
sb->s_op->report_error(event);
fs/fserror.c
73
static inline struct fserror_event *fserror_alloc_event(struct super_block *sb,
fs/fserror.c
88
if (!refcount_inc_not_zero(&sb->s_pending_errors))
fs/fserror.c
90
if (!(sb->s_flags & SB_ACTIVE))
fs/fserror.c
99
event->sb = sb;
fs/fsopen.c
219
struct super_block *sb;
fs/fsopen.c
237
sb = fc->root->d_sb;
fs/fsopen.c
238
ret = security_sb_kern_mount(sb);
fs/fsopen.c
246
up_write(&sb->s_umount);
fs/fsopen.c
253
struct super_block *sb;
fs/fsopen.c
261
sb = fc->root->d_sb;
fs/fsopen.c
262
if (!ns_capable(sb->s_user_ns, CAP_SYS_ADMIN)) {
fs/fsopen.c
267
down_write(&sb->s_umount);
fs/fsopen.c
269
up_write(&sb->s_umount);
fs/fuse/control.c
308
static int fuse_ctl_fill_super(struct super_block *sb, struct fs_context *fsc)
fs/fuse/control.c
314
err = simple_fill_super(sb, FUSE_CTL_SUPER_MAGIC, &empty_descr);
fs/fuse/control.c
320
fuse_control_sb = sb;
fs/fuse/control.c
349
static void fuse_ctl_kill_sb(struct super_block *sb)
fs/fuse/control.c
355
kill_anon_super(sb);
fs/fuse/dax.c
1274
bool fuse_dax_inode_alloc(struct super_block *sb, struct fuse_inode *fi)
fs/fuse/dax.c
1276
struct fuse_conn *fc = get_fuse_conn_super(sb);
fs/fuse/dax.c
758
struct super_block *sb = inode->i_sb;
fs/fuse/dax.c
766
sb_start_pagefault(sb);
fs/fuse/dax.c
791
sb_end_pagefault(sb);
fs/fuse/dev.c
200
bool no_idmap = !fm->sb || (fm->sb->s_iflags & SB_I_NOIDMAP);
fs/fuse/dev.c
641
if (!req->fm->sb || req->fm->sb->s_iflags & SB_I_NOIDMAP) {
fs/fuse/dir.c
1710
WARN_ON_ONCE(!(fm->sb->s_iflags & SB_I_NOIDMAP));
fs/fuse/dir.c
211
shrink_dcache_sb(fm->sb);
fs/fuse/dir.c
550
int fuse_lookup_name(struct super_block *sb, u64 nodeid, const struct qstr *name,
fs/fuse/dir.c
553
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/dir.c
587
*inode = fuse_iget(sb, outarg->nodeid, outarg->generation,
fs/fuse/fuse_i.h
1015
struct super_block *sb;
fs/fuse/fuse_i.h
1035
static inline struct fuse_mount *get_fuse_mount_super(struct super_block *sb)
fs/fuse/fuse_i.h
1037
return sb->s_fs_info;
fs/fuse/fuse_i.h
1040
static inline struct fuse_conn *get_fuse_conn_super(struct super_block *sb)
fs/fuse/fuse_i.h
1042
return get_fuse_mount_super(sb)->fc;
fs/fuse/fuse_i.h
1143
struct inode *fuse_iget(struct super_block *sb, u64 nodeid,
fs/fuse/fuse_i.h
1148
int fuse_lookup_name(struct super_block *sb, u64 nodeid, const struct qstr *name,
fs/fuse/fuse_i.h
1354
int fuse_fill_super_common(struct super_block *sb, struct fuse_fs_context *ctx);
fs/fuse/fuse_i.h
1520
bool fuse_dax_inode_alloc(struct super_block *sb, struct fuse_inode *fi);
fs/fuse/inode.c
102
fi = alloc_inode_sb(sb, fuse_inode_cachep, GFP_KERNEL);
fs/fuse/inode.c
1057
static struct inode *fuse_get_root_inode(struct super_block *sb, unsigned int mode)
fs/fuse/inode.c
1065
return fuse_iget(sb, FUSE_ROOT_ID, 0, &attr, 0, 0, 0);
fs/fuse/inode.c
1073
static struct dentry *fuse_get_dentry(struct super_block *sb,
fs/fuse/inode.c
1076
struct fuse_conn *fc = get_fuse_conn_super(sb);
fs/fuse/inode.c
1084
inode = ilookup5(sb, handle->nodeid, fuse_inode_eq, &handle->nodeid);
fs/fuse/inode.c
1092
err = fuse_lookup_name(sb, handle->nodeid, &name, &outarg,
fs/fuse/inode.c
1152
static struct dentry *fuse_fh_to_dentry(struct super_block *sb,
fs/fuse/inode.c
1164
return fuse_get_dentry(sb, &handle);
fs/fuse/inode.c
1167
static struct dentry *fuse_fh_to_parent(struct super_block *sb,
fs/fuse/inode.c
117
if (IS_ENABLED(CONFIG_FUSE_DAX) && !fuse_dax_inode_alloc(sb, fi))
fs/fuse/inode.c
1178
return fuse_get_dentry(sb, &parent);
fs/fuse/inode.c
1382
fm->sb->s_time_gran = arg->time_gran;
fs/fuse/inode.c
1413
fm->sb->s_flags |= SB_NOSEC;
fs/fuse/inode.c
1444
fm->sb->s_stack_depth = arg->max_stack_depth;
fs/fuse/inode.c
1447
fm->sb->s_export_op = &fuse_export_fid_operations;
fs/fuse/inode.c
1450
fm->sb->s_iflags &= ~SB_I_NOIDMAP;
fs/fuse/inode.c
1467
fm->sb->s_bdi->ra_pages =
fs/fuse/inode.c
1468
min(fm->sb->s_bdi->ra_pages, ra_pages);
fs/fuse/inode.c
1494
ia->in.max_readahead = fm->sb->s_bdi->ra_pages * PAGE_SIZE;
fs/fuse/inode.c
1578
static int fuse_bdi_init(struct fuse_conn *fc, struct super_block *sb)
fs/fuse/inode.c
1583
if (sb->s_bdev) {
fs/fuse/inode.c
1589
bdi_put(sb->s_bdi);
fs/fuse/inode.c
1590
sb->s_bdi = &noop_backing_dev_info;
fs/fuse/inode.c
1592
err = super_setup_bdi_name(sb, "%u:%u%s", MAJOR(fc->dev),
fs/fuse/inode.c
1597
sb->s_bdi->capabilities |= BDI_CAP_STRICTLIMIT;
fs/fuse/inode.c
1611
bdi_set_max_ratio(sb->s_bdi, 1);
fs/fuse/inode.c
1702
static void fuse_sb_defaults(struct super_block *sb)
fs/fuse/inode.c
1704
sb->s_magic = FUSE_SUPER_MAGIC;
fs/fuse/inode.c
1705
sb->s_op = &fuse_super_operations;
fs/fuse/inode.c
1706
sb->s_xattr = fuse_xattr_handlers;
fs/fuse/inode.c
1707
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/fuse/inode.c
1708
sb->s_time_gran = 1;
fs/fuse/inode.c
1709
sb->s_export_op = &fuse_export_operations;
fs/fuse/inode.c
1710
sb->s_iflags |= SB_I_IMA_UNVERIFIABLE_SIGNATURE;
fs/fuse/inode.c
1711
sb->s_iflags |= SB_I_NOIDMAP;
fs/fuse/inode.c
1712
sb->s_iflags |= SB_I_NO_DATA_INTEGRITY;
fs/fuse/inode.c
1713
if (sb->s_user_ns != &init_user_ns)
fs/fuse/inode.c
1714
sb->s_iflags |= SB_I_UNTRUSTED_MOUNTER;
fs/fuse/inode.c
1715
sb->s_flags &= ~(SB_NOSEC | SB_I_VERSION);
fs/fuse/inode.c
1718
static int fuse_fill_super_submount(struct super_block *sb,
fs/fuse/inode.c
1721
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
1728
fuse_sb_defaults(sb);
fs/fuse/inode.c
1729
fm->sb = sb;
fs/fuse/inode.c
1731
WARN_ON(sb->s_bdi != &noop_backing_dev_info);
fs/fuse/inode.c
1732
sb->s_bdi = bdi_get(parent_sb->s_bdi);
fs/fuse/inode.c
1734
sb->s_xattr = parent_sb->s_xattr;
fs/fuse/inode.c
1735
sb->s_export_op = parent_sb->s_export_op;
fs/fuse/inode.c
1736
sb->s_time_gran = parent_sb->s_time_gran;
fs/fuse/inode.c
1737
sb->s_blocksize = parent_sb->s_blocksize;
fs/fuse/inode.c
1738
sb->s_blocksize_bits = parent_sb->s_blocksize_bits;
fs/fuse/inode.c
1739
sb->s_subtype = kstrdup(parent_sb->s_subtype, GFP_KERNEL);
fs/fuse/inode.c
1740
if (parent_sb->s_subtype && !sb->s_subtype)
fs/fuse/inode.c
1744
root = fuse_iget(sb, parent_fi->nodeid, 0, &root_attr, 0, 0,
fs/fuse/inode.c
1754
set_default_d_op(sb, &fuse_dentry_operations);
fs/fuse/inode.c
1755
sb->s_root = d_make_root(root);
fs/fuse/inode.c
1756
if (!sb->s_root)
fs/fuse/inode.c
1781
struct super_block *sb;
fs/fuse/inode.c
1790
sb = sget_fc(fsc, NULL, set_anon_super_fc);
fs/fuse/inode.c
1793
if (IS_ERR(sb))
fs/fuse/inode.c
1794
return PTR_ERR(sb);
fs/fuse/inode.c
1797
err = fuse_fill_super_submount(sb, mp_fi);
fs/fuse/inode.c
1799
deactivate_locked_super(sb);
fs/fuse/inode.c
1807
sb->s_flags |= SB_ACTIVE;
fs/fuse/inode.c
1808
fsc->root = dget(sb->s_root);
fs/fuse/inode.c
1824
int fuse_fill_super_common(struct super_block *sb, struct fuse_fs_context *ctx)
fs/fuse/inode.c
1827
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
1834
if (sb->s_flags & SB_MANDLOCK)
fs/fuse/inode.c
1838
fuse_sb_defaults(sb);
fs/fuse/inode.c
1843
if (!sb_set_blocksize(sb, ctx->blksize))
fs/fuse/inode.c
1848
sb->s_blocksize = PAGE_SIZE;
fs/fuse/inode.c
1849
sb->s_blocksize_bits = PAGE_SHIFT;
fs/fuse/inode.c
1852
sb->s_subtype = ctx->subtype;
fs/fuse/inode.c
1867
fc->dev = sb->s_dev;
fs/fuse/inode.c
1868
fm->sb = sb;
fs/fuse/inode.c
1869
err = fuse_bdi_init(fc, sb);
fs/fuse/inode.c
1874
if (sb->s_flags & SB_POSIXACL)
fs/fuse/inode.c
1876
sb->s_flags |= SB_POSIXACL;
fs/fuse/inode.c
1889
root = fuse_get_root_inode(sb, ctx->rootmode);
fs/fuse/inode.c
1890
set_default_d_op(sb, &fuse_dentry_operations);
fs/fuse/inode.c
1909
sb->s_root = root_dentry;
fs/fuse/inode.c
1931
static int fuse_fill_super(struct super_block *sb, struct fs_context *fsc)
fs/fuse/inode.c
1946
(ctx->file->f_cred->user_ns != sb->s_user_ns))
fs/fuse/inode.c
1950
err = fuse_fill_super_common(sb, ctx);
fs/fuse/inode.c
1956
fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
1965
static int fuse_set_no_super(struct super_block *sb, struct fs_context *fsc)
fs/fuse/inode.c
1970
static int fuse_test_super(struct super_block *sb, struct fs_context *fsc)
fs/fuse/inode.c
1973
return fsc->sget_key == get_fuse_conn_super(sb);
fs/fuse/inode.c
1982
struct super_block *sb;
fs/fuse/inode.c
2022
sb = sget_fc(fsc, fuse_test_super, fuse_set_no_super);
fs/fuse/inode.c
2023
err = PTR_ERR_OR_ZERO(sb);
fs/fuse/inode.c
2024
if (!IS_ERR(sb))
fs/fuse/inode.c
2025
fsc->root = dget(sb->s_root);
fs/fuse/inode.c
203
struct super_block *sb = fsc->root->d_sb;
fs/fuse/inode.c
205
sync_filesystem(sb);
fs/fuse/inode.c
2105
static void fuse_sb_destroy(struct super_block *sb)
fs/fuse/inode.c
2107
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
2110
if (sb->s_root) {
fs/fuse/inode.c
2124
static void fuse_kill_sb_anon(struct super_block *sb)
fs/fuse/inode.c
2126
fuse_sb_destroy(sb);
fs/fuse/inode.c
2127
kill_anon_super(sb);
fs/fuse/inode.c
2128
fuse_mount_destroy(get_fuse_mount_super(sb));
fs/fuse/inode.c
2142
static void fuse_kill_sb_blk(struct super_block *sb)
fs/fuse/inode.c
2144
fuse_sb_destroy(sb);
fs/fuse/inode.c
2145
kill_block_super(sb);
fs/fuse/inode.c
2146
fuse_mount_destroy(get_fuse_mount_super(sb));
fs/fuse/inode.c
465
struct inode *fuse_iget(struct super_block *sb, u64 nodeid,
fs/fuse/inode.c
472
struct fuse_conn *fc = get_fuse_conn_super(sb);
fs/fuse/inode.c
485
inode = new_inode(sb);
fs/fuse/inode.c
504
inode = iget5_locked(sb, nodeid, fuse_inode_eq, fuse_inode_set, &nodeid);
fs/fuse/inode.c
518
if (inode != d_inode(sb->s_root)) {
fs/fuse/inode.c
542
if (!fm_iter->sb)
fs/fuse/inode.c
545
inode = ilookup5(fm_iter->sb, nodeid, fuse_inode_eq, &nodeid);
fs/fuse/inode.c
617
static void fuse_umount_begin(struct super_block *sb)
fs/fuse/inode.c
619
struct fuse_conn *fc = get_fuse_conn_super(sb);
fs/fuse/inode.c
627
if (sb->s_bdev != NULL)
fs/fuse/inode.c
628
retire_super(sb);
fs/fuse/inode.c
659
struct super_block *sb = dentry->d_sb;
fs/fuse/inode.c
660
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
734
static int fuse_sync_fs(struct super_block *sb, int wait)
fs/fuse/inode.c
736
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/inode.c
750
if (!sb->s_root)
fs/fuse/inode.c
763
args.nodeid = get_node_id(sb->s_root->d_inode);
fs/fuse/inode.c
914
struct super_block *sb = root->d_sb;
fs/fuse/inode.c
915
struct fuse_conn *fc = get_fuse_conn_super(sb);
fs/fuse/inode.c
928
if (sb->s_bdev && sb->s_blocksize != FUSE_DEFAULT_BLKSIZE)
fs/fuse/inode.c
929
seq_printf(m, ",blksize=%lu", sb->s_blocksize);
fs/fuse/inode.c
98
static struct inode *fuse_alloc_inode(struct super_block *sb)
fs/fuse/virtio_fs.c
1540
static int virtio_fs_fill_super(struct super_block *sb, struct fs_context *fsc)
fs/fuse/virtio_fs.c
1542
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/virtio_fs.c
1583
err = fuse_fill_super_common(sb, ctx);
fs/fuse/virtio_fs.c
1636
static void virtio_kill_sb(struct super_block *sb)
fs/fuse/virtio_fs.c
1638
struct fuse_mount *fm = get_fuse_mount_super(sb);
fs/fuse/virtio_fs.c
1642
if (sb->s_root) {
fs/fuse/virtio_fs.c
1647
kill_anon_super(sb);
fs/fuse/virtio_fs.c
1651
static int virtio_fs_test_super(struct super_block *sb,
fs/fuse/virtio_fs.c
1655
struct fuse_mount *sb_fm = get_fuse_mount_super(sb);
fs/fuse/virtio_fs.c
1663
struct super_block *sb;
fs/fuse/virtio_fs.c
1707
sb = sget_fc(fsc, virtio_fs_test_super, set_anon_super_fc);
fs/fuse/virtio_fs.c
1710
if (IS_ERR(sb))
fs/fuse/virtio_fs.c
1711
return PTR_ERR(sb);
fs/fuse/virtio_fs.c
1713
if (!sb->s_root) {
fs/fuse/virtio_fs.c
1714
err = virtio_fs_fill_super(sb, fsc);
fs/fuse/virtio_fs.c
1716
deactivate_locked_super(sb);
fs/fuse/virtio_fs.c
1720
sb->s_flags |= SB_ACTIVE;
fs/fuse/virtio_fs.c
1724
fsc->root = dget(sb->s_root);
fs/gfs2/export.c
131
static struct dentry *gfs2_get_dentry(struct super_block *sb,
fs/gfs2/export.c
134
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/export.c
144
static struct dentry *gfs2_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/gfs2/export.c
160
return gfs2_get_dentry(sb, &this);
fs/gfs2/export.c
166
static struct dentry *gfs2_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/gfs2/export.c
181
return gfs2_get_dentry(sb, &parent);
fs/gfs2/export.c
32
struct super_block *sb = inode->i_sb;
fs/gfs2/export.c
49
if (!parent || inode == d_inode(sb->s_root))
fs/gfs2/file.c
1179
struct super_block *sb = inode->i_sb;
fs/gfs2/file.c
1206
error = sb_issue_zeroout(sb, iomap.addr >> inode->i_blkbits,
fs/gfs2/glock.c
2651
struct super_block *sb;
fs/gfs2/glock.c
2697
if (file_inode(i->file)->i_sb == i->sb)
fs/gfs2/glock.c
2800
i->sb = sdp->sd_vfs;
fs/gfs2/glops.c
571
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/glops.c
585
if (down_read_trylock(&sb->s_umount)) {
fs/gfs2/glops.c
586
atomic_inc(&sb->s_active);
fs/gfs2/glops.c
587
up_read(&sb->s_umount);
fs/gfs2/glops.c
589
deactivate_super(sb);
fs/gfs2/inode.c
127
struct inode *gfs2_inode_lookup(struct super_block *sb, unsigned int type,
fs/gfs2/inode.c
137
inode = iget5_locked(sb, no_addr, iget_test, iget_set, &no_addr);
fs/gfs2/inode.c
1419
struct super_block *sb = dir->i_sb;
fs/gfs2/inode.c
1430
if (dir == d_inode(sb->s_root)) {
fs/gfs2/inode.c
261
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/inode.c
265
inode = gfs2_inode_lookup(sb, DT_UNKNOWN, no_addr, no_formal_ino,
fs/gfs2/inode.c
325
struct super_block *sb = dir->i_sb;
fs/gfs2/inode.c
337
dir == d_inode(sb->s_root))) {
fs/gfs2/inode.h
90
struct inode *gfs2_inode_lookup(struct super_block *sb, unsigned type,
fs/gfs2/log.c
830
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/log.c
888
sb->s_blocksize - LH_V1_SIZE - 4);
fs/gfs2/log.c
891
gfs2_log_write(sdp, jd, page, sb->s_blocksize, 0, dblock,
fs/gfs2/log.c
907
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/log.c
909
gfs2_assert_withdraw(sdp, sb->s_writers.frozen != SB_FREEZE_COMPLETE);
fs/gfs2/lops.c
265
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/lops.c
266
struct bio *bio = bio_alloc(sb->s_bdev, BIO_MAX_VECS, opf, GFP_NOIO);
fs/gfs2/lops.c
378
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/lops.c
383
gfs2_log_write(sdp, sdp->sd_jdesc, page, sb->s_blocksize, 0, dblock,
fs/gfs2/ops_fstype.c
1051
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/ops_fstype.c
1055
sprintf(ro, "RDONLY=%d", sb_rdonly(sb));
fs/gfs2/ops_fstype.c
1112
static int gfs2_fill_super(struct super_block *sb, struct fs_context *fc)
fs/gfs2/ops_fstype.c
1121
sdp = init_sbd(sb);
fs/gfs2/ops_fstype.c
1129
sb->s_flags |= SB_RDONLY;
fs/gfs2/ops_fstype.c
1133
sb->s_flags |= SB_POSIXACL;
fs/gfs2/ops_fstype.c
1137
sb->s_flags |= SB_NOSEC;
fs/gfs2/ops_fstype.c
1138
sb->s_magic = GFS2_MAGIC;
fs/gfs2/ops_fstype.c
1139
sb->s_op = &gfs2_super_ops;
fs/gfs2/ops_fstype.c
1141
set_default_d_op(sb, &gfs2_dops);
fs/gfs2/ops_fstype.c
1142
sb->s_export_op = &gfs2_export_ops;
fs/gfs2/ops_fstype.c
1143
sb->s_qcop = &gfs2_quotactl_ops;
fs/gfs2/ops_fstype.c
1144
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
fs/gfs2/ops_fstype.c
1145
sb_dqopt(sb)->flags |= DQUOT_QUOTA_SYS_FILE;
fs/gfs2/ops_fstype.c
1146
sb->s_time_gran = 1;
fs/gfs2/ops_fstype.c
1147
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/gfs2/ops_fstype.c
1151
sdp->sd_sb.sb_bsize = sb_min_blocksize(sb, SECTOR_SIZE);
fs/gfs2/ops_fstype.c
1155
sdp->sd_sb.sb_bsize_shift = sb->s_blocksize_bits;
fs/gfs2/ops_fstype.c
1170
sdp->sd_inode = new_inode(sb);
fs/gfs2/ops_fstype.c
1264
if (!sb_rdonly(sb)) {
fs/gfs2/ops_fstype.c
1274
if (!sb_rdonly(sb))
fs/gfs2/ops_fstype.c
1296
if (sb->s_root)
fs/gfs2/ops_fstype.c
1297
dput(sb->s_root);
fs/gfs2/ops_fstype.c
1298
sb->s_root = NULL;
fs/gfs2/ops_fstype.c
151
struct gfs2_sb_host *sb = &sdp->sd_sb;
fs/gfs2/ops_fstype.c
153
if (sb->sb_magic != GFS2_MAGIC ||
fs/gfs2/ops_fstype.c
154
sb->sb_type != GFS2_METATYPE_SB) {
fs/gfs2/ops_fstype.c
1547
struct super_block *sb = fc->root->d_sb;
fs/gfs2/ops_fstype.c
1548
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/ops_fstype.c
1554
sync_filesystem(sb);
fs/gfs2/ops_fstype.c
1592
if ((sb->s_flags ^ fc->sb_flags) & SB_RDONLY) {
fs/gfs2/ops_fstype.c
160
if (sb->sb_fs_format < GFS2_FS_FORMAT_MIN ||
fs/gfs2/ops_fstype.c
1604
sb->s_flags |= SB_POSIXACL;
fs/gfs2/ops_fstype.c
1606
sb->s_flags &= ~SB_POSIXACL;
fs/gfs2/ops_fstype.c
161
sb->sb_fs_format > GFS2_FS_FORMAT_MAX ||
fs/gfs2/ops_fstype.c
162
sb->sb_multihost_format != GFS2_FORMAT_MULTI) {
fs/gfs2/ops_fstype.c
167
if (sb->sb_bsize < SECTOR_SIZE || sb->sb_bsize > PAGE_SIZE ||
fs/gfs2/ops_fstype.c
168
(sb->sb_bsize & (sb->sb_bsize - 1))) {
fs/gfs2/ops_fstype.c
172
if (sb->sb_bsize_shift != ffs(sb->sb_bsize) - 1) {
fs/gfs2/ops_fstype.c
1741
static void gfs2_evict_inodes(struct super_block *sb)
fs/gfs2/ops_fstype.c
1744
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/ops_fstype.c
1748
spin_lock(&sb->s_inode_list_lock);
fs/gfs2/ops_fstype.c
1749
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/gfs2/ops_fstype.c
1758
spin_unlock(&sb->s_inode_list_lock);
fs/gfs2/ops_fstype.c
1764
spin_lock(&sb->s_inode_list_lock);
fs/gfs2/ops_fstype.c
1766
spin_unlock(&sb->s_inode_list_lock);
fs/gfs2/ops_fstype.c
1770
static void gfs2_kill_sb(struct super_block *sb)
fs/gfs2/ops_fstype.c
1772
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/ops_fstype.c
1775
kill_block_super(sb);
fs/gfs2/ops_fstype.c
1784
shrink_dcache_sb(sb);
fs/gfs2/ops_fstype.c
1786
gfs2_evict_inodes(sb);
fs/gfs2/ops_fstype.c
1797
kill_block_super(sb);
fs/gfs2/ops_fstype.c
181
struct gfs2_sb_host *sb = &sdp->sd_sb;
fs/gfs2/ops_fstype.c
184
sb->sb_magic = be32_to_cpu(str->sb_header.mh_magic);
fs/gfs2/ops_fstype.c
185
sb->sb_type = be32_to_cpu(str->sb_header.mh_type);
fs/gfs2/ops_fstype.c
186
sb->sb_fs_format = be32_to_cpu(str->sb_fs_format);
fs/gfs2/ops_fstype.c
187
sb->sb_multihost_format = be32_to_cpu(str->sb_multihost_format);
fs/gfs2/ops_fstype.c
188
sb->sb_bsize = be32_to_cpu(str->sb_bsize);
fs/gfs2/ops_fstype.c
189
sb->sb_bsize_shift = be32_to_cpu(str->sb_bsize_shift);
fs/gfs2/ops_fstype.c
190
sb->sb_master_dir.no_addr = be64_to_cpu(str->sb_master_dir.no_addr);
fs/gfs2/ops_fstype.c
191
sb->sb_master_dir.no_formal_ino = be64_to_cpu(str->sb_master_dir.no_formal_ino);
fs/gfs2/ops_fstype.c
192
sb->sb_root_dir.no_addr = be64_to_cpu(str->sb_root_dir.no_addr);
fs/gfs2/ops_fstype.c
193
sb->sb_root_dir.no_formal_ino = be64_to_cpu(str->sb_root_dir.no_formal_ino);
fs/gfs2/ops_fstype.c
195
memcpy(sb->sb_lockproto, str->sb_lockproto, GFS2_LOCKNAME_LEN);
fs/gfs2/ops_fstype.c
196
memcpy(sb->sb_locktable, str->sb_locktable, GFS2_LOCKNAME_LEN);
fs/gfs2/ops_fstype.c
221
struct gfs2_sb *sb;
fs/gfs2/ops_fstype.c
224
sb = kmalloc(PAGE_SIZE, GFP_KERNEL);
fs/gfs2/ops_fstype.c
225
if (unlikely(!sb))
fs/gfs2/ops_fstype.c
229
sb, PAGE_SIZE, REQ_OP_READ | REQ_META);
fs/gfs2/ops_fstype.c
232
kfree(sb);
fs/gfs2/ops_fstype.c
235
gfs2_sb_in(sdp, sb);
fs/gfs2/ops_fstype.c
236
kfree(sb);
fs/gfs2/ops_fstype.c
419
static int gfs2_lookup_root(struct super_block *sb, struct dentry **dptr,
fs/gfs2/ops_fstype.c
422
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/ops_fstype.c
426
inode = gfs2_inode_lookup(sb, DT_DIR, no_addr, 0,
fs/gfs2/ops_fstype.c
443
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/ops_fstype.c
463
sb->s_xattr = gfs2_xattr_handlers_max;
fs/gfs2/ops_fstype.c
467
sb->s_xattr = gfs2_xattr_handlers_min;
fs/gfs2/ops_fstype.c
475
if (sdp->sd_sb.sb_bsize < bdev_logical_block_size(sb->s_bdev)) {
fs/gfs2/ops_fstype.c
479
sdp->sd_sb.sb_bsize, bdev_logical_block_size(sb->s_bdev));
fs/gfs2/ops_fstype.c
490
if (!sb_set_blocksize(sb, sdp->sd_sb.sb_bsize))
fs/gfs2/ops_fstype.c
495
ret = gfs2_lookup_root(sb, &sdp->sd_root_dir, no_addr, "root");
fs/gfs2/ops_fstype.c
501
ret = gfs2_lookup_root(sb, &sdp->sd_master_dir, no_addr, "master");
fs/gfs2/ops_fstype.c
506
sb->s_root = dget(sdp->sd_args.ar_meta ? sdp->sd_master_dir : sdp->sd_root_dir);
fs/gfs2/ops_fstype.c
68
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/ops_fstype.c
71
sb->s_fs_info = NULL;
fs/gfs2/ops_fstype.c
75
static struct gfs2_sbd *init_sbd(struct super_block *sb)
fs/gfs2/ops_fstype.c
83
sdp->sd_vfs = sb;
fs/gfs2/ops_fstype.c
87
sb->s_fs_info = sdp;
fs/gfs2/quota.c
1310
int gfs2_quota_sync(struct super_block *sb, int type)
fs/gfs2/quota.c
1312
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/quota.c
1629
static int gfs2_quota_get_state(struct super_block *sb, struct qc_state *state)
fs/gfs2/quota.c
1631
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/quota.c
1662
static int gfs2_get_dqblk(struct super_block *sb, struct kqid qid,
fs/gfs2/quota.c
1665
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/quota.c
1701
static int gfs2_set_dqblk(struct super_block *sb, struct kqid qid,
fs/gfs2/quota.c
1704
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/quota.h
31
int gfs2_quota_sync(struct super_block *sb, int type);
fs/gfs2/rgrp.c
1313
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/rgrp.c
1342
rv = sb_issue_discard(sb,
fs/gfs2/rgrp.c
1360
rv = sb_issue_discard(sb, start, nr_blks, GFP_NOFS, 0);
fs/gfs2/super.c
1240
struct super_block *sb = inode->i_sb;
fs/gfs2/super.c
1241
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
1333
struct super_block *sb = inode->i_sb;
fs/gfs2/super.c
1334
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
1383
struct super_block *sb = inode->i_sb;
fs/gfs2/super.c
1384
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
1391
if (inode->i_nlink || sb_rdonly(sb) || !ip->i_no_addr)
fs/gfs2/super.c
1452
static struct inode *gfs2_alloc_inode(struct super_block *sb)
fs/gfs2/super.c
1456
ip = alloc_inode_sb(sb, gfs2_inode_cachep, GFP_KERNEL);
fs/gfs2/super.c
266
int gfs2_statfs_sync(struct super_block *sb, int type)
fs/gfs2/super.c
268
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
577
static void gfs2_put_super(struct super_block *sb)
fs/gfs2/super.c
579
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
599
if (!sb_rdonly(sb))
fs/gfs2/super.c
656
static int gfs2_sync_fs(struct super_block *sb, int wait)
fs/gfs2/super.c
658
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
660
gfs2_quota_sync(sb, -1);
fs/gfs2/super.c
669
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/super.c
675
error = thaw_super(sb, who, freeze_owner);
fs/gfs2/super.c
688
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/super.c
696
error = freeze_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/gfs2/super.c
715
deactivate_super(sb);
fs/gfs2/super.c
726
static int gfs2_freeze_super(struct super_block *sb, enum freeze_holder who,
fs/gfs2/super.c
729
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
740
error = freeze_super(sb, who, freeze_owner);
fs/gfs2/super.c
774
static int gfs2_freeze_fs(struct super_block *sb)
fs/gfs2/super.c
776
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
795
static int gfs2_thaw_super(struct super_block *sb, enum freeze_holder who,
fs/gfs2/super.c
798
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
808
atomic_inc(&sb->s_active);
fs/gfs2/super.c
818
deactivate_super(sb);
fs/gfs2/super.c
960
struct super_block *sb = dentry->d_sb;
fs/gfs2/super.c
961
struct gfs2_sbd *sdp = sb->s_fs_info;
fs/gfs2/super.c
985
buf->f_fsid = uuid_to_fsid(sb->s_uuid.b);
fs/gfs2/super.h
48
int gfs2_statfs_sync(struct super_block *sb, int type);
fs/gfs2/sys.c
150
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/sys.c
151
int frozen = (sb->s_writers.frozen == SB_UNFROZEN) ? 0 : 1;
fs/gfs2/sys.c
714
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/sys.c
720
sprintf(ro, "RDONLY=%d", sb_rdonly(sb));
fs/gfs2/sys.c
739
&disk_to_dev(sb->s_bdev->bd_disk)->kobj,
fs/gfs2/trans.c
259
struct super_block *sb = sdp->sd_vfs;
fs/gfs2/trans.c
301
if (unlikely(sb->s_writers.frozen == SB_FREEZE_COMPLETE)) {
fs/hfs/bitmap.c
139
u32 hfs_vbm_search_free(struct super_block *sb, u32 goal, u32 *num_bits)
fs/hfs/bitmap.c
148
mutex_lock(&HFS_SB(sb)->bitmap_lock);
fs/hfs/bitmap.c
149
bitmap = HFS_SB(sb)->bitmap;
fs/hfs/bitmap.c
151
pos = hfs_find_set_zero_bits(bitmap, HFS_SB(sb)->fs_ablocks, goal, num_bits);
fs/hfs/bitmap.c
152
if (pos >= HFS_SB(sb)->fs_ablocks) {
fs/hfs/bitmap.c
155
if (pos >= HFS_SB(sb)->fs_ablocks) {
fs/hfs/bitmap.c
162
HFS_SB(sb)->free_ablocks -= *num_bits;
fs/hfs/bitmap.c
163
hfs_bitmap_dirty(sb);
fs/hfs/bitmap.c
165
mutex_unlock(&HFS_SB(sb)->bitmap_lock);
fs/hfs/bitmap.c
193
int hfs_clear_vbm_bits(struct super_block *sb, u16 start, u16 count)
fs/hfs/bitmap.c
205
if ((start + count) > HFS_SB(sb)->fs_ablocks)
fs/hfs/bitmap.c
208
mutex_lock(&HFS_SB(sb)->bitmap_lock);
fs/hfs/bitmap.c
210
curr = HFS_SB(sb)->bitmap + (start / 32);
fs/hfs/bitmap.c
238
HFS_SB(sb)->free_ablocks += len;
fs/hfs/bitmap.c
239
mutex_unlock(&HFS_SB(sb)->bitmap_lock);
fs/hfs/bitmap.c
240
hfs_bitmap_dirty(sb);
fs/hfs/btree.c
102
len = min_t(size_t, folio_size(folio), sb->s_blocksize);
fs/hfs/btree.c
103
memcpy_to_folio(folio, offset, bh->b_data, sb->s_blocksize);
fs/hfs/btree.c
19
struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id, btree_keycmp keycmp)
fs/hfs/btree.c
274
HFS_SB(tree->sb)->alloc_blksz;
fs/hfs/btree.c
276
tree->sb->s_blocksize_bits;
fs/hfs/btree.c
38
tree->sb = sb;
fs/hfs/btree.c
42
tree->inode = iget_locked(sb, id);
fs/hfs/btree.c
47
struct hfs_mdb *mdb = HFS_SB(sb)->mdb;
fs/hfs/btree.c
89
start_block = HFS_SB(sb)->fs_start + (dblock * HFS_SB(sb)->fs_div);
fs/hfs/btree.c
96
bh = sb_bread(sb, start_block);
fs/hfs/btree.h
25
struct super_block *sb;
fs/hfs/btree.h
89
extern struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id,
fs/hfs/catalog.c
108
hfs_cat_build_key(sb, fd.search_key, cnid, NULL);
fs/hfs/catalog.c
109
entry_size = hfs_cat_build_thread(sb, &entry, S_ISDIR(inode->i_mode) ?
fs/hfs/catalog.c
122
hfs_cat_build_key(sb, fd.search_key, dir->i_ino, str);
fs/hfs/catalog.c
142
hfs_cat_build_key(sb, fd.search_key, cnid, NULL);
fs/hfs/catalog.c
187
int hfs_cat_find_brec(struct super_block *sb, u32 cnid,
fs/hfs/catalog.c
193
hfs_cat_build_key(sb, fd->search_key, cnid, NULL);
fs/hfs/catalog.c
215
void hfs_set_next_unused_CNID(struct super_block *sb,
fs/hfs/catalog.c
219
atomic64_cmpxchg(&HFS_SB(sb)->next_id,
fs/hfs/catalog.c
222
atomic64_cmpxchg(&HFS_SB(sb)->next_id,
fs/hfs/catalog.c
23
void hfs_cat_build_key(struct super_block *sb, btree_key *key, u32 parent, const struct qstr *name)
fs/hfs/catalog.c
233
int hfs_correct_next_unused_CNID(struct super_block *sb, u32 cnid)
fs/hfs/catalog.c
242
cnid, atomic64_read(&HFS_SB(sb)->next_id));
fs/hfs/catalog.c
244
if ((cnid + 1) < atomic64_read(&HFS_SB(sb)->next_id)) {
fs/hfs/catalog.c
249
cat_tree = HFS_SB(sb)->cat_tree;
fs/hfs/catalog.c
28
hfs_asc2mac(sb, &key->cat.CName, name);
fs/hfs/catalog.c
313
hfs_set_next_unused_CNID(sb, cnid, found_cnid);
fs/hfs/catalog.c
319
hfs_set_next_unused_CNID(sb, cnid, found_cnid);
fs/hfs/catalog.c
341
struct super_block *sb;
fs/hfs/catalog.c
347
sb = dir->i_sb;
fs/hfs/catalog.c
348
res = hfs_find_init(HFS_SB(sb)->cat_tree, &fd);
fs/hfs/catalog.c
352
hfs_cat_build_key(sb, fd.search_key, dir->i_ino, str);
fs/hfs/catalog.c
363
hfs_free_fork(sb, &file, HFS_FK_DATA);
fs/hfs/catalog.c
365
hfs_free_fork(sb, &file, HFS_FK_RSRC);
fs/hfs/catalog.c
381
hfs_cat_build_key(sb, fd.search_key, cnid, NULL);
fs/hfs/catalog.c
393
res = hfs_correct_next_unused_CNID(sb, cnid);
fs/hfs/catalog.c
414
struct super_block *sb;
fs/hfs/catalog.c
423
sb = src_dir->i_sb;
fs/hfs/catalog.c
424
err = hfs_find_init(HFS_SB(sb)->cat_tree, &src_fd);
fs/hfs/catalog.c
438
hfs_cat_build_key(sb, src_fd.search_key, src_dir->i_ino, src_name);
fs/hfs/catalog.c
451
hfs_cat_build_key(sb, dst_fd.search_key, dst_dir->i_ino, dst_name);
fs/hfs/catalog.c
467
hfs_cat_build_key(sb, src_fd.search_key, src_dir->i_ino, src_name);
fs/hfs/catalog.c
483
hfs_cat_build_key(sb, src_fd.search_key, cnid, NULL);
fs/hfs/catalog.c
492
hfs_cat_build_key(sb, dst_fd.search_key, cnid, NULL);
fs/hfs/catalog.c
493
entry_size = hfs_cat_build_thread(sb, &entry, type == HFS_CDR_FIL ? HFS_CDR_FTH : HFS_CDR_THD,
fs/hfs/catalog.c
65
static int hfs_cat_build_thread(struct super_block *sb,
fs/hfs/catalog.c
72
hfs_asc2mac(sb, &rec->thread.CName, name);
fs/hfs/catalog.c
85
struct super_block *sb;
fs/hfs/catalog.c
95
sb = dir->i_sb;
fs/hfs/catalog.c
96
err = hfs_find_init(HFS_SB(sb)->cat_tree, &fd);
fs/hfs/dir.c
118
len = hfs_mac2asc(sb, strbuf, &fd.key->cat.CName);
fs/hfs/dir.c
257
struct super_block *sb = dir->i_sb;
fs/hfs/dir.c
264
if (unlikely(!is_hfs_cnid_counts_valid(sb))) {
fs/hfs/dir.c
51
struct super_block *sb = inode->i_sb;
fs/hfs/dir.c
62
err = hfs_find_init(HFS_SB(sb)->cat_tree, &fd);
fs/hfs/dir.c
65
hfs_cat_build_key(sb, fd.search_key, inode->i_ino, NULL);
fs/hfs/extent.c
248
static int hfs_free_extents(struct super_block *sb, struct hfs_extent *extent,
fs/hfs/extent.c
269
hfs_clear_vbm_bits(sb, start, count);
fs/hfs/extent.c
275
hfs_clear_vbm_bits(sb, start + count, block_nr);
fs/hfs/extent.c
287
int hfs_free_fork(struct super_block *sb, struct hfs_cat_file *file, int type)
fs/hfs/extent.c
302
total_blocks /= HFS_SB(sb)->alloc_blksz;
fs/hfs/extent.c
310
res = hfs_free_extents(sb, extent, blocks, blocks);
fs/hfs/extent.c
316
res = hfs_find_init(HFS_SB(sb)->ext_tree, &fd);
fs/hfs/extent.c
324
hfs_free_extents(sb, extent, total_blocks - start, total_blocks);
fs/hfs/extent.c
339
struct super_block *sb;
fs/hfs/extent.c
343
sb = inode->i_sb;
fs/hfs/extent.c
345
ablock = (u32)block / HFS_SB(sb)->fs_div;
fs/hfs/extent.c
377
map_bh(bh_result, sb, HFS_SB(sb)->fs_start +
fs/hfs/extent.c
378
dblock * HFS_SB(sb)->fs_div +
fs/hfs/extent.c
379
(u32)block % HFS_SB(sb)->fs_div);
fs/hfs/extent.c
383
HFS_I(inode)->phys_size += sb->s_blocksize;
fs/hfs/extent.c
385
inode_add_bytes(inode, sb->s_blocksize);
fs/hfs/extent.c
393
struct super_block *sb = inode->i_sb;
fs/hfs/extent.c
408
start = hfs_vbm_search_free(sb, goal, &len);
fs/hfs/extent.c
453
set_bit(HFS_FLG_ALT_MDB_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/extent.c
454
set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/extent.c
455
hfs_mark_mdb_dirty(sb);
fs/hfs/extent.c
479
struct super_block *sb = inode->i_sb;
fs/hfs/extent.c
506
size = inode->i_size + HFS_SB(sb)->alloc_blksz - 1;
fs/hfs/extent.c
507
blk_cnt = size / HFS_SB(sb)->alloc_blksz;
fs/hfs/extent.c
513
res = hfs_find_init(HFS_SB(sb)->ext_tree, &fd);
fs/hfs/extent.c
521
hfs_free_extents(sb, HFS_I(inode)->first_extents,
fs/hfs/extent.c
531
hfs_free_extents(sb, HFS_I(inode)->cached_extents,
fs/hfs/extent.c
549
HFS_I(inode)->fs_blocks = (inode->i_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/hfs/extent.c
550
inode_set_bytes(inode, HFS_I(inode)->fs_blocks << sb->s_blocksize_bits);
fs/hfs/hfs_fs.h
131
struct super_block *sb;
fs/hfs/hfs_fs.h
142
extern u32 hfs_vbm_search_free(struct super_block *sb, u32 goal, u32 *num_bits);
fs/hfs/hfs_fs.h
143
extern int hfs_clear_vbm_bits(struct super_block *sb, u16 start, u16 count);
fs/hfs/hfs_fs.h
148
extern int hfs_cat_find_brec(struct super_block *sb, u32 cnid,
fs/hfs/hfs_fs.h
157
extern void hfs_cat_build_key(struct super_block *sb, btree_key *key,
fs/hfs/hfs_fs.h
167
extern int hfs_free_fork(struct super_block *sb,
fs/hfs/hfs_fs.h
193
extern struct inode *hfs_iget(struct super_block *sb, struct hfs_cat_key *key,
fs/hfs/hfs_fs.h
202
extern bool is_hfs_cnid_counts_valid(struct super_block *sb);
fs/hfs/hfs_fs.h
203
extern int hfs_mdb_get(struct super_block *sb);
fs/hfs/hfs_fs.h
204
extern void hfs_mdb_commit(struct super_block *sb);
fs/hfs/hfs_fs.h
205
extern void hfs_mdb_close(struct super_block *sb);
fs/hfs/hfs_fs.h
206
extern void hfs_mdb_put(struct super_block *sb);
fs/hfs/hfs_fs.h
209
extern int hfs_part_find(struct super_block *sb,
fs/hfs/hfs_fs.h
223
extern void hfs_asc2mac(struct super_block *sb,
fs/hfs/hfs_fs.h
225
extern int hfs_mac2asc(struct super_block *sb,
fs/hfs/hfs_fs.h
229
extern void hfs_mark_mdb_dirty(struct super_block *sb);
fs/hfs/hfs_fs.h
258
#define HFS_SB(sb) ((struct hfs_sb_info *)(sb)->s_fs_info)
fs/hfs/hfs_fs.h
264
static inline const char *hfs_mdb_name(struct super_block *sb)
fs/hfs/hfs_fs.h
266
return sb->s_id;
fs/hfs/hfs_fs.h
269
static inline void hfs_bitmap_dirty(struct super_block *sb)
fs/hfs/hfs_fs.h
271
set_bit(HFS_FLG_BITMAP_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/hfs_fs.h
272
hfs_mark_mdb_dirty(sb);
fs/hfs/hfs_fs.h
275
#define sb_bread512(sb, sec, data) ({ \
fs/hfs/hfs_fs.h
282
__block = __start >> (sb)->s_blocksize_bits; \
fs/hfs/hfs_fs.h
283
__offset = __start & ((sb)->s_blocksize - 1); \
fs/hfs/hfs_fs.h
284
__bh = sb_bread((sb), __block); \
fs/hfs/inode.c
185
struct super_block *sb = dir->i_sb;
fs/hfs/inode.c
186
struct inode *inode = new_inode(sb);
fs/hfs/inode.c
200
hfs_cat_build_key(sb, (btree_key *)&HFS_I(inode)->cat_key, dir->i_ino, name);
fs/hfs/inode.c
201
next_id = atomic64_inc_return(&HFS_SB(sb)->next_id);
fs/hfs/inode.c
203
atomic64_dec(&HFS_SB(sb)->next_id);
fs/hfs/inode.c
219
folder_count = atomic64_inc_return(&HFS_SB(sb)->folder_count);
fs/hfs/inode.c
221
atomic64_dec(&HFS_SB(sb)->folder_count);
fs/hfs/inode.c
226
HFS_SB(sb)->root_dirs++;
fs/hfs/inode.c
232
HFS_I(inode)->clump_blocks = HFS_SB(sb)->clumpablks;
fs/hfs/inode.c
233
file_count = atomic64_inc_return(&HFS_SB(sb)->file_count);
fs/hfs/inode.c
235
atomic64_dec(&HFS_SB(sb)->file_count);
fs/hfs/inode.c
240
HFS_SB(sb)->root_files++;
fs/hfs/inode.c
258
set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/inode.c
259
hfs_mark_mdb_dirty(sb);
fs/hfs/inode.c
271
struct super_block *sb = inode->i_sb;
fs/hfs/inode.c
275
atomic64_dec(&HFS_SB(sb)->folder_count);
fs/hfs/inode.c
277
HFS_SB(sb)->root_dirs--;
fs/hfs/inode.c
278
set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/inode.c
279
hfs_mark_mdb_dirty(sb);
fs/hfs/inode.c
283
atomic64_dec(&HFS_SB(sb)->file_count);
fs/hfs/inode.c
285
HFS_SB(sb)->root_files--;
fs/hfs/inode.c
292
set_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags);
fs/hfs/inode.c
293
hfs_mark_mdb_dirty(sb);
fs/hfs/inode.c
299
struct super_block *sb = inode->i_sb;
fs/hfs/inode.c
312
HFS_I(inode)->fs_blocks = (log_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/hfs/inode.c
313
inode_set_bytes(inode, HFS_I(inode)->fs_blocks << sb->s_blocksize_bits);
fs/hfs/inode.c
315
HFS_SB(sb)->alloc_blksz;
fs/hfs/inode.c
316
HFS_I(inode)->clump_blocks = clump_size / HFS_SB(sb)->alloc_blksz;
fs/hfs/inode.c
318
HFS_I(inode)->clump_blocks = HFS_SB(sb)->clumpablks;
fs/hfs/inode.c
417
struct inode *hfs_iget(struct super_block *sb, struct hfs_cat_key *key, hfs_cat_rec *rec)
fs/hfs/inode.c
433
inode = iget5_locked(sb, cnid, hfs_test_inode, hfs_read_inode, &data);
fs/hfs/inode.c
697
struct super_block * sb;
fs/hfs/inode.c
70
struct super_block *sb = inode->i_sb;
fs/hfs/inode.c
709
sb = inode->i_sb;
fs/hfs/inode.c
710
flush_delayed_work(&HFS_SB(sb)->mdb_work);
fs/hfs/inode.c
712
err = sync_blockdev(sb->s_bdev);
fs/hfs/inode.c
79
tree = HFS_SB(sb)->ext_tree;
fs/hfs/inode.c
82
tree = HFS_SB(sb)->cat_tree;
fs/hfs/mdb.c
106
size = sb_min_blocksize(sb, HFS_SECTOR_SIZE);
fs/hfs/mdb.c
110
if (hfs_get_last_session(sb, &part_start, &part_size))
fs/hfs/mdb.c
114
bh = sb_bread512(sb, part_start + HFS_MDB_BLK, mdb);
fs/hfs/mdb.c
125
if (hfs_part_find(sb, &part_start, &part_size))
fs/hfs/mdb.c
129
HFS_SB(sb)->alloc_blksz = size = be32_to_cpu(mdb->drAlBlkSiz);
fs/hfs/mdb.c
136
size = min(HFS_SB(sb)->alloc_blksz, (u32)PAGE_SIZE);
fs/hfs/mdb.c
145
while (HFS_SB(sb)->alloc_blksz & (size - 1))
fs/hfs/mdb.c
148
if (!sb_set_blocksize(sb, size)) {
fs/hfs/mdb.c
153
bh = sb_bread512(sb, part_start + HFS_MDB_BLK, mdb);
fs/hfs/mdb.c
161
HFS_SB(sb)->mdb_bh = bh;
fs/hfs/mdb.c
162
HFS_SB(sb)->mdb = mdb;
fs/hfs/mdb.c
165
HFS_SB(sb)->part_start = part_start;
fs/hfs/mdb.c
166
HFS_SB(sb)->fs_ablocks = be16_to_cpu(mdb->drNmAlBlks);
fs/hfs/mdb.c
167
HFS_SB(sb)->fs_div = HFS_SB(sb)->alloc_blksz >> sb->s_blocksize_bits;
fs/hfs/mdb.c
168
HFS_SB(sb)->clumpablks = be32_to_cpu(mdb->drClpSiz) /
fs/hfs/mdb.c
169
HFS_SB(sb)->alloc_blksz;
fs/hfs/mdb.c
170
if (!HFS_SB(sb)->clumpablks)
fs/hfs/mdb.c
171
HFS_SB(sb)->clumpablks = 1;
fs/hfs/mdb.c
172
HFS_SB(sb)->fs_start = (be16_to_cpu(mdb->drAlBlSt) + part_start) >>
fs/hfs/mdb.c
173
(sb->s_blocksize_bits - HFS_SECTOR_SIZE_BITS);
fs/hfs/mdb.c
176
HFS_SB(sb)->free_ablocks = be16_to_cpu(mdb->drFreeBks);
fs/hfs/mdb.c
177
atomic64_set(&HFS_SB(sb)->next_id, be32_to_cpu(mdb->drNxtCNID));
fs/hfs/mdb.c
178
HFS_SB(sb)->root_files = be16_to_cpu(mdb->drNmFls);
fs/hfs/mdb.c
179
HFS_SB(sb)->root_dirs = be16_to_cpu(mdb->drNmRtDirs);
fs/hfs/mdb.c
180
atomic64_set(&HFS_SB(sb)->file_count, be32_to_cpu(mdb->drFilCnt));
fs/hfs/mdb.c
181
atomic64_set(&HFS_SB(sb)->folder_count, be32_to_cpu(mdb->drDirCnt));
fs/hfs/mdb.c
183
if (!is_hfs_cnid_counts_valid(sb)) {
fs/hfs/mdb.c
185
sb->s_flags |= SB_RDONLY;
fs/hfs/mdb.c
190
bh = sb_bread512(sb, sect, mdb2);
fs/hfs/mdb.c
193
HFS_SB(sb)->alt_mdb_bh = bh;
fs/hfs/mdb.c
194
HFS_SB(sb)->alt_mdb = mdb2;
fs/hfs/mdb.c
199
if (!HFS_SB(sb)->alt_mdb) {
fs/hfs/mdb.c
204
HFS_SB(sb)->bitmap = kzalloc(8192, GFP_KERNEL);
fs/hfs/mdb.c
205
if (!HFS_SB(sb)->bitmap)
fs/hfs/mdb.c
211
size = (HFS_SB(sb)->fs_ablocks + 8) / 8;
fs/hfs/mdb.c
212
ptr = (u8 *)HFS_SB(sb)->bitmap;
fs/hfs/mdb.c
214
bh = sb_bread(sb, off >> sb->s_blocksize_bits);
fs/hfs/mdb.c
219
off2 = off & (sb->s_blocksize - 1);
fs/hfs/mdb.c
220
len = min((int)sb->s_blocksize - off2, size);
fs/hfs/mdb.c
228
HFS_SB(sb)->ext_tree = hfs_btree_open(sb, HFS_EXT_CNID, hfs_ext_keycmp);
fs/hfs/mdb.c
229
if (!HFS_SB(sb)->ext_tree) {
fs/hfs/mdb.c
233
HFS_SB(sb)->cat_tree = hfs_btree_open(sb, HFS_CAT_CNID, hfs_cat_keycmp);
fs/hfs/mdb.c
234
if (!HFS_SB(sb)->cat_tree) {
fs/hfs/mdb.c
242
sb->s_flags |= SB_RDONLY;
fs/hfs/mdb.c
246
sb->s_flags |= SB_RDONLY;
fs/hfs/mdb.c
248
if (!sb_rdonly(sb)) {
fs/hfs/mdb.c
256
mark_buffer_dirty(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
257
sync_dirty_buffer(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
287
void hfs_mdb_commit(struct super_block *sb)
fs/hfs/mdb.c
289
struct hfs_mdb *mdb = HFS_SB(sb)->mdb;
fs/hfs/mdb.c
291
if (sb_rdonly(sb))
fs/hfs/mdb.c
294
lock_buffer(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
295
if (test_and_clear_bit(HFS_FLG_MDB_DIRTY, &HFS_SB(sb)->flags)) {
fs/hfs/mdb.c
298
mdb->drFreeBks = cpu_to_be16(HFS_SB(sb)->free_ablocks);
fs/hfs/mdb.c
300
cpu_to_be32((u32)atomic64_read(&HFS_SB(sb)->next_id));
fs/hfs/mdb.c
301
mdb->drNmFls = cpu_to_be16(HFS_SB(sb)->root_files);
fs/hfs/mdb.c
302
mdb->drNmRtDirs = cpu_to_be16(HFS_SB(sb)->root_dirs);
fs/hfs/mdb.c
304
cpu_to_be32((u32)atomic64_read(&HFS_SB(sb)->file_count));
fs/hfs/mdb.c
306
cpu_to_be32((u32)atomic64_read(&HFS_SB(sb)->folder_count));
fs/hfs/mdb.c
309
mark_buffer_dirty(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
315
if (test_and_clear_bit(HFS_FLG_ALT_MDB_DIRTY, &HFS_SB(sb)->flags) &&
fs/hfs/mdb.c
316
HFS_SB(sb)->alt_mdb) {
fs/hfs/mdb.c
317
hfs_inode_write_fork(HFS_SB(sb)->ext_tree->inode, mdb->drXTExtRec,
fs/hfs/mdb.c
319
hfs_inode_write_fork(HFS_SB(sb)->cat_tree->inode, mdb->drCTExtRec,
fs/hfs/mdb.c
32
static int hfs_get_last_session(struct super_block *sb,
fs/hfs/mdb.c
322
lock_buffer(HFS_SB(sb)->alt_mdb_bh);
fs/hfs/mdb.c
323
memcpy(HFS_SB(sb)->alt_mdb, HFS_SB(sb)->mdb, HFS_SECTOR_SIZE);
fs/hfs/mdb.c
324
HFS_SB(sb)->alt_mdb->drAtrb |= cpu_to_be16(HFS_SB_ATTRIB_UNMNT);
fs/hfs/mdb.c
325
HFS_SB(sb)->alt_mdb->drAtrb &= cpu_to_be16(~HFS_SB_ATTRIB_INCNSTNT);
fs/hfs/mdb.c
326
unlock_buffer(HFS_SB(sb)->alt_mdb_bh);
fs/hfs/mdb.c
328
mark_buffer_dirty(HFS_SB(sb)->alt_mdb_bh);
fs/hfs/mdb.c
329
sync_dirty_buffer(HFS_SB(sb)->alt_mdb_bh);
fs/hfs/mdb.c
332
if (test_and_clear_bit(HFS_FLG_BITMAP_DIRTY, &HFS_SB(sb)->flags)) {
fs/hfs/mdb.c
338
block = be16_to_cpu(HFS_SB(sb)->mdb->drVBMSt) + HFS_SB(sb)->part_start;
fs/hfs/mdb.c
339
off = (block << HFS_SECTOR_SIZE_BITS) & (sb->s_blocksize - 1);
fs/hfs/mdb.c
340
block >>= sb->s_blocksize_bits - HFS_SECTOR_SIZE_BITS;
fs/hfs/mdb.c
341
size = (HFS_SB(sb)->fs_ablocks + 7) / 8;
fs/hfs/mdb.c
342
ptr = (u8 *)HFS_SB(sb)->bitmap;
fs/hfs/mdb.c
344
bh = sb_bread(sb, block);
fs/hfs/mdb.c
349
len = min((int)sb->s_blocksize - off, size);
fs/hfs/mdb.c
35
struct cdrom_device_info *cdi = disk_to_cdi(sb->s_bdev->bd_disk);
fs/hfs/mdb.c
363
unlock_buffer(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
366
void hfs_mdb_close(struct super_block *sb)
fs/hfs/mdb.c
369
if (sb_rdonly(sb))
fs/hfs/mdb.c
371
HFS_SB(sb)->mdb->drAtrb |= cpu_to_be16(HFS_SB_ATTRIB_UNMNT);
fs/hfs/mdb.c
372
HFS_SB(sb)->mdb->drAtrb &= cpu_to_be16(~HFS_SB_ATTRIB_INCNSTNT);
fs/hfs/mdb.c
373
mark_buffer_dirty(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
380
void hfs_mdb_put(struct super_block *sb)
fs/hfs/mdb.c
383
hfs_btree_close(HFS_SB(sb)->ext_tree);
fs/hfs/mdb.c
384
hfs_btree_close(HFS_SB(sb)->cat_tree);
fs/hfs/mdb.c
387
brelse(HFS_SB(sb)->mdb_bh);
fs/hfs/mdb.c
388
brelse(HFS_SB(sb)->alt_mdb_bh);
fs/hfs/mdb.c
39
*size = bdev_nr_sectors(sb->s_bdev);
fs/hfs/mdb.c
390
unload_nls(HFS_SB(sb)->nls_io);
fs/hfs/mdb.c
391
unload_nls(HFS_SB(sb)->nls_disk);
fs/hfs/mdb.c
393
kfree(HFS_SB(sb)->bitmap);
fs/hfs/mdb.c
41
if (HFS_SB(sb)->session >= 0) {
fs/hfs/mdb.c
47
te.cdte_track = HFS_SB(sb)->session;
fs/hfs/mdb.c
67
bool is_hfs_cnid_counts_valid(struct super_block *sb)
fs/hfs/mdb.c
69
struct hfs_sb_info *sbi = HFS_SB(sb);
fs/hfs/mdb.c
94
int hfs_mdb_get(struct super_block *sb)
fs/hfs/part_tbl.c
105
bh = sb_bread512(sb, *part_start + HFS_PMAP_BLK + ++i, pm);
fs/hfs/part_tbl.c
58
int hfs_part_find(struct super_block *sb,
fs/hfs/part_tbl.c
66
bh = sb_bread512(sb, *part_start + HFS_PMAP_BLK, data);
fs/hfs/part_tbl.c
82
(HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
fs/hfs/part_tbl.c
98
(HFS_SB(sb)->part < 0 || HFS_SB(sb)->part == i)) {
fs/hfs/super.c
102
struct super_block *sb = dentry->d_sb;
fs/hfs/super.c
103
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/hfs/super.c
106
buf->f_bsize = sb->s_blocksize;
fs/hfs/super.c
107
buf->f_blocks = (u32)HFS_SB(sb)->fs_ablocks * HFS_SB(sb)->fs_div;
fs/hfs/super.c
108
buf->f_bfree = (u32)HFS_SB(sb)->free_ablocks * HFS_SB(sb)->fs_div;
fs/hfs/super.c
110
buf->f_files = HFS_SB(sb)->fs_ablocks;
fs/hfs/super.c
111
buf->f_ffree = HFS_SB(sb)->free_ablocks;
fs/hfs/super.c
120
struct super_block *sb = fc->root->d_sb;
fs/hfs/super.c
122
sync_filesystem(sb);
fs/hfs/super.c
124
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb))
fs/hfs/super.c
128
if (!(HFS_SB(sb)->mdb->drAtrb & cpu_to_be16(HFS_SB_ATTRIB_UNMNT))) {
fs/hfs/super.c
130
sb->s_flags |= SB_RDONLY;
fs/hfs/super.c
132
} else if (HFS_SB(sb)->mdb->drAtrb & cpu_to_be16(HFS_SB_ATTRIB_SLOCK)) {
fs/hfs/super.c
134
sb->s_flags |= SB_RDONLY;
fs/hfs/super.c
169
static struct inode *hfs_alloc_inode(struct super_block *sb)
fs/hfs/super.c
173
i = alloc_inode_sb(sb, hfs_inode_cachep, GFP_KERNEL);
fs/hfs/super.c
316
static int hfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/hfs/super.c
318
struct hfs_sb_info *sbi = HFS_SB(sb);
fs/hfs/super.c
338
sbi->sb = sb;
fs/hfs/super.c
339
sb->s_op = &hfs_super_operations;
fs/hfs/super.c
340
sb->s_xattr = hfs_xattr_handlers;
fs/hfs/super.c
341
sb->s_flags |= SB_NODIRATIME;
fs/hfs/super.c
344
res = hfs_mdb_get(sb);
fs/hfs/super.c
348
hfs_mdb_name(sb));
fs/hfs/super.c
35
static int hfs_sync_fs(struct super_block *sb, int wait)
fs/hfs/super.c
354
res = hfs_find_init(HFS_SB(sb)->cat_tree, &fd);
fs/hfs/super.c
357
res = hfs_cat_find_brec(sb, HFS_ROOT_CNID, &fd);
fs/hfs/super.c
37
is_hfs_cnid_counts_valid(sb);
fs/hfs/super.c
370
root_inode = hfs_iget(sb, &fd.search_key->cat, &rec);
fs/hfs/super.c
375
set_default_d_op(sb, &hfs_dentry_operations);
fs/hfs/super.c
377
sb->s_root = d_make_root(root_inode);
fs/hfs/super.c
378
if (!sb->s_root)
fs/hfs/super.c
38
hfs_mdb_commit(sb);
fs/hfs/super.c
389
hfs_mdb_put(sb);
fs/hfs/super.c
437
static void hfs_kill_super(struct super_block *sb)
fs/hfs/super.c
439
struct hfs_sb_info *hsb = HFS_SB(sb);
fs/hfs/super.c
441
kill_block_super(sb);
fs/hfs/super.c
49
static void hfs_put_super(struct super_block *sb)
fs/hfs/super.c
51
cancel_delayed_work_sync(&HFS_SB(sb)->mdb_work);
fs/hfs/super.c
52
hfs_mdb_close(sb);
fs/hfs/super.c
54
hfs_mdb_put(sb);
fs/hfs/super.c
60
struct super_block *sb;
fs/hfs/super.c
63
sb = sbi->sb;
fs/hfs/super.c
69
is_hfs_cnid_counts_valid(sb);
fs/hfs/super.c
71
hfs_mdb_commit(sb);
fs/hfs/super.c
74
void hfs_mark_mdb_dirty(struct super_block *sb)
fs/hfs/super.c
76
struct hfs_sb_info *sbi = HFS_SB(sb);
fs/hfs/super.c
79
if (sb_rdonly(sb))
fs/hfs/trans.c
100
struct nls_table *nls_io = HFS_SB(sb)->nls_io;
fs/hfs/trans.c
33
int hfs_mac2asc(struct super_block *sb, char *out, const struct hfs_name *in)
fs/hfs/trans.c
35
struct nls_table *nls_disk = HFS_SB(sb)->nls_disk;
fs/hfs/trans.c
36
struct nls_table *nls_io = HFS_SB(sb)->nls_io;
fs/hfs/trans.c
97
void hfs_asc2mac(struct super_block *sb, struct hfs_name *out, const struct qstr *in)
fs/hfs/trans.c
99
struct nls_table *nls_disk = HFS_SB(sb)->nls_disk;
fs/hfsplus/attributes.c
139
int hfsplus_find_attr(struct super_block *sb, u32 cnid,
fs/hfsplus/attributes.c
146
if (!HFSPLUS_SB(sb)->attr_tree) {
fs/hfsplus/attributes.c
152
err = hfsplus_attr_build_key(sb, fd->search_key, cnid, name);
fs/hfsplus/attributes.c
159
err = hfsplus_attr_build_key(sb, fd->search_key, cnid, NULL);
fs/hfsplus/attributes.c
174
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
177
if (!HFSPLUS_SB(sb)->attr_tree)
fs/hfsplus/attributes.c
180
err = hfs_find_init(HFSPLUS_SB(sb)->attr_tree, &fd);
fs/hfsplus/attributes.c
184
err = hfsplus_find_attr(sb, inode->i_ino, name, &fd);
fs/hfsplus/attributes.c
202
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
210
err = hfsplus_attr_build_key(sb, fd->search_key,
fs/hfsplus/attributes.c
253
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
261
if (!HFSPLUS_SB(sb)->attr_tree) {
fs/hfsplus/attributes.c
270
err = hfs_find_init(HFSPLUS_SB(sb)->attr_tree, &fd);
fs/hfsplus/attributes.c
337
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
344
err = hfsplus_attr_build_key(sb, fd->search_key,
fs/hfsplus/attributes.c
367
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
373
if (!HFSPLUS_SB(sb)->attr_tree) {
fs/hfsplus/attributes.c
378
err = hfs_find_init(HFSPLUS_SB(sb)->attr_tree, &fd);
fs/hfsplus/attributes.c
434
struct super_block *sb = inode->i_sb;
fs/hfsplus/attributes.c
442
if (!HFSPLUS_SB(sb)->attr_tree) {
fs/hfsplus/attributes.c
451
err = hfs_find_init(HFSPLUS_SB(sb)->attr_tree, &fd);
fs/hfsplus/attributes.c
50
int hfsplus_attr_build_key(struct super_block *sb, hfsplus_btree_key *key,
fs/hfsplus/attributes.c
58
int res = hfsplus_asc2uni(sb,
fs/hfsplus/bitmap.c
157
hfsplus_mark_mdb_dirty(sb);
fs/hfsplus/bitmap.c
164
int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
fs/hfsplus/bitmap.c
166
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/bitmap.c
19
int hfsplus_block_allocate(struct super_block *sb, u32 size,
fs/hfsplus/bitmap.c
22
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/bitmap.c
236
hfsplus_mark_mdb_dirty(sb);
fs/hfsplus/bnode.c
704
struct super_block *sb = tree->inode->i_sb;
fs/hfsplus/bnode.c
705
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/btree.c
133
struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
fs/hfsplus/btree.c
148
tree->sb = sb;
fs/hfsplus/btree.c
150
inode = hfsplus_iget(sb, id);
fs/hfsplus/btree.c
205
if (test_bit(HFSPLUS_SB_HFSX, &HFSPLUS_SB(sb)->flags) &&
fs/hfsplus/btree.c
210
set_bit(HFSPLUS_SB_CASEFOLD, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/btree.c
363
HFSPLUS_SB(tree->sb)->alloc_blksz_shift;
fs/hfsplus/btree.c
365
hip->alloc_blocks << HFSPLUS_SB(tree->sb)->fs_shift;
fs/hfsplus/catalog.c
176
static int hfsplus_fill_cat_thread(struct super_block *sb,
fs/hfsplus/catalog.c
185
err = hfsplus_asc2uni(sb, &entry->thread.nodeName, HFSPLUS_MAX_STRLEN,
fs/hfsplus/catalog.c
194
int hfsplus_find_cat(struct super_block *sb, u32 cnid,
fs/hfsplus/catalog.c
201
hfsplus_cat_build_key_with_cnid(sb, fd->search_key, cnid);
fs/hfsplus/catalog.c
256
struct super_block *sb = dir->i_sb;
fs/hfsplus/catalog.c
264
err = hfs_find_init(HFSPLUS_SB(sb)->cat_tree, &fd);
fs/hfsplus/catalog.c
276
hfsplus_cat_build_key_with_cnid(sb, fd.search_key, cnid);
fs/hfsplus/catalog.c
277
entry_size = hfsplus_fill_cat_thread(sb, &entry,
fs/hfsplus/catalog.c
296
err = hfsplus_cat_build_key(sb, fd.search_key, dir->i_ino, str);
fs/hfsplus/catalog.c
322
hfsplus_cat_build_key_with_cnid(sb, fd.search_key, cnid);
fs/hfsplus/catalog.c
332
struct super_block *sb = dir->i_sb;
fs/hfsplus/catalog.c
340
err = hfs_find_init(HFSPLUS_SB(sb)->cat_tree, &fd);
fs/hfsplus/catalog.c
355
hfsplus_cat_build_key_with_cnid(sb, fd.search_key, cnid);
fs/hfsplus/catalog.c
371
err = hfsplus_cat_build_key(sb, fd.search_key, dir->i_ino, str);
fs/hfsplus/catalog.c
385
hfsplus_free_fork(sb, cnid, &fork, HFSPLUS_TYPE_DATA);
fs/hfsplus/catalog.c
391
hfsplus_free_fork(sb, cnid, &fork, HFSPLUS_TYPE_RSRC);
fs/hfsplus/catalog.c
408
hfsplus_cat_build_key_with_cnid(sb, fd.search_key, cnid);
fs/hfsplus/catalog.c
424
if (HFSPLUS_SB(sb)->attr_tree)
fs/hfsplus/catalog.c
43
int hfsplus_cat_build_key(struct super_block *sb,
fs/hfsplus/catalog.c
438
struct super_block *sb = src_dir->i_sb;
fs/hfsplus/catalog.c
447
err = hfs_find_init(HFSPLUS_SB(sb)->cat_tree, &src_fd);
fs/hfsplus/catalog.c
461
err = hfsplus_cat_build_key(sb, src_fd.search_key,
fs/hfsplus/catalog.c
479
err = hfsplus_cat_build_key(sb, dst_fd.search_key,
fs/hfsplus/catalog.c
49
err = hfsplus_asc2uni(sb, &key->cat.name, HFSPLUS_MAX_STRLEN,
fs/hfsplus/catalog.c
500
err = hfsplus_cat_build_key(sb, src_fd.search_key,
fs/hfsplus/catalog.c
517
hfsplus_cat_build_key_with_cnid(sb, src_fd.search_key, cnid);
fs/hfsplus/catalog.c
527
hfsplus_cat_build_key_with_cnid(sb, dst_fd.search_key, cnid);
fs/hfsplus/catalog.c
528
entry_size = hfsplus_fill_cat_thread(sb, &entry, type,
fs/hfsplus/catalog.c
60
void hfsplus_cat_build_key_with_cnid(struct super_block *sb,
fs/hfsplus/dir.c
103
err = hfsplus_cat_build_key(sb, fd.search_key,
fs/hfsplus/dir.c
104
HFSPLUS_SB(sb)->hidden_dir->i_ino,
fs/hfsplus/dir.c
133
struct super_block *sb = inode->i_sb;
fs/hfsplus/dir.c
144
err = hfs_find_init(HFSPLUS_SB(sb)->cat_tree, &fd);
fs/hfsplus/dir.c
152
hfsplus_cat_build_key_with_cnid(sb, fd.search_key, inode->i_ino);
fs/hfsplus/dir.c
207
err = hfsplus_uni2asc_str(sb, &fd.key->cat.name, strbuf, &len);
fs/hfsplus/dir.c
217
if (HFSPLUS_SB(sb)->hidden_dir &&
fs/hfsplus/dir.c
218
HFSPLUS_SB(sb)->hidden_dir->i_ino ==
fs/hfsplus/dir.c
35
struct super_block *sb;
fs/hfsplus/dir.c
41
sb = dir->i_sb;
fs/hfsplus/dir.c
44
err = hfs_find_init(HFSPLUS_SB(sb)->cat_tree, &fd);
fs/hfsplus/dir.c
47
err = hfsplus_cat_build_key(sb, fd.search_key, dir->i_ino,
fs/hfsplus/dir.c
80
HFSPLUS_SB(sb)->hidden_dir &&
fs/hfsplus/dir.c
82
HFSPLUS_I(HFSPLUS_SB(sb)->hidden_dir)->
fs/hfsplus/dir.c
85
HFSPLUS_I(d_inode(sb->s_root))->
fs/hfsplus/extents.c
228
struct super_block *sb = inode->i_sb;
fs/hfsplus/extents.c
229
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/extents.c
284
map_bh(bh_result, sb, sector);
fs/hfsplus/extents.c
288
hip->phys_size += sb->s_blocksize;
fs/hfsplus/extents.c
290
inode_add_bytes(inode, sb->s_blocksize);
fs/hfsplus/extents.c
337
static int hfsplus_free_extents(struct super_block *sb,
fs/hfsplus/extents.c
360
err = hfsplus_block_free(sb, start, count);
fs/hfsplus/extents.c
370
err = hfsplus_block_free(sb, start + count, block_nr);
fs/hfsplus/extents.c
391
int hfsplus_free_fork(struct super_block *sb, u32 cnid,
fs/hfsplus/extents.c
407
res = hfsplus_free_extents(sb, fork->extents, blocks, blocks);
fs/hfsplus/extents.c
413
res = hfs_find_init(HFSPLUS_SB(sb)->ext_tree, &fd);
fs/hfsplus/extents.c
425
hfsplus_free_extents(sb, ext_entry, total_blocks - start,
fs/hfsplus/extents.c
438
struct super_block *sb = inode->i_sb;
fs/hfsplus/extents.c
439
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/extents.c
464
start = hfsplus_block_allocate(sb, sbi->total_blocks, goal, &len);
fs/hfsplus/extents.c
466
start = hfsplus_block_allocate(sb, goal, 0, &len);
fs/hfsplus/extents.c
474
res = sb_issue_zeroout(sb, start, len, GFP_NOFS);
fs/hfsplus/extents.c
542
struct super_block *sb = inode->i_sb;
fs/hfsplus/extents.c
570
blk_cnt = (inode->i_size + HFSPLUS_SB(sb)->alloc_blksz - 1) >>
fs/hfsplus/extents.c
571
HFSPLUS_SB(sb)->alloc_blksz_shift;
fs/hfsplus/extents.c
579
res = hfs_find_init(HFSPLUS_SB(sb)->ext_tree, &fd);
fs/hfsplus/extents.c
588
hfsplus_free_extents(sb, hip->first_extents,
fs/hfsplus/extents.c
604
hfsplus_free_extents(sb, hip->cached_extents,
fs/hfsplus/extents.c
623
hip->fs_blocks = (inode->i_size + sb->s_blocksize - 1) >>
fs/hfsplus/extents.c
624
sb->s_blocksize_bits;
fs/hfsplus/extents.c
625
inode_set_bytes(inode, hip->fs_blocks << sb->s_blocksize_bits);
fs/hfsplus/hfsplus_fs.h
173
static inline struct hfsplus_sb_info *HFSPLUS_SB(struct super_block *sb)
fs/hfsplus/hfsplus_fs.h
175
return sb->s_fs_info;
fs/hfsplus/hfsplus_fs.h
276
static inline unsigned short hfsplus_min_io_size(struct super_block *sb)
fs/hfsplus/hfsplus_fs.h
278
return max_t(unsigned short, HFSPLUS_SB(sb)->min_io_size,
fs/hfsplus/hfsplus_fs.h
336
int hfsplus_attr_build_key(struct super_block *sb, hfsplus_btree_key *key,
fs/hfsplus/hfsplus_fs.h
340
int hfsplus_find_attr(struct super_block *sb, u32 cnid, const char *name,
fs/hfsplus/hfsplus_fs.h
352
int hfsplus_block_allocate(struct super_block *sb, u32 size, u32 offset,
fs/hfsplus/hfsplus_fs.h
354
int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count);
fs/hfsplus/hfsplus_fs.h
359
struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id);
fs/hfsplus/hfsplus_fs.h
41
struct super_block *sb;
fs/hfsplus/hfsplus_fs.h
412
int hfsplus_cat_build_key(struct super_block *sb, hfsplus_btree_key *key,
fs/hfsplus/hfsplus_fs.h
414
void hfsplus_cat_build_key_with_cnid(struct super_block *sb,
fs/hfsplus/hfsplus_fs.h
417
int hfsplus_find_cat(struct super_block *sb, u32 cnid,
fs/hfsplus/hfsplus_fs.h
435
int hfsplus_free_fork(struct super_block *sb, u32 cnid,
fs/hfsplus/hfsplus_fs.h
449
struct inode *hfsplus_new_inode(struct super_block *sb, struct inode *dir,
fs/hfsplus/hfsplus_fs.h
476
int hfs_part_find(struct super_block *sb, sector_t *part_start,
fs/hfsplus/hfsplus_fs.h
480
struct inode *hfsplus_iget(struct super_block *sb, unsigned long ino);
fs/hfsplus/hfsplus_fs.h
481
void hfsplus_mark_mdb_dirty(struct super_block *sb);
fs/hfsplus/hfsplus_fs.h
483
int hfsplus_commit_superblock(struct super_block *sb);
fs/hfsplus/hfsplus_fs.h
495
int hfsplus_uni2asc_str(struct super_block *sb,
fs/hfsplus/hfsplus_fs.h
498
int hfsplus_uni2asc_xattr_str(struct super_block *sb,
fs/hfsplus/hfsplus_fs.h
501
int hfsplus_asc2uni(struct super_block *sb, struct hfsplus_unistr *ustr,
fs/hfsplus/hfsplus_fs.h
508
int hfsplus_submit_bio(struct super_block *sb, sector_t sector, void *buf,
fs/hfsplus/hfsplus_fs.h
510
int hfsplus_read_wrapper(struct super_block *sb);
fs/hfsplus/inode.c
249
struct super_block *sb = inode->i_sb;
fs/hfsplus/inode.c
258
HFSPLUS_SB(sb)->hidden_dir, NULL);
fs/hfsplus/inode.c
425
struct inode *hfsplus_new_inode(struct super_block *sb, struct inode *dir,
fs/hfsplus/inode.c
428
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/inode.c
429
struct inode *inode = new_inode(sb);
fs/hfsplus/inode.c
484
hfsplus_mark_mdb_dirty(sb);
fs/hfsplus/inode.c
491
struct super_block *sb = inode->i_sb;
fs/hfsplus/inode.c
494
HFSPLUS_SB(sb)->folder_count--;
fs/hfsplus/inode.c
495
hfsplus_mark_mdb_dirty(sb);
fs/hfsplus/inode.c
498
HFSPLUS_SB(sb)->file_count--;
fs/hfsplus/inode.c
508
hfsplus_mark_mdb_dirty(sb);
fs/hfsplus/inode.c
513
struct super_block *sb = inode->i_sb;
fs/hfsplus/inode.c
514
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/inode.c
530
(inode->i_size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/hfsplus/inode.c
531
inode_set_bytes(inode, hip->fs_blocks << sb->s_blocksize_bits);
fs/hfsplus/inode.c
65
struct super_block *sb = inode->i_sb;
fs/hfsplus/inode.c
74
tree = HFSPLUS_SB(sb)->ext_tree;
fs/hfsplus/inode.c
77
tree = HFSPLUS_SB(sb)->cat_tree;
fs/hfsplus/inode.c
80
tree = HFSPLUS_SB(sb)->attr_tree;
fs/hfsplus/part_tbl.c
113
res = hfsplus_submit_bio(sb,
fs/hfsplus/part_tbl.c
128
int hfs_part_find(struct super_block *sb,
fs/hfsplus/part_tbl.c
134
buf = kmalloc(hfsplus_min_io_size(sb), GFP_KERNEL);
fs/hfsplus/part_tbl.c
138
res = hfsplus_submit_bio(sb, *part_start + HFS_PMAP_BLK,
fs/hfsplus/part_tbl.c
145
res = hfs_parse_old_pmap(sb, data, part_start, part_size);
fs/hfsplus/part_tbl.c
148
res = hfs_parse_new_pmap(sb, buf, data, part_start, part_size);
fs/hfsplus/part_tbl.c
70
static int hfs_parse_old_pmap(struct super_block *sb, struct old_pmap *pm,
fs/hfsplus/part_tbl.c
73
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/part_tbl.c
91
static int hfs_parse_new_pmap(struct super_block *sb, void *buf,
fs/hfsplus/part_tbl.c
94
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/part_tbl.c
96
int buf_size = hfsplus_min_io_size(sb);
fs/hfsplus/super.c
196
int hfsplus_commit_superblock(struct super_block *sb)
fs/hfsplus/super.c
198
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
22
static struct inode *hfsplus_alloc_inode(struct super_block *sb);
fs/hfsplus/super.c
221
error2 = hfsplus_submit_bio(sb,
fs/hfsplus/super.c
229
error2 = hfsplus_submit_bio(sb,
fs/hfsplus/super.c
243
static int hfsplus_sync_fs(struct super_block *sb, int wait)
fs/hfsplus/super.c
245
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
275
error2 = hfsplus_commit_superblock(sb);
fs/hfsplus/super.c
280
blkdev_issue_flush(sb->s_bdev);
fs/hfsplus/super.c
303
void hfsplus_mark_mdb_dirty(struct super_block *sb)
fs/hfsplus/super.c
305
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
308
if (sb_rdonly(sb))
fs/hfsplus/super.c
328
static void hfsplus_put_super(struct super_block *sb)
fs/hfsplus/super.c
330
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
336
if (!sb_rdonly(sb) && sbi->s_vhdr) {
fs/hfsplus/super.c
343
hfsplus_sync_fs(sb, 1);
fs/hfsplus/super.c
358
struct super_block *sb = dentry->d_sb;
fs/hfsplus/super.c
359
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
360
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/hfsplus/super.c
363
buf->f_bsize = sb->s_blocksize;
fs/hfsplus/super.c
377
struct super_block *sb = fc->root->d_sb;
fs/hfsplus/super.c
379
sync_filesystem(sb);
fs/hfsplus/super.c
380
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb))
fs/hfsplus/super.c
383
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
388
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
395
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
400
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
427
static int hfsplus_fill_super(struct super_block *sb, struct fs_context *fc)
fs/hfsplus/super.c
430
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
462
if (hfsplus_read_wrapper(sb)) {
fs/hfsplus/super.c
470
sb->s_magic = HFSPLUS_VOLHEAD_SIG;
fs/hfsplus/super.c
502
sb->s_op = &hfsplus_sops;
fs/hfsplus/super.c
503
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/hfsplus/super.c
507
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
512
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
514
!sb_rdonly(sb)) {
fs/hfsplus/super.c
516
sb->s_flags |= SB_RDONLY;
fs/hfsplus/super.c
522
sbi->ext_tree = hfs_btree_open(sb, HFSPLUS_EXT_CNID);
fs/hfsplus/super.c
527
sbi->cat_tree = hfs_btree_open(sb, HFSPLUS_CAT_CNID);
fs/hfsplus/super.c
534
sbi->attr_tree = hfs_btree_open(sb, HFSPLUS_ATTR_CNID);
fs/hfsplus/super.c
541
sb->s_xattr = hfsplus_xattr_handlers;
fs/hfsplus/super.c
543
inode = hfsplus_iget(sb, HFSPLUS_ALLOC_CNID);
fs/hfsplus/super.c
552
root = hfsplus_iget(sb, HFSPLUS_ROOT_CNID);
fs/hfsplus/super.c
559
set_default_d_op(sb, &hfsplus_dentry_operations);
fs/hfsplus/super.c
560
sb->s_root = d_make_root(root);
fs/hfsplus/super.c
561
if (!sb->s_root) {
fs/hfsplus/super.c
571
err = hfsplus_cat_build_key(sb, fd.search_key, HFSPLUS_ROOT_CNID, &str);
fs/hfsplus/super.c
580
inode = hfsplus_iget(sb, be32_to_cpu(entry.folder.id));
fs/hfsplus/super.c
589
if (!sb_rdonly(sb)) {
fs/hfsplus/super.c
595
hfsplus_sync_fs(sb, 1);
fs/hfsplus/super.c
599
sbi->hidden_dir = hfsplus_new_inode(sb, root, S_IFDIR);
fs/hfsplus/super.c
641
dput(sb->s_root);
fs/hfsplus/super.c
642
sb->s_root = NULL;
fs/hfsplus/super.c
65
struct inode *hfsplus_iget(struct super_block *sb, unsigned long ino)
fs/hfsplus/super.c
665
static struct inode *hfsplus_alloc_inode(struct super_block *sb)
fs/hfsplus/super.c
669
i = alloc_inode_sb(sb, hfsplus_inode_cachep, GFP_KERNEL);
fs/hfsplus/super.c
71
inode = iget_locked(sb, ino);
fs/hfsplus/super.c
714
static void hfsplus_kill_super(struct super_block *sb)
fs/hfsplus/super.c
716
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/super.c
718
kill_block_super(sb);
fs/hfsplus/unicode.c
150
static int hfsplus_uni2asc(struct super_block *sb,
fs/hfsplus/unicode.c
155
struct nls_table *nls = HFSPLUS_SB(sb)->nls;
fs/hfsplus/unicode.c
173
compose = !test_bit(HFSPLUS_SB_NODECOMPOSE, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
287
inline int hfsplus_uni2asc_str(struct super_block *sb,
fs/hfsplus/unicode.c
291
return hfsplus_uni2asc(sb, ustr, HFSPLUS_MAX_STRLEN, astr, len_p);
fs/hfsplus/unicode.c
295
inline int hfsplus_uni2asc_xattr_str(struct super_block *sb,
fs/hfsplus/unicode.c
299
return hfsplus_uni2asc(sb, (const struct hfsplus_unistr *)ustr,
fs/hfsplus/unicode.c
308
static inline int asc2unichar(struct super_block *sb, const char *astr, int len,
fs/hfsplus/unicode.c
311
int size = HFSPLUS_SB(sb)->nls->char2uni(astr, len, uc);
fs/hfsplus/unicode.c
396
int hfsplus_asc2uni(struct super_block *sb,
fs/hfsplus/unicode.c
405
decompose = !test_bit(HFSPLUS_SB_NODECOMPOSE, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
407
size = asc2unichar(sb, astr, len, &c);
fs/hfsplus/unicode.c
439
struct super_block *sb = dentry->d_sb;
fs/hfsplus/unicode.c
448
casefold = test_bit(HFSPLUS_SB_CASEFOLD, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
449
decompose = !test_bit(HFSPLUS_SB_NODECOMPOSE, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
455
size = asc2unichar(sb, astr, len, &c);
fs/hfsplus/unicode.c
493
struct super_block *sb = dentry->d_sb;
fs/hfsplus/unicode.c
502
casefold = test_bit(HFSPLUS_SB_CASEFOLD, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
503
decompose = !test_bit(HFSPLUS_SB_NODECOMPOSE, &HFSPLUS_SB(sb)->flags);
fs/hfsplus/unicode.c
513
size = asc2unichar(sb, astr1, len1, &c);
fs/hfsplus/unicode.c
528
size = asc2unichar(sb, astr2, len2, &c);
fs/hfsplus/unicode_test.c
1005
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1058
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1097
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1134
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1184
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1227
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1274
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1325
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1363
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1404
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1447
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
1507
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/unicode_test.c
389
struct super_block sb;
fs/hfsplus/unicode_test.c
403
ptr->sb.s_fs_info = &ptr->sb_info;
fs/hfsplus/unicode_test.c
451
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
461
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
470
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
500
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
512
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
525
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
556
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
565
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
574
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
604
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
641
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
657
result = hfsplus_uni2asc_str(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
717
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
724
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
731
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
738
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
764
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
772
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
782
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
812
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
822
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
831
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
838
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1, 0, "test", 4);
fs/hfsplus/unicode_test.c
861
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
868
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
875
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
882
result = hfsplus_asc2uni(&mock_sb->sb,
fs/hfsplus/unicode_test.c
911
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str1,
fs/hfsplus/unicode_test.c
919
result = hfsplus_asc2uni(&mock_sb->sb, &mock_env->str2,
fs/hfsplus/unicode_test.c
937
static void setup_mock_dentry(struct super_block *sb)
fs/hfsplus/unicode_test.c
940
test_dentry.d_sb = sb;
fs/hfsplus/unicode_test.c
961
setup_mock_dentry(&mock_sb->sb);
fs/hfsplus/wrapper.c
101
struct cdrom_device_info *cdi = disk_to_cdi(sb->s_bdev->bd_disk);
fs/hfsplus/wrapper.c
105
*size = bdev_nr_sectors(sb->s_bdev);
fs/hfsplus/wrapper.c
107
if (HFSPLUS_SB(sb)->session >= 0) {
fs/hfsplus/wrapper.c
113
te.cdte_track = HFSPLUS_SB(sb)->session;
fs/hfsplus/wrapper.c
134
int hfsplus_read_wrapper(struct super_block *sb)
fs/hfsplus/wrapper.c
136
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/wrapper.c
143
blocksize = sb_min_blocksize(sb, HFSPLUS_SECTOR_SIZE);
fs/hfsplus/wrapper.c
149
if (hfsplus_get_last_session(sb, &part_start, &part_size))
fs/hfsplus/wrapper.c
153
sbi->s_vhdr_buf = kmalloc(hfsplus_min_io_size(sb), GFP_KERNEL);
fs/hfsplus/wrapper.c
156
sbi->s_backup_vhdr_buf = kmalloc(hfsplus_min_io_size(sb), GFP_KERNEL);
fs/hfsplus/wrapper.c
161
error = hfsplus_submit_bio(sb, part_start + HFSPLUS_VOLHEAD_SECTOR,
fs/hfsplus/wrapper.c
188
if (hfs_part_find(sb, &part_start, &part_size))
fs/hfsplus/wrapper.c
193
error = hfsplus_submit_bio(sb, part_start + part_size - 2,
fs/hfsplus/wrapper.c
222
if (sb_set_blocksize(sb, blocksize) != blocksize) {
fs/hfsplus/wrapper.c
228
part_start >> (sb->s_blocksize_bits - HFSPLUS_SECTOR_SHIFT);
fs/hfsplus/wrapper.c
231
sbi->fs_shift = sbi->alloc_blksz_shift - sb->s_blocksize_bits;
fs/hfsplus/wrapper.c
48
int hfsplus_submit_bio(struct super_block *sb, sector_t sector,
fs/hfsplus/wrapper.c
51
u64 io_size = hfsplus_min_io_size(sb);
fs/hfsplus/wrapper.c
63
return bdev_rw_virt(sb->s_bdev, sector, buf, io_size, opf);
fs/hfsplus/wrapper.c
98
static int hfsplus_get_last_session(struct super_block *sb,
fs/hfsplus/xattr.c
121
static int hfsplus_create_attributes_file(struct super_block *sb)
fs/hfsplus/xattr.c
124
struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
fs/hfsplus/xattr.c
169
attr_file = hfsplus_iget(sb, HFSPLUS_ATTR_CNID);
fs/hfsplus/xattr.c
183
clump_size = hfsplus_calc_btree_clump_size(sb->s_blocksize,
fs/hfsplus/xattr.c
241
sbi->attr_tree = hfs_btree_open(sb, HFSPLUS_ATTR_CNID);
fs/hostfs/hostfs_kern.c
212
static struct inode *hostfs_alloc_inode(struct super_block *sb)
fs/hostfs/hostfs_kern.c
216
hi = alloc_inode_sb(sb, hostfs_inode_cache, GFP_KERNEL_ACCOUNT);
fs/hostfs/hostfs_kern.c
570
static struct inode *hostfs_iget(struct super_block *sb, char *name)
fs/hostfs/hostfs_kern.c
579
inode = iget5_locked(sb, st.ino, hostfs_inode_test, hostfs_inode_set,
fs/hostfs/hostfs_kern.c
928
static int hostfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/hostfs/hostfs_kern.c
930
struct hostfs_fs_info *fsi = sb->s_fs_info;
fs/hostfs/hostfs_kern.c
934
sb->s_blocksize = 1024;
fs/hostfs/hostfs_kern.c
935
sb->s_blocksize_bits = 10;
fs/hostfs/hostfs_kern.c
936
sb->s_magic = HOSTFS_SUPER_MAGIC;
fs/hostfs/hostfs_kern.c
937
sb->s_op = &hostfs_sbops;
fs/hostfs/hostfs_kern.c
938
sb->s_d_flags = DCACHE_DONTCACHE;
fs/hostfs/hostfs_kern.c
939
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/hostfs/hostfs_kern.c
940
err = super_setup_bdi(sb);
fs/hostfs/hostfs_kern.c
944
root_inode = hostfs_iget(sb, fsi->host_root_path);
fs/hostfs/hostfs_kern.c
956
root_inode = hostfs_iget(sb, name);
fs/hostfs/hostfs_kern.c
962
sb->s_root = d_make_root(root_inode);
fs/hostfs/hostfs_kern.c
963
if (sb->s_root == NULL)
fs/hpfs/file.c
125
struct super_block *sb = inode->i_sb;
fs/hpfs/file.c
136
hpfs_lock(sb);
fs/hpfs/file.c
139
n_secs = hpfs_search_hotfix_map_for_range(sb, s,
fs/hpfs/file.c
142
s = hpfs_search_hotfix_map(sb, s);
fs/hpfs/file.c
155
hpfs_unlock(sb);
fs/hpfs/hpfs_fn.h
320
static inline struct hpfs_sb_info *hpfs_sb(struct super_block *sb)
fs/hpfs/hpfs_fn.h
322
return sb->s_fs_info;
fs/hpfs/inode.c
123
if (hpfs_sb(sb)->sb_chk >= 2) {
fs/hpfs/inode.c
125
if (hpfs_map_fnode(sb, hpfs_inode->i_parent_dir, &bh0)) brelse(bh0);
fs/hpfs/inode.c
16
struct super_block *sb = i->i_sb;
fs/hpfs/inode.c
19
i->i_uid = hpfs_sb(sb)->sb_uid;
fs/hpfs/inode.c
20
i->i_gid = hpfs_sb(sb)->sb_gid;
fs/hpfs/inode.c
21
i->i_mode = hpfs_sb(sb)->sb_mode;
fs/hpfs/inode.c
48
struct super_block *sb = i->i_sb;
fs/hpfs/inode.c
53
if (!(fnode = hpfs_map_fnode(sb, i->i_ino, &bh))) {
fs/hpfs/inode.c
91
umode_t mode = hpfs_sb(sb)->sb_mode;
fs/hpfs/super.c
234
static struct inode *hpfs_alloc_inode(struct super_block *sb)
fs/hpfs/super.c
237
ei = alloc_inode_sb(sb, hpfs_inode_cachep, GFP_NOFS);
fs/hpfs/super.c
29
struct hpfs_spare_block *sb;
fs/hpfs/super.c
30
if ((sb = hpfs_map_sector(s, 17, &bh, 0))) {
fs/hpfs/super.c
31
sb->dirty = 1;
fs/hpfs/super.c
32
sb->old_wrote = 0;
fs/hpfs/super.c
46
struct hpfs_spare_block *sb;
fs/hpfs/super.c
49
if ((sb = hpfs_map_sector(s, 17, &bh, 0))) {
fs/hpfs/super.c
50
sb->dirty = hpfs_sb(s)->sb_chkdsk > 1 - hpfs_sb(s)->sb_was_error;
fs/hpfs/super.c
51
sb->old_wrote = hpfs_sb(s)->sb_chkdsk >= 2 && !hpfs_sb(s)->sb_was_error;
fs/hpfs/super.c
723
struct super_block *sb = fc->root->d_sb;
fs/hpfs/super.c
724
struct hpfs_sb_info *sbi = hpfs_sb(sb);
fs/hugetlbfs/inode.c
1156
static void hugetlbfs_put_super(struct super_block *sb)
fs/hugetlbfs/inode.c
1158
struct hugetlbfs_sb_info *sbi = HUGETLBFS_SB(sb);
fs/hugetlbfs/inode.c
1161
sb->s_fs_info = NULL;
fs/hugetlbfs/inode.c
1197
static struct inode *hugetlbfs_alloc_inode(struct super_block *sb)
fs/hugetlbfs/inode.c
1199
struct hugetlbfs_sb_info *sbinfo = HUGETLBFS_SB(sb);
fs/hugetlbfs/inode.c
1204
p = alloc_inode_sb(sb, hugetlbfs_inode_cachep, GFP_KERNEL);
fs/hugetlbfs/inode.c
1405
hugetlbfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/hugetlbfs/inode.c
1413
sb->s_fs_info = sbinfo;
fs/hugetlbfs/inode.c
1435
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/hugetlbfs/inode.c
1436
sb->s_blocksize = huge_page_size(ctx->hstate);
fs/hugetlbfs/inode.c
1437
sb->s_blocksize_bits = huge_page_shift(ctx->hstate);
fs/hugetlbfs/inode.c
1438
sb->s_magic = HUGETLBFS_MAGIC;
fs/hugetlbfs/inode.c
1439
sb->s_op = &hugetlbfs_ops;
fs/hugetlbfs/inode.c
1440
sb->s_d_flags = DCACHE_DONTCACHE;
fs/hugetlbfs/inode.c
1441
sb->s_time_gran = 1;
fs/hugetlbfs/inode.c
1447
sb->s_stack_depth = FILESYSTEM_MAX_STACK_DEPTH;
fs/hugetlbfs/inode.c
1448
sb->s_root = d_make_root(hugetlbfs_get_root(sb, ctx));
fs/hugetlbfs/inode.c
1449
if (!sb->s_root)
fs/hugetlbfs/inode.c
896
static struct inode *hugetlbfs_get_root(struct super_block *sb,
fs/hugetlbfs/inode.c
901
inode = new_inode(sb);
fs/hugetlbfs/inode.c
925
static struct inode *hugetlbfs_get_inode(struct super_block *sb,
fs/hugetlbfs/inode.c
943
inode = new_inode(sb);
fs/inode.c
1029
long prune_icache_sb(struct super_block *sb, struct shrink_control *sc)
fs/inode.c
1034
freed = list_lru_shrink_walk(&sb->s_inode_lru, sc,
fs/inode.c
1045
static struct inode *find_inode(struct super_block *sb,
fs/inode.c
1061
if (inode->i_sb != sb)
fs/inode.c
1089
static struct inode *find_inode_fast(struct super_block *sb,
fs/inode.c
1105
if (inode->i_sb != sb)
fs/inode.c
1181
struct inode *new_inode(struct super_block *sb)
fs/inode.c
1185
inode = alloc_inode(sb);
fs/inode.c
1381
struct inode *iget5_locked(struct super_block *sb, unsigned long hashval,
fs/inode.c
1385
struct inode *inode = ilookup5(sb, hashval, test, data);
fs/inode.c
1388
struct inode *new = alloc_inode(sb);
fs/inode.c
1411
struct inode *iget5_locked_rcu(struct super_block *sb, unsigned long hashval,
fs/inode.c
1415
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1422
inode = find_inode(sb, head, test, data, false, &isnew);
fs/inode.c
1435
new = alloc_inode(sb);
fs/inode.c
1458
struct inode *iget_locked(struct super_block *sb, unsigned long ino)
fs/inode.c
1460
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1467
inode = find_inode_fast(sb, head, ino, false, &isnew);
fs/inode.c
1480
inode = alloc_inode(sb);
fs/inode.c
1486
old = find_inode_fast(sb, head, ino, true, &isnew);
fs/inode.c
1530
static int test_inode_iunique(struct super_block *sb, unsigned long ino)
fs/inode.c
1532
struct hlist_head *b = inode_hashtable + hash(sb, ino);
fs/inode.c
1536
if (inode->i_ino == ino && inode->i_sb == sb)
fs/inode.c
1556
ino_t iunique(struct super_block *sb, ino_t max_reserved)
fs/inode.c
1573
} while (!test_inode_iunique(sb, res));
fs/inode.c
1619
struct inode *ilookup5_nowait(struct super_block *sb, unsigned long hashval,
fs/inode.c
1622
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1626
inode = find_inode(sb, head, test, data, true, isnew);
fs/inode.c
1650
struct inode *ilookup5(struct super_block *sb, unsigned long hashval,
fs/inode.c
1659
inode = ilookup5_nowait(sb, hashval, test, data, &isnew);
fs/inode.c
1680
struct inode *ilookup(struct super_block *sb, unsigned long ino)
fs/inode.c
1682
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1689
inode = find_inode_fast(sb, head, ino, false, &isnew);
fs/inode.c
1728
struct inode *find_inode_nowait(struct super_block *sb,
fs/inode.c
1734
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1740
if (inode->i_sb != sb)
fs/inode.c
1776
struct inode *find_inode_rcu(struct super_block *sb, unsigned long hashval,
fs/inode.c
1779
struct hlist_head *head = inode_hashtable + hash(sb, hashval);
fs/inode.c
1786
if (inode->i_sb == sb &&
fs/inode.c
1814
struct inode *find_inode_by_ino_rcu(struct super_block *sb,
fs/inode.c
1817
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1825
inode->i_sb == sb &&
fs/inode.c
1835
struct super_block *sb = inode->i_sb;
fs/inode.c
1837
struct hlist_head *head = inode_hashtable + hash(sb, ino);
fs/inode.c
1849
if (old->i_sb != sb)
fs/inode.c
1924
struct super_block *sb = inode->i_sb;
fs/inode.c
1938
(sb->s_flags & SB_ACTIVE)) {
fs/inode.c
228
int inode_init_always_gfp(struct super_block *sb, struct inode *inode, gfp_t gfp)
fs/inode.c
234
inode->i_sb = sb;
fs/inode.c
235
inode->i_blkbits = sb->s_blocksize_bits;
fs/inode.c
245
if (sb->s_xattr)
fs/inode.c
247
if (sb->s_type->fs_flags & FS_MGTIME)
fs/inode.c
271
lockdep_set_class(&inode->i_lock, &sb->s_type->i_lock_key);
fs/inode.c
274
lockdep_set_class(&inode->i_rwsem, &sb->s_type->i_mutex_key);
fs/inode.c
2795
struct super_block *sb = inode->i_sb;
fs/inode.c
2796
unsigned int gran = sb->s_time_gran;
fs/inode.c
2798
t.tv_sec = clamp(t.tv_sec, sb->s_time_min, sb->s_time_max);
fs/inode.c
2799
if (unlikely(t.tv_sec == sb->s_time_max || t.tv_sec == sb->s_time_min))
fs/inode.c
291
&sb->s_type->invalidate_lock_key,
fs/inode.c
293
if (sb->s_iflags & SB_I_STABLE_WRITES)
fs/inode.c
3031
struct super_block *sb;
fs/inode.c
3041
if (get_kernel_nofault(sb, &inode->i_sb) ||
fs/inode.c
3052
if (!sb ||
fs/inode.c
3053
get_kernel_nofault(s_type, &sb->s_type) || !s_type ||
fs/inode.c
341
struct inode *alloc_inode(struct super_block *sb)
fs/inode.c
343
const struct super_operations *ops = sb->s_op;
fs/inode.c
347
inode = ops->alloc_inode(sb);
fs/inode.c
349
inode = alloc_inode_sb(sb, inode_cachep, GFP_KERNEL);
fs/inode.c
354
if (unlikely(inode_init_always(sb, inode))) {
fs/inode.c
656
struct super_block *sb = inode->i_sb;
fs/inode.c
658
spin_lock(&sb->s_inode_list_lock);
fs/inode.c
659
list_add(&inode->i_sb_list, &sb->s_inodes);
fs/inode.c
660
spin_unlock(&sb->s_inode_list_lock);
fs/inode.c
666
struct super_block *sb = inode->i_sb;
fs/inode.c
669
spin_lock(&sb->s_inode_list_lock);
fs/inode.c
671
spin_unlock(&sb->s_inode_list_lock);
fs/inode.c
675
static unsigned long hash(struct super_block *sb, unsigned long hashval)
fs/inode.c
679
tmp = (hashval * (unsigned long)sb) ^ (GOLDEN_RATIO_PRIME + hashval) /
fs/inode.c
902
void evict_inodes(struct super_block *sb)
fs/inode.c
908
spin_lock(&sb->s_inode_list_lock);
fs/inode.c
909
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/inode.c
934
spin_unlock(&sb->s_inode_list_lock);
fs/inode.c
940
spin_unlock(&sb->s_inode_list_lock);
fs/internal.h
136
extern bool super_trylock_shared(struct super_block *sb);
fs/internal.h
138
void put_super(struct super_block *sb);
fs/internal.h
140
int sb_init_dio_done_wq(struct super_block *sb);
fs/internal.h
149
static inline void sb_start_ro_state_change(struct super_block *sb)
fs/internal.h
151
WRITE_ONCE(sb->s_readonly_remount, 1);
fs/internal.h
169
static inline void sb_end_ro_state_change(struct super_block *sb)
fs/internal.h
180
WRITE_ONCE(sb->s_readonly_remount, 0);
fs/internal.h
212
extern long prune_icache_sb(struct super_block *sb, struct shrink_control *sc);
fs/internal.h
227
extern long prune_dcache_sb(struct super_block *sb, struct shrink_control *sc);
fs/ioctl.c
387
struct super_block *sb = file_inode(filp)->i_sb;
fs/ioctl.c
389
if (!ns_capable(sb->s_user_ns, CAP_SYS_ADMIN))
fs/ioctl.c
393
if (sb->s_op->freeze_fs == NULL && sb->s_op->freeze_super == NULL)
fs/ioctl.c
397
if (sb->s_op->freeze_super)
fs/ioctl.c
398
return sb->s_op->freeze_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/ioctl.c
399
return freeze_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/ioctl.c
404
struct super_block *sb = file_inode(filp)->i_sb;
fs/ioctl.c
406
if (!ns_capable(sb->s_user_ns, CAP_SYS_ADMIN))
fs/ioctl.c
410
if (sb->s_op->thaw_super)
fs/ioctl.c
411
return sb->s_op->thaw_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/ioctl.c
412
return thaw_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/ioctl.c
457
struct super_block *sb = file_inode(file)->i_sb;
fs/ioctl.c
458
struct fsuuid2 u = { .len = sb->s_uuid_len, };
fs/ioctl.c
460
if (!sb->s_uuid_len)
fs/ioctl.c
463
memcpy(&u.uuid[0], &sb->s_uuid, sb->s_uuid_len);
fs/ioctl.c
470
struct super_block *sb = file_inode(file)->i_sb;
fs/ioctl.c
472
if (!strlen(sb->s_sysfs_name))
fs/ioctl.c
477
u.len = scnprintf(u.name, sizeof(u.name), "%s/%s", sb->s_type->name, sb->s_sysfs_name);
fs/ioctl.c
61
struct super_block *sb = inode->i_sb;
fs/ioctl.c
82
sb->s_id, filp);
fs/isofs/export.c
162
static struct dentry *isofs_fh_to_dentry(struct super_block *sb,
fs/isofs/export.c
170
return isofs_export_iget(sb, ifid->block, ifid->offset,
fs/isofs/export.c
174
static struct dentry *isofs_fh_to_parent(struct super_block *sb,
fs/isofs/export.c
182
return isofs_export_iget(sb,
fs/isofs/export.c
20
isofs_export_iget(struct super_block *sb,
fs/isofs/export.c
29
inode = isofs_iget(sb, block, offset);
fs/isofs/inode.c
1000
struct super_block *sb = dentry->d_sb;
fs/isofs/inode.c
1001
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/isofs/inode.c
1004
buf->f_bsize = sb->s_blocksize;
fs/isofs/inode.c
1005
buf->f_blocks = (ISOFS_SB(sb)->s_nzones
fs/isofs/inode.c
1006
<< (ISOFS_SB(sb)->s_log_zone_size - sb->s_blocksize_bits));
fs/isofs/inode.c
1009
buf->f_files = ISOFS_SB(sb)->s_ninodes;
fs/isofs/inode.c
1271
struct super_block *sb = inode->i_sb;
fs/isofs/inode.c
1272
struct isofs_sb_info *sbi = ISOFS_SB(sb);
fs/isofs/inode.c
1499
struct inode *__isofs_iget(struct super_block *sb,
fs/isofs/inode.c
1509
if (offset >= 1ul << sb->s_blocksize_bits)
fs/isofs/inode.c
1515
hashval = (block << sb->s_blocksize_bits) | offset;
fs/isofs/inode.c
1517
inode = iget5_locked(sb, hashval, &isofs_iget5_test,
fs/isofs/inode.c
503
static unsigned int isofs_get_last_session(struct super_block *sb, s32 session)
fs/isofs/inode.c
505
struct cdrom_device_info *cdi = disk_to_cdi(sb->s_bdev->bd_disk);
fs/isofs/inode.c
52
static void isofs_put_super(struct super_block *sb)
fs/isofs/inode.c
54
struct isofs_sb_info *sbi = ISOFS_SB(sb);
fs/isofs/inode.c
550
static bool rootdir_empty(struct super_block *sb, unsigned long block)
fs/isofs/inode.c
556
bh = sb_bread(sb, block);
fs/isofs/inode.c
61
sb->s_fs_info = NULL;
fs/isofs/inode.c
71
static struct inode *isofs_alloc_inode(struct super_block *sb)
fs/isofs/inode.c
74
ei = alloc_inode_sb(sb, isofs_inode_cachep, GFP_KERNEL);
fs/isofs/isofs.h
126
struct inode *__isofs_iget(struct super_block *sb,
fs/isofs/isofs.h
131
static inline struct inode *isofs_iget(struct super_block *sb,
fs/isofs/isofs.h
135
return __isofs_iget(sb, block, offset, 0);
fs/isofs/isofs.h
138
static inline struct inode *isofs_iget_reloc(struct super_block *sb,
fs/isofs/isofs.h
142
return __isofs_iget(sb, block, offset, 1);
fs/isofs/isofs.h
65
static inline struct isofs_sb_info *ISOFS_SB(struct super_block *sb)
fs/isofs/isofs.h
67
return sb->s_fs_info;
fs/jbd2/journal.c
118
static __be32 jbd2_superblock_csum(journal_superblock_t *sb)
fs/jbd2/journal.c
123
old_csum = sb->s_checksum;
fs/jbd2/journal.c
124
sb->s_checksum = 0;
fs/jbd2/journal.c
125
csum = jbd2_chksum(~0, (char *)sb, sizeof(journal_superblock_t));
fs/jbd2/journal.c
126
sb->s_checksum = old_csum;
fs/jbd2/journal.c
1311
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
1315
if (sb->s_header.h_magic != cpu_to_be32(JBD2_MAGIC_NUMBER) ||
fs/jbd2/journal.c
1316
sb->s_blocksize != cpu_to_be32(journal->j_blocksize)) {
fs/jbd2/journal.c
1321
if (be32_to_cpu(sb->s_header.h_blocktype) != JBD2_SUPERBLOCK_V1 &&
fs/jbd2/journal.c
1322
be32_to_cpu(sb->s_header.h_blocktype) != JBD2_SUPERBLOCK_V2) {
fs/jbd2/journal.c
1327
if (be32_to_cpu(sb->s_maxlen) > journal->j_total_len) {
fs/jbd2/journal.c
1332
if (be32_to_cpu(sb->s_first) == 0 ||
fs/jbd2/journal.c
1333
be32_to_cpu(sb->s_first) >= journal->j_total_len) {
fs/jbd2/journal.c
1336
be32_to_cpu(sb->s_first));
fs/jbd2/journal.c
1347
if ((sb->s_feature_ro_compat &
fs/jbd2/journal.c
1349
(sb->s_feature_incompat &
fs/jbd2/journal.c
1356
jbd2_journal_get_num_fc_blks(sb) : 0;
fs/jbd2/journal.c
1357
if (be32_to_cpu(sb->s_maxlen) < JBD2_MIN_JOURNAL_BLOCKS ||
fs/jbd2/journal.c
1358
be32_to_cpu(sb->s_maxlen) - JBD2_MIN_JOURNAL_BLOCKS < num_fc_blks) {
fs/jbd2/journal.c
1360
be32_to_cpu(sb->s_maxlen), num_fc_blks);
fs/jbd2/journal.c
1381
if (sb->s_checksum_type != JBD2_CRC32C_CHKSUM) {
fs/jbd2/journal.c
1387
if (sb->s_checksum != jbd2_superblock_csum(sb)) {
fs/jbd2/journal.c
1462
journal_superblock_t *sb;
fs/jbd2/journal.c
1475
sb = (journal_superblock_t *)bh->b_data;
fs/jbd2/journal.c
1476
journal->j_superblock = sb;
fs/jbd2/journal.c
1483
journal->j_tail_sequence = be32_to_cpu(sb->s_sequence);
fs/jbd2/journal.c
1484
journal->j_tail = be32_to_cpu(sb->s_start);
fs/jbd2/journal.c
1485
journal->j_first = be32_to_cpu(sb->s_first);
fs/jbd2/journal.c
1486
journal->j_errno = be32_to_cpu(sb->s_errno);
fs/jbd2/journal.c
1487
journal->j_last = be32_to_cpu(sb->s_maxlen);
fs/jbd2/journal.c
1489
if (be32_to_cpu(sb->s_maxlen) < journal->j_total_len)
fs/jbd2/journal.c
1490
journal->j_total_len = be32_to_cpu(sb->s_maxlen);
fs/jbd2/journal.c
1493
journal->j_csum_seed = jbd2_chksum(~0, sb->s_uuid,
fs/jbd2/journal.c
1494
sizeof(sb->s_uuid));
fs/jbd2/journal.c
1499
journal->j_fc_last = be32_to_cpu(sb->s_maxlen);
fs/jbd2/journal.c
1501
jbd2_journal_get_num_fc_blks(sb);
fs/jbd2/journal.c
1708
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
1711
first = be32_to_cpu(sb->s_first);
fs/jbd2/journal.c
1712
last = be32_to_cpu(sb->s_maxlen);
fs/jbd2/journal.c
1757
if (sb->s_start == 0) {
fs/jbd2/journal.c
1787
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
1822
sb->s_checksum = jbd2_superblock_csum(sb);
fs/jbd2/journal.c
1856
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
1872
sb->s_sequence = cpu_to_be32(tail_tid);
fs/jbd2/journal.c
1873
sb->s_start = cpu_to_be32(tail_block);
fs/jbd2/journal.c
1898
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
1903
if (sb->s_start == 0) { /* Is it already empty? */
fs/jbd2/journal.c
1911
sb->s_sequence = cpu_to_be32(journal->j_tail_sequence);
fs/jbd2/journal.c
1912
sb->s_start = cpu_to_be32(0);
fs/jbd2/journal.c
1913
sb->s_head = cpu_to_be32(journal->j_head);
fs/jbd2/journal.c
2042
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
2050
sb->s_errno = cpu_to_be32(errcode);
fs/jbd2/journal.c
2067
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
2072
err = jbd2_journal_create_slab(be32_to_cpu(sb->s_blocksize));
fs/jbd2/journal.c
2214
journal_superblock_t *sb;
fs/jbd2/journal.c
2221
sb = journal->j_superblock;
fs/jbd2/journal.c
2223
if (((be32_to_cpu(sb->s_feature_compat) & compat) == compat) &&
fs/jbd2/journal.c
2224
((be32_to_cpu(sb->s_feature_ro_compat) & ro) == ro) &&
fs/jbd2/journal.c
2225
((be32_to_cpu(sb->s_feature_incompat) & incompat) == incompat))
fs/jbd2/journal.c
2262
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/journal.c
2265
num_fc_blks = jbd2_journal_get_num_fc_blks(sb);
fs/jbd2/journal.c
2301
((incompat & (f)) && !(sb->s_feature_incompat & cpu_to_be32(f)))
fs/jbd2/journal.c
2303
((compat & (f)) && !(sb->s_feature_compat & cpu_to_be32(f)))
fs/jbd2/journal.c
2304
journal_superblock_t *sb;
fs/jbd2/journal.c
2326
sb = journal->j_superblock;
fs/jbd2/journal.c
2339
sb->s_checksum_type = JBD2_CRC32C_CHKSUM;
fs/jbd2/journal.c
2340
sb->s_feature_compat &=
fs/jbd2/journal.c
2342
journal->j_csum_seed = jbd2_chksum(~0, sb->s_uuid,
fs/jbd2/journal.c
2343
sizeof(sb->s_uuid));
fs/jbd2/journal.c
2348
sb->s_feature_incompat &=
fs/jbd2/journal.c
2352
sb->s_feature_compat |= cpu_to_be32(compat);
fs/jbd2/journal.c
2353
sb->s_feature_ro_compat |= cpu_to_be32(ro);
fs/jbd2/journal.c
2354
sb->s_feature_incompat |= cpu_to_be32(incompat);
fs/jbd2/journal.c
2360
sb->s_checksum = jbd2_superblock_csum(sb);
fs/jbd2/journal.c
2383
journal_superblock_t *sb;
fs/jbd2/journal.c
2388
sb = journal->j_superblock;
fs/jbd2/journal.c
2391
sb->s_feature_compat &= ~cpu_to_be32(compat);
fs/jbd2/journal.c
2392
sb->s_feature_ro_compat &= ~cpu_to_be32(ro);
fs/jbd2/journal.c
2393
sb->s_feature_incompat &= ~cpu_to_be32(incompat);
fs/jbd2/journal.c
2399
sb->s_checksum = jbd2_superblock_csum(sb);
fs/jbd2/recovery.c
297
journal_superblock_t *sb = journal->j_superblock;
fs/jbd2/recovery.c
300
be32_to_cpu(sb->s_sequence), be32_to_cpu(sb->s_head));
fs/jbd2/recovery.c
301
journal->j_transaction_sequence = be32_to_cpu(sb->s_sequence) + 1;
fs/jbd2/recovery.c
302
journal->j_head = be32_to_cpu(sb->s_head);
fs/jbd2/recovery.c
594
journal_superblock_t * sb;
fs/jbd2/recovery.c
609
sb = journal->j_superblock;
fs/jbd2/recovery.c
610
next_commit_ID = be32_to_cpu(sb->s_sequence);
fs/jbd2/recovery.c
611
next_log_block = be32_to_cpu(sb->s_start);
fs/jffs2/fs.c
253
struct inode *jffs2_iget(struct super_block *sb, unsigned long ino)
fs/jffs2/fs.c
265
inode = iget_locked(sb, ino);
fs/jffs2/fs.c
396
int jffs2_do_remount_fs(struct super_block *sb, struct fs_context *fc)
fs/jffs2/fs.c
398
struct jffs2_sb_info *c = JFFS2_SB_INFO(sb);
fs/jffs2/fs.c
400
if (c->flags & JFFS2_SB_FLAG_RO && !sb_rdonly(sb))
fs/jffs2/fs.c
407
if (!sb_rdonly(sb)) {
fs/jffs2/fs.c
426
struct super_block *sb = dir_i->i_sb;
fs/jffs2/fs.c
434
c = JFFS2_SB_INFO(sb);
fs/jffs2/fs.c
436
inode = new_inode(sb);
fs/jffs2/fs.c
514
int jffs2_do_fill_super(struct super_block *sb, struct fs_context *fc)
fs/jffs2/fs.c
521
c = JFFS2_SB_INFO(sb);
fs/jffs2/fs.c
578
root_i = jffs2_iget(sb, 1);
fs/jffs2/fs.c
588
sb->s_root = d_make_root(root_i);
fs/jffs2/fs.c
589
if (!sb->s_root)
fs/jffs2/fs.c
592
sb->s_maxbytes = 0xFFFFFFFF;
fs/jffs2/fs.c
593
sb->s_blocksize = PAGE_SIZE;
fs/jffs2/fs.c
594
sb->s_blocksize_bits = PAGE_SHIFT;
fs/jffs2/fs.c
595
sb->s_magic = JFFS2_SUPER_MAGIC;
fs/jffs2/fs.c
596
sb->s_time_min = 0;
fs/jffs2/fs.c
597
sb->s_time_max = U32_MAX;
fs/jffs2/fs.c
599
if (!sb_rdonly(sb))
fs/jffs2/os-linux.h
175
int jffs2_do_remount_fs(struct super_block *sb, struct fs_context *fc);
fs/jffs2/os-linux.h
176
int jffs2_do_fill_super(struct super_block *sb, struct fs_context *fc);
fs/jffs2/os-linux.h
24
#define JFFS2_SB_INFO(sb) (sb->s_fs_info)
fs/jffs2/super.c
100
struct jffs2_sb_info *c = JFFS2_SB_INFO(sb);
fs/jffs2/super.c
113
static struct inode *jffs2_nfs_get_inode(struct super_block *sb, uint64_t ino,
fs/jffs2/super.c
119
return jffs2_iget(sb, ino);
fs/jffs2/super.c
122
static struct dentry *jffs2_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/jffs2/super.c
125
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/jffs2/super.c
129
static struct dentry *jffs2_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/jffs2/super.c
132
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/jffs2/super.c
236
struct super_block *sb = fc->root->d_sb;
fs/jffs2/super.c
238
sync_filesystem(sb);
fs/jffs2/super.c
241
return jffs2_do_remount_fs(sb, fc);
fs/jffs2/super.c
259
static int jffs2_fill_super(struct super_block *sb, struct fs_context *fc)
fs/jffs2/super.c
261
struct jffs2_sb_info *c = sb->s_fs_info;
fs/jffs2/super.c
265
sb->s_mtd->index, sb->s_mtd->name);
fs/jffs2/super.c
267
c->mtd = sb->s_mtd;
fs/jffs2/super.c
268
c->os_priv = sb;
fs/jffs2/super.c
283
sb->s_op = &jffs2_super_operations;
fs/jffs2/super.c
284
sb->s_export_op = &jffs2_export_ops;
fs/jffs2/super.c
285
sb->s_flags = sb->s_flags | SB_NOATIME;
fs/jffs2/super.c
286
sb->s_xattr = jffs2_xattr_handlers;
fs/jffs2/super.c
288
sb->s_flags |= SB_POSIXACL;
fs/jffs2/super.c
290
return jffs2_do_fill_super(sb, fc);
fs/jffs2/super.c
323
static void jffs2_put_super (struct super_block *sb)
fs/jffs2/super.c
325
struct jffs2_sb_info *c = JFFS2_SB_INFO(sb);
fs/jffs2/super.c
345
static void jffs2_kill_sb(struct super_block *sb)
fs/jffs2/super.c
347
struct jffs2_sb_info *c = JFFS2_SB_INFO(sb);
fs/jffs2/super.c
348
if (c && !sb_rdonly(sb))
fs/jffs2/super.c
350
kill_mtd_super(sb);
fs/jffs2/super.c
38
static struct inode *jffs2_alloc_inode(struct super_block *sb)
fs/jffs2/super.c
42
f = alloc_inode_sb(sb, jffs2_inode_cachep, GFP_KERNEL);
fs/jffs2/super.c
98
static int jffs2_sync_fs(struct super_block *sb, int wait)
fs/jffs2/wbuf.c
1163
struct super_block *sb = OFNI_BS_2SFFJ(c);
fs/jffs2/wbuf.c
1165
if (!sb_rdonly(sb)) {
fs/jffs2/wbuf.c
1173
struct super_block *sb = OFNI_BS_2SFFJ(c);
fs/jffs2/wbuf.c
1176
if (sb_rdonly(sb))
fs/jfs/inode.c
24
struct inode *jfs_iget(struct super_block *sb, unsigned long ino)
fs/jfs/inode.c
29
inode = iget_locked(sb, ino);
fs/jfs/ioctl.c
112
struct super_block *sb = inode->i_sb;
fs/jfs/ioctl.c
119
if (!bdev_max_discard_sectors(sb->s_bdev)) {
fs/jfs/ioctl.c
129
bdev_discard_granularity(sb->s_bdev));
fs/jfs/jfs_discard.c
110
up_read(&sb->s_umount);
fs/jfs/jfs_discard.c
111
range->len = trimmed << sb->s_blocksize_bits;
fs/jfs/jfs_discard.c
34
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_discard.c
37
r = sb_issue_discard(sb, blkno, nblocks, GFP_NOFS, 0);
fs/jfs/jfs_discard.c
40
sb, (unsigned long long)blkno,
fs/jfs/jfs_discard.c
45
sb, (unsigned long long)blkno,
fs/jfs/jfs_discard.c
69
struct super_block *sb = ipbmap->i_sb;
fs/jfs/jfs_discard.c
80
start = range->start >> sb->s_blocksize_bits;
fs/jfs/jfs_discard.c
81
end = start + (range->len >> sb->s_blocksize_bits) - 1;
fs/jfs/jfs_discard.c
82
minlen = range->minlen >> sb->s_blocksize_bits;
fs/jfs/jfs_discard.c
86
down_read(&sb->s_umount);
fs/jfs/jfs_discard.c
92
range->len < sb->s_blocksize) {
fs/jfs/jfs_discard.c
93
up_read(&sb->s_umount);
fs/jfs/jfs_dmap.c
1578
struct super_block *sb = ipbmap->i_sb;
fs/jfs/jfs_dmap.c
1643
if (!(JFS_SBI(sb)->flag & JFS_DISCARD))
fs/jfs/jfs_dmap.c
352
struct super_block *sb = ipbmap->i_sb;
fs/jfs/jfs_dmap.c
369
if (JFS_SBI(sb)->flag & JFS_DISCARD)
fs/jfs/jfs_dmap.c
370
if (JFS_SBI(sb)->minblks_trim <= nblocks)
fs/jfs/jfs_dmap.c
4036
struct super_block *sb = ipbmap->i_sb;
fs/jfs/jfs_dmap.c
4042
nblocks = ipbmap->i_size >> JFS_SBI(sb)->l2bsize;
fs/jfs/jfs_dmap.c
4043
npages = nblocks >> JFS_SBI(sb)->l2nbperpage;
fs/jfs/jfs_dmap.c
463
log = (struct jfs_log *) JFS_SBI(tblk->sb)->log;
fs/jfs/jfs_dtree.c
1641
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
1735
xsize = xlen << JFS_SBI(sb)->l2bsize;
fs/jfs/jfs_dtree.c
1874
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
1909
xsize = xlen << JFS_SBI(sb)->l2bsize;
fs/jfs/jfs_dtree.c
324
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
325
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_dtree.c
583
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_dtree.c
598
if ((JFS_SBI(sb)->mntflag & JFS_OS2) == JFS_OS2) {
fs/jfs/jfs_dtree.c
646
JFS_SBI(sb)->mntflag);
fs/jfs/jfs_dtree.c
773
jfs_error(sb, "stack overrun!\n");
fs/jfs/jfs_extent.c
199
struct super_block *sb = ip->i_sb;
fs/jfs/jfs_extent.c
200
int nbperpage = JFS_SBI(sb)->nbperpage;
fs/jfs/jfs_extent.c
213
prev = ((offset & ~POFFSET) >> JFS_SBI(sb)->l2bsize) - nbperpage;
fs/jfs/jfs_filsys.h
142
#define LBLK2PBLK(sb,b) ((b) << (sb->s_blocksize_bits - L2PBSIZE))
fs/jfs/jfs_filsys.h
143
#define PBLK2LBLK(sb,b) ((b) >> (sb->s_blocksize_bits - L2PBSIZE))
fs/jfs/jfs_imap.c
2445
struct super_block *sb;
fs/jfs/jfs_imap.c
2456
sb = ipimap->i_sb;
fs/jfs/jfs_imap.c
2457
sbi = JFS_SBI(sb);
fs/jfs/jfs_imap.c
2521
tid = txBegin(sb, COMMIT_FORCE);
fs/jfs/jfs_imap.c
2597
duplicateIXtree(sb, blkno, xlen, &xaddr);
fs/jfs/jfs_imap.c
2809
log = JFS_SBI(tblk->sb)->log;
fs/jfs/jfs_imap.c
2982
static void duplicateIXtree(struct super_block *sb, s64 blkno,
fs/jfs/jfs_imap.c
2991
if (JFS_SBI(sb)->mntflag & JFS_BAD_SAIT) /* s_flag */
fs/jfs/jfs_imap.c
2993
ip = diReadSpecial(sb, FILESYSTEM_I, 1);
fs/jfs/jfs_imap.c
2995
JFS_SBI(sb)->mntflag |= JFS_BAD_SAIT;
fs/jfs/jfs_imap.c
2996
if (readSuper(sb, &bh))
fs/jfs/jfs_imap.c
3008
tid = txBegin(sb, COMMIT_FORCE);
fs/jfs/jfs_imap.c
3011
JFS_SBI(sb)->mntflag |= JFS_BAD_SAIT;
fs/jfs/jfs_imap.c
418
struct inode *diReadSpecial(struct super_block *sb, ino_t inum, int secondary)
fs/jfs/jfs_imap.c
420
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_imap.c
426
ip = new_inode(sb);
fs/jfs/jfs_incore.h
160
struct super_block *sb; /* Point back to vfs super block */
fs/jfs/jfs_incore.h
215
static inline struct jfs_sb_info *JFS_SBI(struct super_block *sb)
fs/jfs/jfs_incore.h
217
return sb->s_fs_info;
fs/jfs/jfs_inode.c
102
inode->i_generation = JFS_SBI(sb)->gengen++;
fs/jfs/jfs_inode.c
43
struct super_block *sb = parent->i_sb;
fs/jfs/jfs_inode.c
48
inode = new_inode(sb);
fs/jfs/jfs_inode.h
25
extern struct dentry *jfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/jfs/jfs_inode.h
27
extern struct dentry *jfs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/jfs/jfs_logmgr.c
1058
int lmLogOpen(struct super_block *sb)
fs/jfs/jfs_logmgr.c
1063
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_logmgr.c
1066
return open_dummy_log(sb);
fs/jfs/jfs_logmgr.c
1069
return open_inline_log(sb);
fs/jfs/jfs_logmgr.c
1154
static int open_inline_log(struct super_block *sb)
fs/jfs/jfs_logmgr.c
1165
log->bdev_file = sb->s_bdev_file;
fs/jfs/jfs_logmgr.c
1166
log->base = addressPXD(&JFS_SBI(sb)->logpxd);
fs/jfs/jfs_logmgr.c
1167
log->size = lengthPXD(&JFS_SBI(sb)->logpxd) >>
fs/jfs/jfs_logmgr.c
1168
(L2LOGPSIZE - sb->s_blocksize_bits);
fs/jfs/jfs_logmgr.c
1169
log->l2bsize = sb->s_blocksize_bits;
fs/jfs/jfs_logmgr.c
1170
ASSERT(L2LOGPSIZE >= sb->s_blocksize_bits);
fs/jfs/jfs_logmgr.c
1181
list_add(&JFS_SBI(sb)->log_list, &log->sb_list);
fs/jfs/jfs_logmgr.c
1182
JFS_SBI(sb)->log = log;
fs/jfs/jfs_logmgr.c
1187
static int open_dummy_log(struct super_block *sb)
fs/jfs/jfs_logmgr.c
1213
list_add(&JFS_SBI(sb)->log_list, &dummy_log->sb_list);
fs/jfs/jfs_logmgr.c
1214
JFS_SBI(sb)->log = dummy_log;
fs/jfs/jfs_logmgr.c
1434
int lmLogClose(struct super_block *sb)
fs/jfs/jfs_logmgr.c
1436
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_logmgr.c
1453
sync_blockdev(sb->s_bdev);
fs/jfs/jfs_logmgr.c
170
static int open_inline_log(struct super_block *sb);
fs/jfs/jfs_logmgr.c
171
static int open_dummy_log(struct super_block *sb);
fs/jfs/jfs_logmgr.h
490
extern int lmLogOpen(struct super_block *sb);
fs/jfs/jfs_logmgr.h
491
extern int lmLogClose(struct super_block *sb);
fs/jfs/jfs_metapage.c
764
mp->sb = inode->i_sb;
fs/jfs/jfs_metapage.c
841
jfs_error(mp->sb, "metapage_write_one() failed\n");
fs/jfs/jfs_metapage.c
885
jfs_error(mp->sb, "metapage_write_one() failed\n");
fs/jfs/jfs_metapage.h
28
struct super_block *sb;
fs/jfs/jfs_mount.c
108
ipbmap = diReadSpecial(sb, BMAP_I, 0);
fs/jfs/jfs_mount.c
138
ipaimap2 = diReadSpecial(sb, AGGREGATE_I, 1);
fs/jfs/jfs_mount.c
166
ipimap = diReadSpecial(sb, FILESYSTEM_I, 0);
fs/jfs/jfs_mount.c
221
int jfs_mount_rw(struct super_block *sb, int remount)
fs/jfs/jfs_mount.c
223
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_mount.c
232
if (chkSuper(sb) || (sbi->state != FM_CLEAN))
fs/jfs/jfs_mount.c
257
if ((rc = lmLogOpen(sb)))
fs/jfs/jfs_mount.c
263
if ((rc = updateSuper(sb, FM_MOUNT))) {
fs/jfs/jfs_mount.c
265
lmLogClose(sb);
fs/jfs/jfs_mount.c
272
logMOUNT(sb);
fs/jfs/jfs_mount.c
287
static int chkSuper(struct super_block *sb)
fs/jfs/jfs_mount.c
290
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_mount.c
299
if ((rc = readSuper(sb, &bh)))
fs/jfs/jfs_mount.c
349
!sb_rdonly(sb)) {
fs/jfs/jfs_mount.c
402
int updateSuper(struct super_block *sb, uint state)
fs/jfs/jfs_mount.c
405
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_mount.c
423
if ((rc = readSuper(sb, &bh)))
fs/jfs/jfs_mount.c
458
int readSuper(struct super_block *sb, struct buffer_head **bpp)
fs/jfs/jfs_mount.c
461
*bpp = sb_bread(sb, SUPER1_OFF >> sb->s_blocksize_bits);
fs/jfs/jfs_mount.c
466
*bpp = sb_bread(sb, SUPER2_OFF >> sb->s_blocksize_bits);
fs/jfs/jfs_mount.c
488
static int logMOUNT(struct super_block *sb)
fs/jfs/jfs_mount.c
490
struct jfs_log *log = JFS_SBI(sb)->log;
fs/jfs/jfs_mount.c
497
lrd.aggregate = cpu_to_le32(new_encode_dev(sb->s_bdev->bd_dev));
fs/jfs/jfs_mount.c
55
static int logMOUNT(struct super_block *sb);
fs/jfs/jfs_mount.c
70
int jfs_mount(struct super_block *sb)
fs/jfs/jfs_mount.c
73
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_mount.c
83
if ((rc = chkSuper(sb))) {
fs/jfs/jfs_mount.c
87
ipaimap = diReadSpecial(sb, AGGREGATE_I, 0);
fs/jfs/jfs_txnmgr.c
1142
struct super_block *sb;
fs/jfs/jfs_txnmgr.c
1151
sb = cd.sb = iplist[0]->i_sb;
fs/jfs/jfs_txnmgr.c
1155
tid = txBegin(sb, 0);
fs/jfs/jfs_txnmgr.c
1161
log = JFS_SBI(sb)->log;
fs/jfs/jfs_txnmgr.c
1452
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1538
PXDlength(pxd, mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1589
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1618
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1708
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
1797
mp->logical_size >> tblk->sb->
fs/jfs/jfs_txnmgr.c
1918
mp->logical_size >> tblk->sb->s_blocksize_bits);
fs/jfs/jfs_txnmgr.c
2258
ipimap = JFS_SBI(tblk->sb)->ipimap;
fs/jfs/jfs_txnmgr.c
2638
jfs_error(tblk->sb, "\n");
fs/jfs/jfs_txnmgr.c
2667
log = (struct jfs_log *) JFS_SBI(tblk->sb)->log;
fs/jfs/jfs_txnmgr.c
2715
sbi = JFS_SBI(tblk->sb);
fs/jfs/jfs_txnmgr.c
2785
if (!(JFS_SBI(tblk->sb)->commit_state & IN_LAZYCOMMIT) &&
fs/jfs/jfs_txnmgr.c
2812
void txQuiesce(struct super_block *sb)
fs/jfs/jfs_txnmgr.c
2816
struct jfs_log *log = JFS_SBI(sb)->log;
fs/jfs/jfs_txnmgr.c
2868
void txResume(struct super_block *sb)
fs/jfs/jfs_txnmgr.c
2870
struct jfs_log *log = JFS_SBI(sb)->log;
fs/jfs/jfs_txnmgr.c
349
tid_t txBegin(struct super_block *sb, int flag)
fs/jfs/jfs_txnmgr.c
356
log = JFS_SBI(sb)->log;
fs/jfs/jfs_txnmgr.c
359
jfs_error(sb, "read-only filesystem\n");
fs/jfs/jfs_txnmgr.c
426
tblk->sb = sb;
fs/jfs/jfs_txnmgr.c
453
void txBeginAnon(struct super_block *sb)
fs/jfs/jfs_txnmgr.c
457
log = JFS_SBI(sb)->log;
fs/jfs/jfs_txnmgr.c
507
log = JFS_SBI(tblk->sb)->log;
fs/jfs/jfs_txnmgr.c
912
log = JFS_SBI(tblk->sb)->log;
fs/jfs/jfs_txnmgr.h
264
struct super_block *sb; /* superblock */
fs/jfs/jfs_txnmgr.h
34
struct super_block *sb; /* super block */
fs/jfs/jfs_umount.c
107
updateSuper(sb, FM_CLEAN);
fs/jfs/jfs_umount.c
114
rc = lmLogClose(sb);
fs/jfs/jfs_umount.c
121
int jfs_umount_rw(struct super_block *sb)
fs/jfs/jfs_umount.c
123
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_umount.c
150
updateSuper(sb, FM_CLEAN);
fs/jfs/jfs_umount.c
152
return lmLogClose(sb);
fs/jfs/jfs_umount.c
37
int jfs_umount(struct super_block *sb)
fs/jfs/jfs_umount.c
39
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/jfs_umount.c
47
jfs_info("UnMount JFS: sb:0x%p", sb);
fs/jfs/namei.c
1011
xaddr += JFS_SBI(sb)->nbperpage;
fs/jfs/namei.c
1480
static struct inode *jfs_nfs_get_inode(struct super_block *sb,
fs/jfs/namei.c
1487
inode = jfs_iget(sb, ino);
fs/jfs/namei.c
1499
struct dentry *jfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/jfs/namei.c
1502
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/jfs/namei.c
1506
struct dentry *jfs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/jfs/namei.c
1509
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/jfs/namei.c
893
struct super_block *sb;
fs/jfs/namei.c
985
sb = ip->i_sb;
fs/jfs/namei.c
986
bmask = JFS_SBI(sb)->bsize - 1;
fs/jfs/namei.c
989
xlen = xsize >> JFS_SBI(sb)->l2bsize;
fs/jfs/resize.c
198
txQuiesce(sb);
fs/jfs/resize.c
201
sbi->direct_inode->i_size = bdev_nr_bytes(sb->s_bdev);
fs/jfs/resize.c
224
if ((rc = readSuper(sb, &bh)))
fs/jfs/resize.c
255
log->size = newLogSize >> (L2LOGPSIZE - sb->s_blocksize_bits);
fs/jfs/resize.c
385
tid = txBegin(sb, COMMIT_FORCE);
fs/jfs/resize.c
467
ipbmap2 = diReadSpecial(sb, BMAP_I, 1);
fs/jfs/resize.c
482
if ((rc = readSuper(sb, &bh)))
fs/jfs/resize.c
50
int jfs_extendfs(struct super_block *sb, s64 newLVSize, int newLogSize)
fs/jfs/resize.c
508
bh2 = sb_bread(sb, SUPER2_OFF >> sb->s_blocksize_bits);
fs/jfs/resize.c
526
jfs_error(sb, "\n");
fs/jfs/resize.c
53
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/resize.c
532
txResume(sb);
fs/jfs/resize.c
89
VolumeSize = sb_bdev_nr_blocks(sb);
fs/jfs/resize.c
98
bh = sb_bread(sb, newLVSize - 1);
fs/jfs/super.c
101
static struct inode *jfs_alloc_inode(struct super_block *sb)
fs/jfs/super.c
105
jfs_inode = alloc_inode_sb(sb, jfs_inode_cachep, GFP_NOFS);
fs/jfs/super.c
157
static int jfs_quota_off(struct super_block *sb, int type);
fs/jfs/super.c
158
static int jfs_quota_on(struct super_block *sb, int type, int format_id,
fs/jfs/super.c
161
static void jfs_quota_off_umount(struct super_block *sb)
fs/jfs/super.c
166
jfs_quota_off(sb, type);
fs/jfs/super.c
180
static inline void jfs_quota_off_umount(struct super_block *sb)
fs/jfs/super.c
185
static void jfs_put_super(struct super_block *sb)
fs/jfs/super.c
187
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/super.c
192
jfs_quota_off_umount(sb);
fs/jfs/super.c
194
rc = jfs_umount(sb);
fs/jfs/super.c
364
struct super_block *sb = fc->root->d_sb;
fs/jfs/super.c
370
sync_filesystem(sb);
fs/jfs/super.c
373
JFS_SBI(sb)->flag = ctx->flag;
fs/jfs/super.c
374
JFS_SBI(sb)->uid = ctx->uid;
fs/jfs/super.c
375
JFS_SBI(sb)->gid = ctx->gid;
fs/jfs/super.c
376
JFS_SBI(sb)->umask = ctx->umask;
fs/jfs/super.c
377
JFS_SBI(sb)->minblks_trim = ctx->minblks_trim;
fs/jfs/super.c
379
unload_nls(JFS_SBI(sb)->nls_tab);
fs/jfs/super.c
380
JFS_SBI(sb)->nls_tab = ctx->nls_map;
fs/jfs/super.c
385
if (sb_rdonly(sb)) {
fs/jfs/super.c
391
ctx->newLVSize = sb_bdev_nr_blocks(sb);
fs/jfs/super.c
396
rc = jfs_extendfs(sb, ctx->newLVSize, 0);
fs/jfs/super.c
401
if (sb_rdonly(sb) && !readonly) {
fs/jfs/super.c
406
truncate_inode_pages(JFS_SBI(sb)->direct_inode->i_mapping, 0);
fs/jfs/super.c
408
JFS_SBI(sb)->flag = flag;
fs/jfs/super.c
409
ret = jfs_mount_rw(sb, 1);
fs/jfs/super.c
412
sb->s_flags &= ~SB_RDONLY;
fs/jfs/super.c
414
dquot_resume(sb, -1);
fs/jfs/super.c
417
if (!sb_rdonly(sb) && readonly) {
fs/jfs/super.c
418
rc = dquot_suspend(sb, -1);
fs/jfs/super.c
421
rc = jfs_umount_rw(sb);
fs/jfs/super.c
422
JFS_SBI(sb)->flag = flag;
fs/jfs/super.c
425
if ((JFS_SBI(sb)->flag & JFS_NOINTEGRITY) != (flag & JFS_NOINTEGRITY)) {
fs/jfs/super.c
426
if (!sb_rdonly(sb)) {
fs/jfs/super.c
427
rc = jfs_umount_rw(sb);
fs/jfs/super.c
431
JFS_SBI(sb)->flag = flag;
fs/jfs/super.c
432
ret = jfs_mount_rw(sb, 1);
fs/jfs/super.c
436
JFS_SBI(sb)->flag = flag;
fs/jfs/super.c
441
static int jfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/jfs/super.c
450
jfs_info("In jfs_read_super: s_flags=0x%lx", sb->s_flags);
fs/jfs/super.c
456
sb->s_fs_info = sbi;
fs/jfs/super.c
457
sb->s_max_links = JFS_LINK_MAX;
fs/jfs/super.c
458
sb->s_time_min = 0;
fs/jfs/super.c
459
sb->s_time_max = U32_MAX;
fs/jfs/super.c
460
sbi->sb = sb;
fs/jfs/super.c
474
if (!bdev_max_discard_sectors(sb->s_bdev)) {
fs/jfs/super.c
483
sb->s_flags |= SB_POSIXACL;
fs/jfs/super.c
494
sb_set_blocksize(sb, PSIZE);
fs/jfs/super.c
499
sb->s_op = &jfs_super_operations;
fs/jfs/super.c
500
sb->s_export_op = &jfs_export_operations;
fs/jfs/super.c
501
sb->s_xattr = jfs_xattr_handlers;
fs/jfs/super.c
503
sb->dq_op = &dquot_operations;
fs/jfs/super.c
504
sb->s_qcop = &jfs_quotactl_ops;
fs/jfs/super.c
505
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
fs/jfs/super.c
511
inode = new_inode(sb);
fs/jfs/super.c
516
inode->i_size = bdev_nr_bytes(sb->s_bdev);
fs/jfs/super.c
523
rc = jfs_mount(sb);
fs/jfs/super.c
529
if (sb_rdonly(sb))
fs/jfs/super.c
532
rc = jfs_mount_rw(sb, 0);
fs/jfs/super.c
542
sb->s_magic = JFS_SUPER_MAGIC;
fs/jfs/super.c
545
set_default_d_op(sb, &jfs_ci_dentry_operations);
fs/jfs/super.c
547
inode = jfs_iget(sb, ROOT_I);
fs/jfs/super.c
552
sb->s_root = d_make_root(inode);
fs/jfs/super.c
553
if (!sb->s_root)
fs/jfs/super.c
559
sb->s_maxbytes = min(((loff_t)sb->s_blocksize) << 40, MAX_LFS_FILESIZE);
fs/jfs/super.c
560
sb->s_time_gran = 1;
fs/jfs/super.c
567
rc = jfs_umount(sb);
fs/jfs/super.c
582
static int jfs_freeze(struct super_block *sb)
fs/jfs/super.c
584
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/super.c
588
if (!sb_rdonly(sb)) {
fs/jfs/super.c
589
txQuiesce(sb);
fs/jfs/super.c
592
jfs_error(sb, "lmLogShutdown failed\n");
fs/jfs/super.c
595
txResume(sb);
fs/jfs/super.c
599
rc = updateSuper(sb, FM_CLEAN);
fs/jfs/super.c
612
static int jfs_unfreeze(struct super_block *sb)
fs/jfs/super.c
614
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/super.c
618
if (!sb_rdonly(sb)) {
fs/jfs/super.c
619
rc = updateSuper(sb, FM_MOUNT);
fs/jfs/super.c
62
static void jfs_handle_error(struct super_block *sb)
fs/jfs/super.c
621
jfs_error(sb, "updateSuper failed\n");
fs/jfs/super.c
626
jfs_error(sb, "lmLogInit failed\n");
fs/jfs/super.c
628
txResume(sb);
fs/jfs/super.c
638
static int jfs_sync_fs(struct super_block *sb, int wait)
fs/jfs/super.c
64
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/super.c
640
struct jfs_log *log = JFS_SBI(sb)->log;
fs/jfs/super.c
648
dquot_writeback_dquots(sb, -1);
fs/jfs/super.c
66
if (sb_rdonly(sb))
fs/jfs/super.c
69
updateSuper(sb, FM_DIRTY);
fs/jfs/super.c
694
static ssize_t jfs_quota_read(struct super_block *sb, int type, char *data,
fs/jfs/super.c
697
struct inode *inode = sb_dqopt(sb)->files[type];
fs/jfs/super.c
698
sector_t blk = off >> sb->s_blocksize_bits;
fs/jfs/super.c
700
int offset = off & (sb->s_blocksize - 1);
fs/jfs/super.c
713
tocopy = min_t(size_t, sb->s_blocksize - offset, toread);
fs/jfs/super.c
723
bh = sb_bread(sb, tmp_bh.b_blocknr);
fs/jfs/super.c
73
sb->s_id);
fs/jfs/super.c
738
static ssize_t jfs_quota_write(struct super_block *sb, int type,
fs/jfs/super.c
741
struct inode *inode = sb_dqopt(sb)->files[type];
fs/jfs/super.c
742
sector_t blk = off >> sb->s_blocksize_bits;
fs/jfs/super.c
744
int offset = off & (sb->s_blocksize - 1);
fs/jfs/super.c
752
tocopy = min_t(size_t, sb->s_blocksize - offset, towrite);
fs/jfs/super.c
759
if (offset || tocopy != sb->s_blocksize)
fs/jfs/super.c
76
sb->s_id);
fs/jfs/super.c
760
bh = sb_bread(sb, tmp_bh.b_blocknr);
fs/jfs/super.c
762
bh = sb_getblk(sb, tmp_bh.b_blocknr);
fs/jfs/super.c
77
sb->s_flags |= SB_RDONLY;
fs/jfs/super.c
797
static int jfs_quota_on(struct super_block *sb, int type, int format_id,
fs/jfs/super.c
803
err = dquot_quota_on(sb, type, format_id, path);
fs/jfs/super.c
818
static int jfs_quota_off(struct super_block *sb, int type)
fs/jfs/super.c
820
struct inode *inode = sb_dqopt(sb)->files[type];
fs/jfs/super.c
826
err = dquot_quota_off(sb, type);
fs/jfs/super.c
83
void jfs_error(struct super_block *sb, const char *fmt, ...)
fs/jfs/super.c
839
return dquot_quota_off(sb, type);
fs/jfs/super.c
872
struct super_block *sb = fc->root->d_sb;
fs/jfs/super.c
875
ctx->uid = JFS_SBI(sb)->uid;
fs/jfs/super.c
876
ctx->gid = JFS_SBI(sb)->gid;
fs/jfs/super.c
877
ctx->umask = JFS_SBI(sb)->umask;
fs/jfs/super.c
879
ctx->minblks_trim = JFS_SBI(sb)->minblks_trim;
fs/jfs/super.c
880
ctx->flag = JFS_SBI(sb)->flag;
fs/jfs/super.c
94
sb->s_id, __builtin_return_address(0), &vaf);
fs/jfs/super.c
98
jfs_handle_error(sb);
fs/jfs/xattr.c
204
struct super_block *sb = ip->i_sb;
fs/jfs/xattr.c
206
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/xattr.c
225
nblocks = (size + (sb->s_blocksize - 1)) >> sb->s_blocksize_bits;
fs/jfs/xattr.c
253
((((nb + sb->s_blocksize - 1)) >> sb->s_blocksize_bits))
fs/jfs/xattr.c
254
<< sb->s_blocksize_bits;
fs/jfs/xattr.c
353
struct super_block *sb = ip->i_sb;
fs/jfs/xattr.c
355
struct jfs_sb_info *sbi = JFS_SBI(sb);
fs/jfs/xattr.c
370
jfs_error(sb, "nbytes is 0\n");
fs/jfs/xattr.c
393
((((nb + sb->s_blocksize - 1)) >> sb->s_blocksize_bits))
fs/jfs/xattr.c
394
<< sb->s_blocksize_bits;
fs/jfs/xattr.c
429
struct super_block *sb = inode->i_sb;
fs/jfs/xattr.c
472
jfs_error(sb, "invalid ea.flag\n");
fs/jfs/xattr.c
475
current_blocks = (ea_size + sb->s_blocksize - 1) >>
fs/jfs/xattr.c
476
sb->s_blocksize_bits;
fs/jfs/xattr.c
486
ea_buf->max_size = (size + sb->s_blocksize - 1) &
fs/jfs/xattr.c
487
~(sb->s_blocksize - 1);
fs/jfs/xattr.c
505
blocks_needed = (min_size + sb->s_blocksize - 1) >>
fs/jfs/xattr.c
506
sb->s_blocksize_bits;
fs/jfs/xattr.c
529
blocks_needed << sb->s_blocksize_bits,
fs/jfs/xattr.c
537
ea_buf->max_size = (min_size + sb->s_blocksize - 1) &
fs/jfs/xattr.c
538
~(sb->s_blocksize - 1);
fs/jfs/xattr.c
550
lengthDXD(&ji->ea) << sb->s_blocksize_bits,
fs/jfs/xattr.c
557
ea_buf->max_size = (ea_size + sb->s_blocksize - 1) &
fs/jfs/xattr.c
558
~(sb->s_blocksize - 1);
fs/kernfs/file.c
947
inode = ilookup(info->sb, kernfs_ino(kn));
fs/kernfs/file.c
955
p_inode = ilookup(info->sb, kernfs_ino(parent));
fs/kernfs/inode.c
249
struct inode *kernfs_get_inode(struct super_block *sb, struct kernfs_node *kn)
fs/kernfs/inode.c
253
inode = iget_locked(sb, kernfs_ino(kn));
fs/kernfs/kernfs-internal.h
86
struct super_block *sb;
fs/kernfs/mount.c
101
struct kernfs_super_info *info = kernfs_info(sb);
fs/kernfs/mount.c
141
inode = kernfs_get_inode(sb, kn);
fs/kernfs/mount.c
146
static struct dentry *kernfs_fh_to_dentry(struct super_block *sb,
fs/kernfs/mount.c
150
return __kernfs_fh_to_dentry(sb, fid, fh_len, fh_type, false);
fs/kernfs/mount.c
153
static struct dentry *kernfs_fh_to_parent(struct super_block *sb,
fs/kernfs/mount.c
157
return __kernfs_fh_to_dentry(sb, fid, fh_len, fh_type, true);
fs/kernfs/mount.c
183
struct kernfs_root *kernfs_root_from_sb(struct super_block *sb)
fs/kernfs/mount.c
185
if (sb->s_op == &kernfs_sops)
fs/kernfs/mount.c
186
return kernfs_info(sb)->root;
fs/kernfs/mount.c
223
struct super_block *sb)
fs/kernfs/mount.c
229
BUG_ON(sb->s_op != &kernfs_sops);
fs/kernfs/mount.c
231
dentry = dget(sb->s_root);
fs/kernfs/mount.c
283
static int kernfs_fill_super(struct super_block *sb, struct kernfs_fs_context *kfc)
fs/kernfs/mount.c
285
struct kernfs_super_info *info = kernfs_info(sb);
fs/kernfs/mount.c
290
info->sb = sb;
fs/kernfs/mount.c
292
sb->s_iflags |= SB_I_NOEXEC | SB_I_NODEV;
fs/kernfs/mount.c
293
sb->s_blocksize = PAGE_SIZE;
fs/kernfs/mount.c
294
sb->s_blocksize_bits = PAGE_SHIFT;
fs/kernfs/mount.c
295
sb->s_magic = kfc->magic;
fs/kernfs/mount.c
296
sb->s_op = &kernfs_sops;
fs/kernfs/mount.c
297
sb->s_xattr = kernfs_xattr_handlers;
fs/kernfs/mount.c
299
sb->s_export_op = &kernfs_export_ops;
fs/kernfs/mount.c
300
sb->s_time_gran = 1;
fs/kernfs/mount.c
301
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/kernfs/mount.c
304
sb->s_shrink->seeks = 0;
fs/kernfs/mount.c
308
inode = kernfs_get_inode(sb, info->root->kn);
fs/kernfs/mount.c
321
sb->s_root = root;
fs/kernfs/mount.c
322
set_default_d_op(sb, &kernfs_dops);
fs/kernfs/mount.c
326
static int kernfs_test_super(struct super_block *sb, struct fs_context *fc)
fs/kernfs/mount.c
328
struct kernfs_super_info *sb_info = kernfs_info(sb);
fs/kernfs/mount.c
334
static int kernfs_set_super(struct super_block *sb, struct fs_context *fc)
fs/kernfs/mount.c
339
return set_anon_super_fc(sb, fc);
fs/kernfs/mount.c
348
const void *kernfs_super_ns(struct super_block *sb)
fs/kernfs/mount.c
350
struct kernfs_super_info *info = kernfs_info(sb);
fs/kernfs/mount.c
369
struct super_block *sb;
fs/kernfs/mount.c
382
sb = sget_fc(fc, kernfs_test_super, kernfs_set_super);
fs/kernfs/mount.c
383
if (IS_ERR(sb))
fs/kernfs/mount.c
384
return PTR_ERR(sb);
fs/kernfs/mount.c
386
if (!sb->s_root) {
fs/kernfs/mount.c
387
struct kernfs_super_info *info = kernfs_info(sb);
fs/kernfs/mount.c
392
error = kernfs_fill_super(sb, kfc);
fs/kernfs/mount.c
394
deactivate_locked_super(sb);
fs/kernfs/mount.c
397
sb->s_flags |= SB_ACTIVE;
fs/kernfs/mount.c
401
super_set_uuid(sb, uuid.b, sizeof(uuid));
fs/kernfs/mount.c
408
fc->root = dget(sb->s_root);
fs/kernfs/mount.c
427
void kernfs_kill_sb(struct super_block *sb)
fs/kernfs/mount.c
429
struct kernfs_super_info *info = kernfs_info(sb);
fs/kernfs/mount.c
440
kill_anon_super(sb);
fs/kernfs/mount.c
97
static struct dentry *__kernfs_fh_to_dentry(struct super_block *sb,
fs/libfs.c
1488
struct dentry *generic_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/libfs.c
1490
(struct super_block *sb, u64 ino, u32 gen))
fs/libfs.c
1500
inode = get_inode(sb, fid->i32.ino, fid->i32.gen);
fs/libfs.c
1521
struct dentry *generic_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/libfs.c
1523
(struct super_block *sb, u64 ino, u32 gen))
fs/libfs.c
1532
inode = get_inode(sb, fid->i32.parent_ino,
fs/libfs.c
1862
struct super_block *sb = dentry->d_sb;
fs/libfs.c
1863
const struct unicode_map *um = sb->s_encoding;
fs/libfs.c
1870
if (ret < 0 && sb_has_strict_encoding(sb))
fs/libfs.c
1907
const struct super_block *sb = parent->i_sb;
fs/libfs.c
1908
const struct unicode_map *um = sb->s_encoding;
fs/libfs.c
1950
if (res < 0 && sb_has_strict_encoding(sb)) {
fs/libfs.c
1976
void generic_set_sb_d_ops(struct super_block *sb)
fs/libfs.c
1979
if (sb->s_encoding) {
fs/libfs.c
1980
set_default_d_op(sb, &generic_ci_dentry_ops);
fs/libfs.c
1985
if (sb->s_cop) {
fs/libfs.c
1986
set_default_d_op(sb, &generic_encrypted_dentry_ops);
fs/libfs.c
2157
struct super_block *sb,
fs/libfs.c
2162
const struct stashed_operations *sops = sb->s_fs_info;
fs/libfs.c
2165
inode = new_inode_pseudo(sb);
fs/libfs.c
2184
dentry = d_alloc_anon(sb);
fs/lockd/svcsubs.c
463
struct super_block *sb = datap;
fs/lockd/svcsubs.c
465
return sb == nlmsvc_file_inode(file)->i_sb;
fs/lockd/svcsubs.c
475
nlmsvc_unlock_all_by_sb(struct super_block *sb)
fs/lockd/svcsubs.c
479
ret = nlm_traverse_files(sb, nlmsvc_always_match, nlmsvc_match_sb);
fs/minix/bitmap.c
102
return (count_free(sbi->s_zmap, sb->s_blocksize, bits)
fs/minix/bitmap.c
107
minix_V1_raw_inode(struct super_block *sb, ino_t ino, struct buffer_head **bh)
fs/minix/bitmap.c
110
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/bitmap.c
115
sb->s_id, (long)ino);
fs/minix/bitmap.c
121
*bh = sb_bread(sb, block);
fs/minix/bitmap.c
131
minix_V2_raw_inode(struct super_block *sb, ino_t ino, struct buffer_head **bh)
fs/minix/bitmap.c
134
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/bitmap.c
136
int minix2_inodes_per_block = sb->s_blocksize / sizeof(struct minix2_inode);
fs/minix/bitmap.c
141
sb->s_id, (long)ino);
fs/minix/bitmap.c
147
*bh = sb_bread(sb, block);
fs/minix/bitmap.c
185
struct super_block *sb = inode->i_sb;
fs/minix/bitmap.c
188
int k = sb->s_blocksize_bits + 3;
fs/minix/bitmap.c
215
struct super_block *sb = dir->i_sb;
fs/minix/bitmap.c
216
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/bitmap.c
217
struct inode *inode = new_inode(sb);
fs/minix/bitmap.c
219
int bits_per_zone = 8 * sb->s_blocksize;
fs/minix/bitmap.c
263
unsigned long minix_count_free_inodes(struct super_block *sb)
fs/minix/bitmap.c
265
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/bitmap.c
268
return count_free(sbi->s_imap, sb->s_blocksize, bits);
fs/minix/bitmap.c
44
struct super_block *sb = inode->i_sb;
fs/minix/bitmap.c
45
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/bitmap.c
47
int k = sb->s_blocksize_bits + 3;
fs/minix/bitmap.c
65
sb->s_id, block);
fs/minix/bitmap.c
97
unsigned long minix_count_free_blocks(struct super_block *sb)
fs/minix/bitmap.c
99
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/dir.c
159
struct super_block * sb = dir->i_sb;
fs/minix/dir.c
160
struct minix_sb_info * sbi = minix_sb(sb);
fs/minix/dir.c
204
struct super_block * sb = dir->i_sb;
fs/minix/dir.c
205
struct minix_sb_info * sbi = minix_sb(sb);
fs/minix/dir.c
86
struct super_block *sb = inode->i_sb;
fs/minix/dir.c
87
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/inode.c
135
struct super_block *sb = fc->root->d_sb;
fs/minix/inode.c
136
struct minix_sb_info * sbi = sb->s_fs_info;
fs/minix/inode.c
138
sync_filesystem(sb);
fs/minix/inode.c
140
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb))
fs/minix/inode.c
170
static bool minix_check_superblock(struct super_block *sb)
fs/minix/inode.c
172
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/inode.c
189
block = minix_blocks_needed(sbi->s_ninodes, sb->s_blocksize);
fs/minix/inode.c
198
sb->s_blocksize);
fs/minix/inode.c
211
sb->s_maxbytes > (7 + 512 + 512*512) * BLOCK_SIZE)
fs/minix/inode.c
418
struct super_block *sb = dentry->d_sb;
fs/minix/inode.c
419
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/inode.c
420
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/minix/inode.c
421
buf->f_type = sb->s_magic;
fs/minix/inode.c
422
buf->f_bsize = sb->s_blocksize;
fs/minix/inode.c
424
buf->f_bfree = minix_count_free_blocks(sb);
fs/minix/inode.c
427
buf->f_ffree = minix_count_free_inodes(sb);
fs/minix/inode.c
58
static void minix_put_super(struct super_block *sb)
fs/minix/inode.c
609
struct inode *minix_iget(struct super_block *sb, unsigned long ino)
fs/minix/inode.c
61
struct minix_sb_info *sbi = minix_sb(sb);
fs/minix/inode.c
613
inode = iget_locked(sb, ino);
fs/minix/inode.c
63
if (!sb_rdonly(sb)) {
fs/minix/inode.c
707
struct super_block *sb = path->dentry->d_sb;
fs/minix/inode.c
712
stat->blocks = (BLOCK_SIZE / 512) * V1_minix_blocks(stat->size, sb);
fs/minix/inode.c
714
stat->blocks = (sb->s_blocksize / 512) * V2_minix_blocks(stat->size, sb);
fs/minix/inode.c
715
stat->blksize = sb->s_blocksize;
fs/minix/inode.c
74
sb->s_fs_info = NULL;
fs/minix/inode.c
80
static struct inode *minix_alloc_inode(struct super_block *sb)
fs/minix/inode.c
83
ei = alloc_inode_sb(sb, minix_inode_cachep, GFP_KERNEL);
fs/minix/itree_common.c
302
struct super_block *sb = inode->i_sb;
fs/minix/itree_common.c
312
iblock = (inode->i_size + sb->s_blocksize -1) >> sb->s_blocksize_bits;
fs/minix/itree_common.c
357
static inline unsigned nblocks(loff_t size, struct super_block *sb)
fs/minix/itree_common.c
359
int k = sb->s_blocksize_bits - 10;
fs/minix/itree_common.c
36
struct super_block *sb = inode->i_sb;
fs/minix/itree_common.c
361
blocks = (size + sb->s_blocksize - 1) >> (BLOCK_SIZE_BITS + k);
fs/minix/itree_common.c
365
blocks += sb->s_blocksize/sizeof(block_t) - 1;
fs/minix/itree_common.c
366
blocks /= sb->s_blocksize/sizeof(block_t);
fs/minix/itree_common.c
46
bh = sb_bread(sb, block_to_cpu(p->key));
fs/minix/itree_v1.c
64
unsigned V1_minix_blocks(loff_t size, struct super_block *sb)
fs/minix/itree_v1.c
66
return nblocks(size, sb);
fs/minix/itree_v2.c
25
#define INDIRCOUNT(sb) (1 << ((sb)->s_blocksize_bits - 2))
fs/minix/itree_v2.c
30
struct super_block *sb = inode->i_sb;
fs/minix/itree_v2.c
34
block, sb->s_bdev);
fs/minix/itree_v2.c
37
if ((u64)block * (u64)sb->s_blocksize >= sb->s_maxbytes)
fs/minix/itree_v2.c
42
} else if ((block -= DIRCOUNT) < INDIRCOUNT(sb)) {
fs/minix/itree_v2.c
45
} else if ((block -= INDIRCOUNT(sb)) < INDIRCOUNT(sb) * INDIRCOUNT(sb)) {
fs/minix/itree_v2.c
47
offsets[n++] = block / INDIRCOUNT(sb);
fs/minix/itree_v2.c
48
offsets[n++] = block % INDIRCOUNT(sb);
fs/minix/itree_v2.c
50
block -= INDIRCOUNT(sb) * INDIRCOUNT(sb);
fs/minix/itree_v2.c
52
offsets[n++] = (block / INDIRCOUNT(sb)) / INDIRCOUNT(sb);
fs/minix/itree_v2.c
53
offsets[n++] = (block / INDIRCOUNT(sb)) % INDIRCOUNT(sb);
fs/minix/itree_v2.c
54
offsets[n++] = block % INDIRCOUNT(sb);
fs/minix/itree_v2.c
72
unsigned V2_minix_blocks(loff_t size, struct super_block *sb)
fs/minix/itree_v2.c
74
return nblocks(size, sb);
fs/minix/minix.h
53
unsigned long minix_count_free_inodes(struct super_block *sb);
fs/minix/minix.h
56
unsigned long minix_count_free_blocks(struct super_block *sb);
fs/minix/minix.h
85
static inline struct minix_sb_info *minix_sb(struct super_block *sb)
fs/minix/minix.h
87
return sb->s_fs_info;
fs/namei.c
599
static int sb_permission(struct super_block *sb, struct inode *inode, int mask)
fs/namei.c
605
if (sb_rdonly(sb) && (S_ISREG(mode) || S_ISDIR(mode) || S_ISLNK(mode)))
fs/namei.c
812
struct super_block *sb = mnt->mnt_sb;
fs/namei.c
815
if (mnt->mnt_root == sb->s_root)
fs/namespace.c
1842
static int do_umount_root(struct super_block *sb)
fs/namespace.c
1846
down_write(&sb->s_umount);
fs/namespace.c
1847
if (!sb_rdonly(sb)) {
fs/namespace.c
1850
fc = fs_context_for_reconfigure(sb->s_root, SB_RDONLY,
fs/namespace.c
1861
up_write(&sb->s_umount);
fs/namespace.c
1867
struct super_block *sb = mnt->mnt.mnt_sb;
fs/namespace.c
1910
if (flags & MNT_FORCE && sb->s_op->umount_begin) {
fs/namespace.c
1911
sb->s_op->umount_begin(sb);
fs/namespace.c
1928
if (!ns_capable(sb->s_user_ns, CAP_SYS_ADMIN))
fs/namespace.c
1930
return do_umount_root(sb);
fs/namespace.c
2019
struct super_block *sb = path->dentry->d_sb;
fs/namespace.c
2029
if (flags & MNT_FORCE && !ns_capable(sb->s_user_ns, CAP_SYS_ADMIN))
fs/namespace.c
3290
struct super_block *sb = mnt->mnt_sb;
fs/namespace.c
3293
(!(sb->s_iflags & SB_I_TS_EXPIRY_WARNED)) &&
fs/namespace.c
3294
(ktime_get_real_seconds() + TIME_UPTIME_SEC_MAX > sb->s_time_max)) {
fs/namespace.c
3306
sb->s_type->name,
fs/namespace.c
3308
mntpath, &sb->s_time_max,
fs/namespace.c
3309
(unsigned long long)sb->s_time_max);
fs/namespace.c
3311
sb->s_iflags |= SB_I_TS_EXPIRY_WARNED;
fs/namespace.c
3324
struct super_block *sb = path->mnt->mnt_sb;
fs/namespace.c
3341
down_read(&sb->s_umount);
fs/namespace.c
3347
up_read(&sb->s_umount);
fs/namespace.c
3363
struct super_block *sb = path->mnt->mnt_sb;
fs/namespace.c
3388
down_write(&sb->s_umount);
fs/namespace.c
3390
if (ns_capable(sb->s_user_ns, CAP_SYS_ADMIN)) {
fs/namespace.c
3398
up_write(&sb->s_umount);
fs/namespace.c
3753
static bool mount_too_revealing(const struct super_block *sb, int *new_mnt_flags);
fs/namespace.c
3762
struct super_block *sb;
fs/namespace.c
3769
sb = fc->root->d_sb;
fs/namespace.c
3770
error = security_sb_kern_mount(sb);
fs/namespace.c
3774
if (unlikely(mount_too_revealing(sb, &mnt_flags))) {
fs/namespace.c
5265
struct super_block *sb = s->mnt->mnt_sb;
fs/namespace.c
5268
s->sm.sb_dev_major = MAJOR(sb->s_dev);
fs/namespace.c
5269
s->sm.sb_dev_minor = MINOR(sb->s_dev);
fs/namespace.c
5270
s->sm.sb_magic = sb->s_magic;
fs/namespace.c
5271
s->sm.sb_flags = sb->s_flags & (SB_RDONLY|SB_SYNCHRONOUS|SB_DIRSYNC|SB_LAZYTIME);
fs/namespace.c
5332
struct super_block *sb = s->mnt->mnt_sb;
fs/namespace.c
5334
seq_puts(seq, sb->s_type->name);
fs/namespace.c
5340
struct super_block *sb = s->mnt->mnt_sb;
fs/namespace.c
5342
if (sb->s_subtype)
fs/namespace.c
5343
seq_puts(seq, sb->s_subtype);
fs/namespace.c
5348
struct super_block *sb = s->mnt->mnt_sb;
fs/namespace.c
5351
if (sb->s_op->show_devname) {
fs/namespace.c
5355
ret = sb->s_op->show_devname(seq, s->mnt->mnt_root);
fs/namespace.c
5381
struct super_block *sb = mnt->mnt_sb;
fs/namespace.c
5385
err = security_sb_show_options(seq, sb);
fs/namespace.c
5389
if (sb->s_op->show_options) {
fs/namespace.c
5390
err = sb->s_op->show_options(seq, mnt->mnt_root);
fs/namespace.c
5441
struct super_block *sb = mnt->mnt_sb;
fs/namespace.c
5445
if (!sb->s_op->show_options)
fs/namespace.c
5448
err = sb->s_op->show_options(seq, mnt->mnt_root);
fs/namespace.c
5463
struct super_block *sb = mnt->mnt_sb;
fs/namespace.c
5467
err = security_sb_show_options(seq, sb);
fs/namespace.c
6309
const struct super_block *sb,
fs/namespace.c
6320
if (mnt->mnt.mnt_sb->s_type != sb->s_type)
fs/namespace.c
6368
static bool mount_too_revealing(const struct super_block *sb, int *new_mnt_flags)
fs/namespace.c
6378
s_iflags = sb->s_iflags;
fs/namespace.c
6388
return !mnt_already_visible(ns, sb, new_mnt_flags);
fs/namespace.c
694
int sb_prepare_remount_readonly(struct super_block *sb)
fs/namespace.c
699
if (atomic_long_read(&sb->s_remove_count))
fs/namespace.c
704
for (struct mount *m = sb->s_mounts; m; m = m->mnt_next_for_sb) {
fs/namespace.c
711
if (!err && atomic_long_read(&sb->s_remove_count))
fs/namespace.c
715
sb_start_ro_state_change(sb);
fs/namespace.c
716
for (struct mount *m = sb->s_mounts; m; m = m->mnt_next_for_sb) {
fs/nfs/blocklayout/rpc_pipefs.c
144
static int nfs4blocklayout_register_sb(struct super_block *sb,
fs/nfs/blocklayout/rpc_pipefs.c
150
dir = rpc_d_lookup_sb(sb, NFS_PIPE_DIRNAME);
fs/nfs/blocklayout/rpc_pipefs.c
161
struct super_block *sb = ptr;
fs/nfs/blocklayout/rpc_pipefs.c
162
struct net *net = sb->s_fs_info;
fs/nfs/blocklayout/rpc_pipefs.c
176
ret = nfs4blocklayout_register_sb(sb, nn->bl_device_pipe);
fs/nfs/cache_lib.c
116
int nfs_cache_register_sb(struct super_block *sb, struct cache_detail *cd)
fs/nfs/cache_lib.c
121
dir = rpc_d_lookup_sb(sb, "cache");
fs/nfs/cache_lib.c
143
void nfs_cache_unregister_sb(struct super_block *sb, struct cache_detail *cd)
fs/nfs/cache_lib.h
29
extern int nfs_cache_register_sb(struct super_block *sb,
fs/nfs/cache_lib.h
31
extern void nfs_cache_unregister_sb(struct super_block *sb,
fs/nfs/dns_resolve.c
427
struct super_block *sb = ptr;
fs/nfs/dns_resolve.c
428
struct net *net = sb->s_fs_info;
fs/nfs/dns_resolve.c
441
ret = nfs_cache_register_sb(sb, cd);
fs/nfs/dns_resolve.c
444
nfs_cache_unregister_sb(sb, cd);
fs/nfs/export.c
100
inode = nfs_ilookup(sb, fattr, server_fh);
fs/nfs/export.c
104
rpc_ops = NFS_SB(sb)->nfs_client->rpc_ops;
fs/nfs/export.c
105
ret = rpc_ops->getattr(NFS_SB(sb), server_fh, fattr, NULL);
fs/nfs/export.c
108
trace_nfs_fh_to_dentry(sb, server_fh, fattr->fileid, ret);
fs/nfs/export.c
113
inode = nfs_fhget(sb, server_fh, fattr);
fs/nfs/export.c
128
struct super_block *sb = inode->i_sb;
fs/nfs/export.c
129
struct nfs_server *server = NFS_SB(sb);
fs/nfs/export.c
148
pinode = nfs_fhget(sb, &fh, fattr);
fs/nfs/export.c
64
nfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/nfs/export.c
88
fattr = nfs_alloc_fattr_with_label(NFS_SB(sb));
fs/nfs/fscache.c
109
!nfs_append_int(key, &len, sb->s_flags & NFS_SB_MASK) ||
fs/nfs/fscache.c
151
void nfs_fscache_release_super_cookie(struct super_block *sb)
fs/nfs/fscache.c
153
struct nfs_server *nfss = NFS_SB(sb);
fs/nfs/fscache.c
88
int nfs_fscache_get_super_cookie(struct super_block *sb, const char *uniq, int ulen)
fs/nfs/fscache.c
91
struct nfs_server *nfss = NFS_SB(sb);
fs/nfs/fscache.h
170
static inline void nfs_fscache_release_super_cookie(struct super_block *sb) {}
fs/nfs/getroot.c
129
if (ctx->clone_data.sb) {
fs/nfs/getroot.c
135
error = security_sb_clone_mnt_opts(ctx->clone_data.sb,
fs/nfs/getroot.c
139
clone_server = NFS_SB(ctx->clone_data.sb);
fs/nfs/getroot.c
39
static int nfs_superblock_set_dummy_root(struct super_block *sb, struct inode *inode)
fs/nfs/getroot.c
42
if (sb->s_root == NULL) {
fs/nfs/getroot.c
43
sb->s_root = d_make_root(inode);
fs/nfs/getroot.c
44
if (sb->s_root == NULL)
fs/nfs/getroot.c
55
spin_lock(&d_inode(sb->s_root)->i_lock);
fs/nfs/getroot.c
56
spin_lock(&sb->s_root->d_lock);
fs/nfs/getroot.c
57
hlist_del_init(&sb->s_root->d_u.d_alias);
fs/nfs/getroot.c
58
spin_unlock(&sb->s_root->d_lock);
fs/nfs/getroot.c
59
spin_unlock(&d_inode(sb->s_root)->i_lock);
fs/nfs/inode.c
1248
struct super_block *sb = ctx->dentry->d_sb;
fs/nfs/inode.c
1261
nfs_sb_deactive(sb);
fs/nfs/inode.c
2520
struct inode *nfs_alloc_inode(struct super_block *sb)
fs/nfs/inode.c
2523
nfsi = alloc_inode_sb(sb, nfs_inode_cachep, GFP_KERNEL);
fs/nfs/inode.c
409
nfs_ilookup(struct super_block *sb, struct nfs_fattr *fattr, struct nfs_fh *fh)
fs/nfs/inode.c
423
inode = ilookup5(sb, hash, nfs_find_actor, &desc);
fs/nfs/inode.c
451
nfs_fhget(struct super_block *sb, struct nfs_fh *fh, struct nfs_fattr *fattr)
fs/nfs/inode.c
458
u64 fattr_supported = NFS_SB(sb)->fattr_valid;
fs/nfs/inode.c
461
nfs_attr_check_mountpoint(sb, fattr);
fs/nfs/inode.c
472
inode = iget5_locked(sb, hash, nfs_find_actor, nfs_init_locked, &desc);
fs/nfs/inode.c
496
inode->i_op = NFS_SB(sb)->nfs_client->rpc_ops->file_inode_ops;
fs/nfs/inode.c
498
inode->i_fop = NFS_SB(sb)->nfs_client->rpc_ops->file_ops;
fs/nfs/inode.c
503
inode->i_op = NFS_SB(sb)->nfs_client->rpc_ops->dir_inode_ops;
fs/nfs/internal.h
152
struct super_block *sb;
fs/nfs/internal.h
441
extern struct inode *nfs_alloc_inode(struct super_block *sb);
fs/nfs/internal.h
518
extern bool nfs_sb_active(struct super_block *sb);
fs/nfs/internal.h
519
extern void nfs_sb_deactive(struct super_block *sb);
fs/nfs/internal.h
747
struct super_block *sb = inode->i_sb;
fs/nfs/internal.h
749
if (sb && nfs_sb_active(sb)) {
fs/nfs/internal.h
752
nfs_sb_deactive(sb);
fs/nfs/internal.h
760
struct super_block *sb = inode->i_sb;
fs/nfs/internal.h
763
nfs_sb_deactive(sb);
fs/nfs/internal.h
840
void nfs_super_set_maxbytes(struct super_block *sb, __u64 maxfilesize)
fs/nfs/internal.h
842
sb->s_maxbytes = (loff_t)maxfilesize;
fs/nfs/internal.h
843
if (sb->s_maxbytes > MAX_LFS_FILESIZE || sb->s_maxbytes <= 0)
fs/nfs/internal.h
844
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/nfs/namespace.c
168
ctx->clone_data.sb = path->dentry->d_sb;
fs/nfs/namespace.c
286
server = ctx->nfs_mod->rpc_ops->clone_server(NFS_SB(ctx->clone_data.sb),
fs/nfs/nfs4client.c
1063
struct nfs_server *parent_server = NFS_SB(ctx->clone_data.sb);
fs/nfs/nfs4proc.c
1463
struct super_block *sb = p->dentry->d_sb;
fs/nfs/nfs4proc.c
1477
nfs_sb_deactive(sb);
fs/nfs/nfs4proc.c
3556
struct super_block *sb = calldata->state->inode->i_sb;
fs/nfs/nfs4proc.c
3564
nfs_sb_deactive(sb);
fs/nfs/nfstrace.h
1810
const struct super_block *sb,
fs/nfs/nfstrace.h
1816
TP_ARGS(sb, fh, fileid, error),
fs/nfs/nfstrace.h
1827
__entry->dev = sb->s_dev;
fs/nfs/super.c
1030
struct super_block *sb = fc->root->d_sb;
fs/nfs/super.c
1031
struct nfs_server *nfss = sb->s_fs_info;
fs/nfs/super.c
1034
sync_filesystem(sb);
fs/nfs/super.c
1068
static void nfs_fill_super(struct super_block *sb, struct nfs_fs_context *ctx)
fs/nfs/super.c
1070
struct nfs_server *server = NFS_SB(sb);
fs/nfs/super.c
1072
sb->s_blocksize_bits = 0;
fs/nfs/super.c
1073
sb->s_blocksize = 0;
fs/nfs/super.c
1074
sb->s_xattr = server->nfs_client->cl_nfs_mod->xattr;
fs/nfs/super.c
1075
sb->s_op = server->nfs_client->cl_nfs_mod->sops;
fs/nfs/super.c
1077
sb->s_blocksize =
fs/nfs/super.c
1078
nfs_block_size(server->bsize, &sb->s_blocksize_bits);
fs/nfs/super.c
1082
sb->s_time_gran = 1000;
fs/nfs/super.c
1083
sb->s_time_min = 0;
fs/nfs/super.c
1084
sb->s_time_max = U32_MAX;
fs/nfs/super.c
1091
sb->s_flags |= SB_POSIXACL;
fs/nfs/super.c
1092
sb->s_time_gran = 1;
fs/nfs/super.c
1093
sb->s_time_min = 0;
fs/nfs/super.c
1094
sb->s_time_max = U32_MAX;
fs/nfs/super.c
1095
sb->s_export_op = &nfs_export_ops;
fs/nfs/super.c
1098
sb->s_iflags |= SB_I_NOUMASK;
fs/nfs/super.c
1099
sb->s_time_gran = 1;
fs/nfs/super.c
1100
sb->s_time_min = S64_MIN;
fs/nfs/super.c
1101
sb->s_time_max = S64_MAX;
fs/nfs/super.c
1103
sb->s_export_op = &nfs_export_ops;
fs/nfs/super.c
1107
sb->s_magic = NFS_SUPER_MAGIC;
fs/nfs/super.c
1110
snprintf(sb->s_id, sizeof(sb->s_id),
fs/nfs/super.c
1111
"%u:%u", MAJOR(sb->s_dev), MINOR(sb->s_dev));
fs/nfs/super.c
1113
if (sb->s_blocksize == 0)
fs/nfs/super.c
1114
sb->s_blocksize = nfs_block_bits(server->wsize,
fs/nfs/super.c
1115
&sb->s_blocksize_bits);
fs/nfs/super.c
1117
nfs_super_set_maxbytes(sb, server->maxfilesize);
fs/nfs/super.c
1118
nfs_sysfs_move_server_to_sb(sb);
fs/nfs/super.c
1223
static int nfs_compare_super(struct super_block *sb, struct fs_context *fc)
fs/nfs/super.c
1225
struct nfs_server *server = fc->s_fs_info, *old = NFS_SB(sb);
fs/nfs/super.c
1237
security_sb_mnt_opts_compat(sb, fc->security))
fs/nfs/super.c
1239
return nfs_compare_mount_options(sb, server, fc);
fs/nfs/super.c
1243
static int nfs_get_cache_cookie(struct super_block *sb,
fs/nfs/super.c
1246
struct nfs_server *nfss = NFS_SB(sb);
fs/nfs/super.c
1255
if (ctx->clone_data.sb) {
fs/nfs/super.c
1256
struct nfs_server *mnt_s = NFS_SB(ctx->clone_data.sb);
fs/nfs/super.c
1272
return nfs_fscache_get_super_cookie(sb, uniq, ulen);
fs/nfs/super.c
1275
static int nfs_get_cache_cookie(struct super_block *sb,
fs/nfs/super.c
194
bool nfs_sb_active(struct super_block *sb)
fs/nfs/super.c
196
struct nfs_server *server = NFS_SB(sb);
fs/nfs/super.c
198
if (!atomic_inc_not_zero(&sb->s_active))
fs/nfs/super.c
201
atomic_dec(&sb->s_active);
fs/nfs/super.c
206
void nfs_sb_deactive(struct super_block *sb)
fs/nfs/super.c
208
struct nfs_server *server = NFS_SB(sb);
fs/nfs/super.c
211
deactivate_super(sb);
fs/nfs/super.c
736
void nfs_umount_begin(struct super_block *sb)
fs/nfs/super.c
741
server = NFS_SB(sb);
fs/nfs/unlink.c
290
struct super_block *sb = data->old_dir->i_sb;
fs/nfs/unlink.c
312
nfs_sb_deactive(sb);
fs/nfs/unlink.c
67
struct super_block *sb = dentry->d_sb;
fs/nfs/unlink.c
73
nfs_sb_deactive(sb);
fs/nfsd/blocklayout.c
205
nfsd4_block_get_device_info_simple(struct super_block *sb,
fs/nfsd/blocklayout.c
221
return sb->s_export_op->get_uuid(sb, b->simple.sig, &b->simple.sig_len,
fs/nfsd/blocklayout.c
226
nfsd4_block_proc_getdeviceinfo(struct super_block *sb,
fs/nfsd/blocklayout.c
231
if (bdev_is_partition(sb->s_bdev))
fs/nfsd/blocklayout.c
233
return nfserrno(nfsd4_block_get_device_info_simple(sb, gdp));
fs/nfsd/blocklayout.c
30
struct super_block *sb = inode->i_sb;
fs/nfsd/blocklayout.c
313
nfsd4_block_get_device_info_scsi(struct super_block *sb,
fs/nfsd/blocklayout.c
333
ret = nfsd4_block_get_unique_id(sb->s_bdev->bd_disk, b);
fs/nfsd/blocklayout.c
338
ops = sb->s_bdev->bd_disk->fops->pr_ops;
fs/nfsd/blocklayout.c
341
sb->s_id);
fs/nfsd/blocklayout.c
345
ret = ops->pr_register(sb->s_bdev, 0, NFSD_MDS_PR_KEY, true);
fs/nfsd/blocklayout.c
348
sb->s_id);
fs/nfsd/blocklayout.c
35
error = sb->s_export_op->map_blocks(inode, offset, length, &iomap,
fs/nfsd/blocklayout.c
352
ret = ops->pr_reserve(sb->s_bdev, NFSD_MDS_PR_KEY,
fs/nfsd/blocklayout.c
356
sb->s_id);
fs/nfsd/blocklayout.c
369
nfsd4_scsi_proc_getdeviceinfo(struct super_block *sb,
fs/nfsd/blocklayout.c
374
if (bdev_is_partition(sb->s_bdev))
fs/nfsd/blocklayout.c
376
return nfserrno(nfsd4_block_get_device_info_scsi(sb, clp, gdp));
fs/nfsd/flexfilelayout.c
81
nfsd4_ff_proc_getdeviceinfo(struct super_block *sb, struct svc_rqst *rqstp,
fs/nfsd/nfs3proc.c
682
struct super_block *sb = argp->fh.fh_dentry->d_sb;
fs/nfsd/nfs3proc.c
685
if (sb->s_magic == MSDOS_SUPER_MAGIC) {
fs/nfsd/nfs3proc.c
688
resp->f_maxfilesize = sb->s_maxbytes;
fs/nfsd/nfs3proc.c
719
struct super_block *sb = argp->fh.fh_dentry->d_sb;
fs/nfsd/nfs3proc.c
722
switch (sb->s_magic) {
fs/nfsd/nfs4layouts.c
129
struct super_block *sb = exp->ex_path.mnt->mnt_sb;
fs/nfsd/nfs4layouts.c
139
if (sb->s_export_op->get_uuid &&
fs/nfsd/nfs4layouts.c
140
sb->s_export_op->map_blocks &&
fs/nfsd/nfs4layouts.c
141
sb->s_export_op->commit_blocks)
fs/nfsd/nfs4layouts.c
145
if (sb->s_export_op->map_blocks &&
fs/nfsd/nfs4layouts.c
146
sb->s_export_op->commit_blocks &&
fs/nfsd/nfs4layouts.c
147
sb->s_bdev &&
fs/nfsd/nfs4layouts.c
148
sb->s_bdev->bd_disk->fops->pr_ops &&
fs/nfsd/nfs4layouts.c
149
sb->s_bdev->bd_disk->fops->get_unique_id)
fs/nfsd/nfs4proc.c
1544
const struct super_block *sb)
fs/nfsd/nfs4proc.c
1547
file_inode(copy->nf_src->nf_file)->i_sb == sb)
fs/nfsd/nfs4proc.c
1550
file_inode(copy->nf_dst->nf_file)->i_sb == sb)
fs/nfsd/nfs4proc.c
1560
void nfsd4_cancel_copy_by_sb(struct net *net, struct super_block *sb)
fs/nfsd/nfs4proc.c
1576
if (nfsd4_copy_on_sb(copy, sb)) {
fs/nfsd/nfs4proc.c
1615
extern void nfs_sb_deactive(struct super_block *sb);
fs/nfsd/nfs4recover.c
1995
struct super_block *sb = ptr;
fs/nfsd/nfs4recover.c
1996
struct net *net = sb->s_fs_info;
fs/nfsd/nfs4recover.c
2011
ret = nfsd4_cld_register_sb(sb, cn->cn_pipe);
fs/nfsd/nfs4recover.c
850
nfsd4_cld_register_sb(struct super_block *sb, struct rpc_pipe *pipe)
fs/nfsd/nfs4recover.c
855
dir = rpc_d_lookup_sb(sb, NFSD_PIPE_DIR);
fs/nfsd/nfs4recover.c
866
struct super_block *sb;
fs/nfsd/nfs4recover.c
869
sb = rpc_get_sb_net(net);
fs/nfsd/nfs4recover.c
870
if (!sb)
fs/nfsd/nfs4recover.c
872
err = nfsd4_cld_register_sb(sb, pipe);
fs/nfsd/nfs4recover.c
880
struct super_block *sb;
fs/nfsd/nfs4recover.c
882
sb = rpc_get_sb_net(net);
fs/nfsd/nfs4recover.c
883
if (sb) {
fs/nfsd/nfs4state.c
1742
struct super_block *sb,
fs/nfsd/nfs4state.c
1752
stid->sc_file->fi_inode->i_sb == sb) {
fs/nfsd/nfs4state.c
1774
void nfsd4_revoke_states(struct nfsd_net *nn, struct super_block *sb)
fs/nfsd/nfs4state.c
1787
struct nfs4_stid *stid = find_one_sb_stid(clp, sb,
fs/nfsd/nfs4xdr.c
3397
struct super_block *sb = args->exp->ex_path.mnt->mnt_sb;
fs/nfsd/nfs4xdr.c
3399
return nfsd4_encode_uint64_t(xdr, sb->s_maxbytes);
fs/nfsd/nfsctl.c
1123
static struct inode *nfsd_get_inode(struct super_block *sb, umode_t mode)
fs/nfsd/nfsctl.c
1125
struct inode *inode = new_inode(sb);
fs/nfsd/nfsctl.c
1128
inode->i_ino = iunique(sb, NFSD_MaxReserved);
fs/nfsd/nfsctl.c
1286
static int nfsd_fill_super(struct super_block *sb, struct fs_context *fc)
fs/nfsd/nfsctl.c
1324
ret = simple_fill_super(sb, 0x6e667364, nfsd_files);
fs/nfsd/nfsctl.c
1327
_nfsd_symlink(sb->s_root, "supported_krb5_enctypes",
fs/nfsd/nfsctl.c
1329
dentry = nfsd_mkdir(sb->s_root, NULL, "clients");
fs/nfsd/nfsctl.c
1360
static void nfsd_umount(struct super_block *sb)
fs/nfsd/nfsctl.c
1362
struct net *net = sb->s_fs_info;
fs/nfsd/nfsctl.c
1366
kill_anon_super(sb);
fs/nfsd/pnfs.h
25
__be32 (*proc_getdeviceinfo)(struct super_block *sb,
fs/nfsd/state.h
852
void nfsd4_revoke_states(struct nfsd_net *nn, struct super_block *sb);
fs/nfsd/state.h
853
void nfsd4_cancel_copy_by_sb(struct net *net, struct super_block *sb);
fs/nfsd/state.h
855
static inline void nfsd4_revoke_states(struct nfsd_net *nn, struct super_block *sb)
fs/nfsd/state.h
858
static inline void nfsd4_cancel_copy_by_sb(struct net *net, struct super_block *sb)
fs/nfsd/vfs.c
1432
struct super_block *sb = file_inode(file)->i_sb;
fs/nfsd/vfs.c
1446
if (sb->s_export_op)
fs/nfsd/vfs.c
1447
exp_op_flags = sb->s_export_op->flags;
fs/nilfs2/cpfile.c
1134
int nilfs_cpfile_read(struct super_block *sb, size_t cpsize,
fs/nilfs2/cpfile.c
1140
if (cpsize > sb->s_blocksize) {
fs/nilfs2/cpfile.c
1141
nilfs_err(sb, "too large checkpoint size: %zu bytes", cpsize);
fs/nilfs2/cpfile.c
1144
nilfs_err(sb, "too small checkpoint size: %zu bytes", cpsize);
fs/nilfs2/cpfile.c
1148
cpfile = nilfs_iget_locked(sb, NULL, NILFS_CPFILE_INO);
fs/nilfs2/cpfile.h
33
int nilfs_cpfile_read(struct super_block *sb, size_t cpsize,
fs/nilfs2/dat.c
488
int nilfs_dat_read(struct super_block *sb, size_t entry_size,
fs/nilfs2/dat.c
496
if (entry_size > sb->s_blocksize) {
fs/nilfs2/dat.c
497
nilfs_err(sb, "too large DAT entry size: %zu bytes",
fs/nilfs2/dat.c
501
nilfs_err(sb, "too small DAT entry size: %zu bytes",
fs/nilfs2/dat.c
506
dat = nilfs_iget_locked(sb, NULL, NILFS_DAT_INO);
fs/nilfs2/dat.h
44
int nilfs_dat_read(struct super_block *sb, size_t entry_size,
fs/nilfs2/dir.c
113
struct super_block *sb = dir->i_sb;
fs/nilfs2/dir.c
152
nilfs_error(sb,
fs/nilfs2/dir.c
171
nilfs_error(sb,
fs/nilfs2/dir.c
179
nilfs_error(sb,
fs/nilfs2/dir.c
239
struct super_block *sb = inode->i_sb;
fs/nilfs2/dir.c
254
nilfs_error(sb, "bad page in #%lu", inode->i_ino);
fs/nilfs2/dir.c
263
nilfs_error(sb, "zero-length directory entry");
fs/nilfs2/ifile.c
129
struct super_block *sb = ifile->i_sb;
fs/nilfs2/ifile.c
132
if (unlikely(!NILFS_VALID_INODE(sb, ino))) {
fs/nilfs2/ifile.c
133
nilfs_error(sb, "bad inode number: %lu", (unsigned long)ino);
fs/nilfs2/ifile.c
139
nilfs_warn(sb, "error %d reading inode: ino=%lu",
fs/nilfs2/ifile.c
181
int nilfs_ifile_read(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/ifile.c
188
ifile = nilfs_iget_locked(sb, root, NILFS_IFILE_INO);
fs/nilfs2/ifile.c
205
nilfs = sb->s_fs_info;
fs/nilfs2/ifile.h
40
int nilfs_ifile_read(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/inode.c
302
struct super_block *sb = dir->i_sb;
fs/nilfs2/inode.c
310
inode = new_inode(sb);
fs/nilfs2/inode.c
445
static int __nilfs_read_inode(struct super_block *sb,
fs/nilfs2/inode.c
449
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/inode.c
484
nilfs_error(sb,
fs/nilfs2/inode.c
535
struct inode *nilfs_ilookup(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/inode.c
542
return ilookup5(sb, ino, nilfs_iget_test, &args);
fs/nilfs2/inode.c
545
struct inode *nilfs_iget_locked(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/inode.c
552
return iget5_locked(sb, ino, nilfs_iget_test, nilfs_iget_set, &args);
fs/nilfs2/inode.c
555
struct inode *nilfs_iget(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/inode.c
561
inode = nilfs_iget_locked(sb, root, ino);
fs/nilfs2/inode.c
573
err = __nilfs_read_inode(sb, root, ino, inode);
fs/nilfs2/inode.c
582
struct inode *nilfs_iget_for_gc(struct super_block *sb, unsigned long ino,
fs/nilfs2/inode.c
591
inode = iget5_locked(sb, ino, nilfs_iget_test, nilfs_iget_set, &args);
fs/nilfs2/inode.c
801
struct super_block *sb = inode->i_sb;
fs/nilfs2/inode.c
809
blocksize = sb->s_blocksize;
fs/nilfs2/inode.c
810
blkoff = (inode->i_size + blocksize - 1) >> sb->s_blocksize_bits;
fs/nilfs2/inode.c
811
nilfs_transaction_begin(sb, &ti, 0); /* never fails */
fs/nilfs2/inode.c
823
nilfs_transaction_commit(sb);
fs/nilfs2/inode.c
857
struct super_block *sb = inode->i_sb;
fs/nilfs2/inode.c
868
nilfs_transaction_begin(sb, &ti, 0); /* never fails */
fs/nilfs2/inode.c
872
nilfs = sb->s_fs_info;
fs/nilfs2/inode.c
873
if (unlikely(sb_rdonly(sb) || !nilfs->ns_writer)) {
fs/nilfs2/inode.c
885
nilfs_transaction_abort(sb);
fs/nilfs2/inode.c
902
nilfs_transaction_commit(sb);
fs/nilfs2/inode.c
914
struct super_block *sb = inode->i_sb;
fs/nilfs2/inode.c
921
err = nilfs_transaction_begin(sb, &ti, 0);
fs/nilfs2/inode.c
941
return nilfs_transaction_commit(sb);
fs/nilfs2/inode.c
944
nilfs_transaction_abort(sb);
fs/nilfs2/ioctl.c
1234
static int nilfs_ioctl_get_fslabel(struct super_block *sb, void __user *argp)
fs/nilfs2/ioctl.c
1236
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/ioctl.c
1264
static int nilfs_ioctl_set_fslabel(struct super_block *sb, struct file *filp,
fs/nilfs2/ioctl.c
1268
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/ioctl.c
1287
nilfs_err(sb, "unable to set label with more than %zu bytes",
fs/nilfs2/ioctl.c
1294
sbp = nilfs_prepare_super(sb, false);
fs/nilfs2/ioctl.c
1304
ret = nilfs_commit_super(sb, NILFS_SB_COMMIT_ALL);
fs/nilfs2/ioctl.c
590
static int nilfs_ioctl_move_blocks(struct super_block *sb,
fs/nilfs2/ioctl.c
594
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/ioctl.c
606
inode = nilfs_iget_for_gc(sb, ino, cno);
fs/nilfs2/mdt.c
108
err = nilfs_transaction_commit(sb);
fs/nilfs2/mdt.c
110
nilfs_transaction_abort(sb);
fs/nilfs2/mdt.c
400
struct super_block *sb;
fs/nilfs2/mdt.c
421
sb = inode->i_sb;
fs/nilfs2/mdt.c
424
err = nilfs_construct_segment(sb);
fs/nilfs2/mdt.c
75
struct super_block *sb = inode->i_sb;
fs/nilfs2/mdt.c
80
nilfs_transaction_begin(sb, &ti, 0);
fs/nilfs2/namei.c
145
struct super_block *sb = dir->i_sb;
fs/nilfs2/namei.c
150
if (l > sb->s_blocksize)
fs/nilfs2/namei.c
479
static struct dentry *nilfs_get_dentry(struct super_block *sb, u64 cno,
fs/nilfs2/namei.c
485
if (ino < NILFS_FIRST_INO(sb) && ino != NILFS_ROOT_INO)
fs/nilfs2/namei.c
488
root = nilfs_lookup_root(sb->s_fs_info, cno);
fs/nilfs2/namei.c
492
inode = nilfs_iget(sb, root, ino);
fs/nilfs2/namei.c
504
static struct dentry *nilfs_fh_to_dentry(struct super_block *sb, struct fid *fh,
fs/nilfs2/namei.c
514
return nilfs_get_dentry(sb, fid->cno, fid->ino, fid->gen);
fs/nilfs2/namei.c
517
static struct dentry *nilfs_fh_to_parent(struct super_block *sb, struct fid *fh,
fs/nilfs2/namei.c
526
return nilfs_get_dentry(sb, fid->cno, fid->parent_ino, fid->parent_gen);
fs/nilfs2/nilfs.h
138
#define NILFS_FIRST_INO(sb) (((struct the_nilfs *)sb->s_fs_info)->ns_first_ino)
fs/nilfs2/nilfs.h
140
#define NILFS_MDT_INODE(sb, ino) \
fs/nilfs2/nilfs.h
142
#define NILFS_VALID_INODE(sb, ino) \
fs/nilfs2/nilfs.h
143
((ino) >= NILFS_FIRST_INO(sb) || \
fs/nilfs2/nilfs.h
289
struct inode *nilfs_ilookup(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/nilfs.h
291
struct inode *nilfs_iget_locked(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/nilfs.h
293
struct inode *nilfs_iget(struct super_block *sb, struct nilfs_root *root,
fs/nilfs2/nilfs.h
295
extern struct inode *nilfs_iget_for_gc(struct super_block *sb,
fs/nilfs2/nilfs.h
328
void __nilfs_msg(struct super_block *sb, const char *fmt, ...);
fs/nilfs2/nilfs.h
330
void __nilfs_error(struct super_block *sb, const char *function,
fs/nilfs2/nilfs.h
335
#define nilfs_msg(sb, level, fmt, ...) \
fs/nilfs2/nilfs.h
336
__nilfs_msg(sb, level fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
337
#define nilfs_error(sb, fmt, ...) \
fs/nilfs2/nilfs.h
338
__nilfs_error(sb, __func__, fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
342
#define nilfs_msg(sb, level, fmt, ...) \
fs/nilfs2/nilfs.h
345
(void)(sb); \
fs/nilfs2/nilfs.h
347
#define nilfs_error(sb, fmt, ...) \
fs/nilfs2/nilfs.h
350
__nilfs_error(sb, "", " "); \
fs/nilfs2/nilfs.h
355
#define nilfs_crit(sb, fmt, ...) \
fs/nilfs2/nilfs.h
356
nilfs_msg(sb, KERN_CRIT, fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
357
#define nilfs_err(sb, fmt, ...) \
fs/nilfs2/nilfs.h
358
nilfs_msg(sb, KERN_ERR, fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
359
#define nilfs_warn(sb, fmt, ...) \
fs/nilfs2/nilfs.h
360
nilfs_msg(sb, KERN_WARNING, fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
361
#define nilfs_info(sb, fmt, ...) \
fs/nilfs2/nilfs.h
362
nilfs_msg(sb, KERN_INFO, fmt, ##__VA_ARGS__)
fs/nilfs2/nilfs.h
366
extern int nilfs_store_magic(struct super_block *sb,
fs/nilfs2/nilfs.h
372
struct nilfs_super_block **nilfs_prepare_super(struct super_block *sb,
fs/nilfs2/nilfs.h
374
int nilfs_commit_super(struct super_block *sb, int flag);
fs/nilfs2/nilfs.h
375
int nilfs_cleanup_super(struct super_block *sb);
fs/nilfs2/nilfs.h
376
int nilfs_resize_fs(struct super_block *sb, __u64 newsize);
fs/nilfs2/nilfs.h
377
int nilfs_attach_checkpoint(struct super_block *sb, __u64 cno, int curr_mnt,
fs/nilfs2/nilfs.h
379
int nilfs_checkpoint_is_mounted(struct super_block *sb, __u64 cno);
fs/nilfs2/recovery.c
440
struct super_block *sb,
fs/nilfs2/recovery.c
462
nilfs_err(sb,
fs/nilfs2/recovery.c
48
static int nilfs_warn_segment_error(struct super_block *sb, int err)
fs/nilfs2/recovery.c
523
struct super_block *sb,
fs/nilfs2/recovery.c
536
inode = nilfs_iget(sb, root, rb->ino);
fs/nilfs2/recovery.c
54
nilfs_err(sb, "I/O error reading segment");
fs/nilfs2/recovery.c
576
nilfs_warn(sb,
fs/nilfs2/recovery.c
605
struct super_block *sb,
fs/nilfs2/recovery.c
677
nilfs, sb, root, &dsync_blocks,
fs/nilfs2/recovery.c
709
nilfs_info(sb, "salvaged %lu blocks", nsalvaged_blocks);
fs/nilfs2/recovery.c
720
nilfs_err(sb,
fs/nilfs2/recovery.c
75
nilfs_err(sb, "unrecognized segment error %d", err);
fs/nilfs2/recovery.c
78
nilfs_warn(sb, "invalid segment: %s", msg);
fs/nilfs2/recovery.c
795
struct super_block *sb,
fs/nilfs2/recovery.c
804
err = nilfs_attach_checkpoint(sb, ri->ri_cno, true, &root);
fs/nilfs2/recovery.c
806
nilfs_err(sb, "error %d loading the latest checkpoint", err);
fs/nilfs2/recovery.c
810
err = nilfs_do_roll_forward(nilfs, sb, root, ri);
fs/nilfs2/recovery.c
815
err = nilfs_prepare_segment_for_recovery(nilfs, sb, ri);
fs/nilfs2/recovery.c
817
nilfs_err(sb, "error %d preparing segment for recovery",
fs/nilfs2/recovery.c
822
err = nilfs_attach_log_writer(sb, root);
fs/nilfs2/recovery.c
827
err = nilfs_construct_segment(sb);
fs/nilfs2/recovery.c
828
nilfs_detach_log_writer(sb);
fs/nilfs2/recovery.c
831
nilfs_err(sb, "error %d writing segment for recovery",
fs/nilfs2/segbuf.c
34
struct nilfs_segment_buffer *nilfs_segbuf_new(struct super_block *sb)
fs/nilfs2/segbuf.c
42
segbuf->sb_super = sb;
fs/nilfs2/segment.c
141
static int nilfs_prepare_segment_lock(struct super_block *sb,
fs/nilfs2/segment.c
156
nilfs_warn(sb, "journal info from a different FS");
fs/nilfs2/segment.c
199
int nilfs_transaction_begin(struct super_block *sb,
fs/nilfs2/segment.c
204
int ret = nilfs_prepare_segment_lock(sb, ti);
fs/nilfs2/segment.c
212
trace_nilfs2_transaction_transition(sb, trace_ti,
fs/nilfs2/segment.c
218
sb_start_intwrite(sb);
fs/nilfs2/segment.c
220
nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
229
trace_nilfs2_transaction_transition(sb, trace_ti, trace_ti->ti_count,
fs/nilfs2/segment.c
2314
int nilfs_construct_segment(struct super_block *sb)
fs/nilfs2/segment.c
2316
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
2320
if (sb_rdonly(sb) || unlikely(!sci))
fs/nilfs2/segment.c
2344
int nilfs_construct_dsync_segment(struct super_block *sb, struct inode *inode,
fs/nilfs2/segment.c
2347
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
2353
if (sb_rdonly(sb) || unlikely(!sci))
fs/nilfs2/segment.c
2356
nilfs_transaction_lock(sb, &ti, 0);
fs/nilfs2/segment.c
2363
nilfs_transaction_unlock(sb);
fs/nilfs2/segment.c
2372
nilfs_transaction_unlock(sb);
fs/nilfs2/segment.c
2384
nilfs_transaction_unlock(sb);
fs/nilfs2/segment.c
239
sb_end_intwrite(sb);
fs/nilfs2/segment.c
2508
int nilfs_clean_segments(struct super_block *sb, struct nilfs_argv *argv,
fs/nilfs2/segment.c
2511
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
2519
nilfs_transaction_lock(sb, &ti, 1);
fs/nilfs2/segment.c
2542
nilfs_warn(sb, "error %d cleaning segments", err);
fs/nilfs2/segment.c
2550
nilfs_warn(sb,
fs/nilfs2/segment.c
256
int nilfs_transaction_commit(struct super_block *sb)
fs/nilfs2/segment.c
2561
nilfs_transaction_unlock(sb);
fs/nilfs2/segment.c
259
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
266
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
2698
static struct nilfs_sc_info *nilfs_segctor_new(struct super_block *sb,
fs/nilfs2/segment.c
2701
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
2708
sci->sc_super = sb;
fs/nilfs2/segment.c
279
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
2832
int nilfs_attach_log_writer(struct super_block *sb, struct nilfs_root *root)
fs/nilfs2/segment.c
2834
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
2849
sci = nilfs_segctor_new(sb, root);
fs/nilfs2/segment.c
285
err = nilfs_construct_segment(sb);
fs/nilfs2/segment.c
2857
nilfs_err(sb, "error %d creating segctord thread", err);
fs/nilfs2/segment.c
2858
nilfs_detach_log_writer(sb);
fs/nilfs2/segment.c
2875
void nilfs_detach_log_writer(struct super_block *sb)
fs/nilfs2/segment.c
2877
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
288
sb_end_intwrite(sb);
fs/nilfs2/segment.c
2891
nilfs_warn(sb,
fs/nilfs2/segment.c
292
void nilfs_transaction_abort(struct super_block *sb)
fs/nilfs2/segment.c
295
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
300
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
306
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
312
sb_end_intwrite(sb);
fs/nilfs2/segment.c
315
void nilfs_relax_pressure_in_lock(struct super_block *sb)
fs/nilfs2/segment.c
317
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
320
if (sb_rdonly(sb) || unlikely(!sci) || !sci->sc_flush_request)
fs/nilfs2/segment.c
338
static void nilfs_transaction_lock(struct super_block *sb,
fs/nilfs2/segment.c
343
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
354
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
369
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.c
373
static void nilfs_transaction_unlock(struct super_block *sb)
fs/nilfs2/segment.c
376
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/segment.c
384
trace_nilfs2_transaction_transition(sb, ti, ti->ti_count,
fs/nilfs2/segment.h
232
int nilfs_attach_log_writer(struct super_block *sb, struct nilfs_root *root);
fs/nilfs2/segment.h
233
void nilfs_detach_log_writer(struct super_block *sb);
fs/nilfs2/segment.h
240
int nilfs_salvage_orphan_logs(struct the_nilfs *nilfs, struct super_block *sb,
fs/nilfs2/sufile.c
1211
int nilfs_sufile_read(struct super_block *sb, size_t susize,
fs/nilfs2/sufile.c
1220
if (susize > sb->s_blocksize) {
fs/nilfs2/sufile.c
1221
nilfs_err(sb, "too large segment usage size: %zu bytes",
fs/nilfs2/sufile.c
1225
nilfs_err(sb, "too small segment usage size: %zu bytes",
fs/nilfs2/sufile.c
1230
sufile = nilfs_iget_locked(sb, NULL, NILFS_SUFILE_INO);
fs/nilfs2/sufile.c
1250
nilfs_err(sb,
fs/nilfs2/sufile.h
53
int nilfs_sufile_read(struct super_block *sb, size_t susize,
fs/nilfs2/super.c
100
nilfs_commit_super(sb, NILFS_SB_COMMIT_ALL);
fs/nilfs2/super.c
1004
int nilfs_checkpoint_is_mounted(struct super_block *sb, __u64 cno)
fs/nilfs2/super.c
1006
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
1021
inode = nilfs_ilookup(sb, root, NILFS_ROOT_INO);
fs/nilfs2/super.c
1046
nilfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/nilfs2/super.c
1054
nilfs = alloc_nilfs(sb);
fs/nilfs2/super.c
1058
sb->s_fs_info = nilfs;
fs/nilfs2/super.c
1060
err = init_nilfs(nilfs, sb);
fs/nilfs2/super.c
1067
sb->s_op = &nilfs_sops;
fs/nilfs2/super.c
1068
sb->s_export_op = &nilfs_export_ops;
fs/nilfs2/super.c
1069
sb->s_root = NULL;
fs/nilfs2/super.c
1070
sb->s_time_gran = 1;
fs/nilfs2/super.c
1071
sb->s_max_links = NILFS_LINK_MAX;
fs/nilfs2/super.c
1073
sb->s_bdi = bdi_get(sb->s_bdev->bd_disk->bdi);
fs/nilfs2/super.c
1075
err = load_nilfs(nilfs, sb);
fs/nilfs2/super.c
1079
super_set_uuid(sb, nilfs->ns_sbp[0]->s_uuid,
fs/nilfs2/super.c
1081
super_set_sysfs_name_bdev(sb);
fs/nilfs2/super.c
1084
err = nilfs_attach_checkpoint(sb, cno, true, &fsroot);
fs/nilfs2/super.c
1086
nilfs_err(sb,
fs/nilfs2/super.c
1092
if (!sb_rdonly(sb)) {
fs/nilfs2/super.c
1093
err = nilfs_attach_log_writer(sb, fsroot);
fs/nilfs2/super.c
1098
err = nilfs_get_root_dentry(sb, fsroot, &sb->s_root);
fs/nilfs2/super.c
1104
if (!sb_rdonly(sb)) {
fs/nilfs2/super.c
1106
nilfs_setup_super(sb, true);
fs/nilfs2/super.c
1113
nilfs_detach_log_writer(sb);
fs/nilfs2/super.c
1132
struct super_block *sb = fc->root->d_sb;
fs/nilfs2/super.c
1133
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
1136
sync_filesystem(sb);
fs/nilfs2/super.c
1141
nilfs_warn(sb,
fs/nilfs2/super.c
1145
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb))
fs/nilfs2/super.c
1148
sb->s_flags |= SB_RDONLY;
fs/nilfs2/super.c
1155
nilfs_cleanup_super(sb);
fs/nilfs2/super.c
1171
nilfs_warn(sb,
fs/nilfs2/super.c
1178
sb->s_flags &= ~SB_RDONLY;
fs/nilfs2/super.c
1180
root = NILFS_I(d_inode(sb->s_root))->i_root;
fs/nilfs2/super.c
1181
err = nilfs_attach_log_writer(sb, root);
fs/nilfs2/super.c
1183
sb->s_flags |= SB_RDONLY;
fs/nilfs2/super.c
1188
nilfs_setup_super(sb, true);
fs/nilfs2/super.c
1192
sb->s_flags = (sb->s_flags & ~SB_POSIXACL);
fs/nilfs2/super.c
123
void __nilfs_error(struct super_block *sb, const char *function,
fs/nilfs2/super.c
126
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
136
sb->s_id, function, &vaf);
fs/nilfs2/super.c
140
if (!sb_rdonly(sb)) {
fs/nilfs2/super.c
141
nilfs_set_error(sb);
fs/nilfs2/super.c
145
sb->s_flags |= SB_RDONLY;
fs/nilfs2/super.c
151
sb->s_id);
fs/nilfs2/super.c
154
struct inode *nilfs_alloc_inode(struct super_block *sb)
fs/nilfs2/super.c
158
ii = alloc_inode_sb(sb, nilfs_inode_cachep, GFP_NOFS);
fs/nilfs2/super.c
178
static int nilfs_sync_super(struct super_block *sb, int flag)
fs/nilfs2/super.c
180
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
193
nilfs_err(sb, "unable to write superblock: err=%d", err);
fs/nilfs2/super.c
251
struct nilfs_super_block **nilfs_prepare_super(struct super_block *sb,
fs/nilfs2/super.c
254
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
263
nilfs_crit(sb, "superblock broke");
fs/nilfs2/super.c
277
int nilfs_commit_super(struct super_block *sb, int flag)
fs/nilfs2/super.c
279
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
302
return nilfs_sync_super(sb, flag);
fs/nilfs2/super.c
315
int nilfs_cleanup_super(struct super_block *sb)
fs/nilfs2/super.c
317
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
322
sbp = nilfs_prepare_super(sb, 0);
fs/nilfs2/super.c
335
ret = nilfs_commit_super(sb, flag);
fs/nilfs2/super.c
347
static int nilfs_move_2nd_super(struct super_block *sb, loff_t sb2off)
fs/nilfs2/super.c
349
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
375
nsbh = sb_getblk(sb, newblocknr);
fs/nilfs2/super.c
377
nilfs_warn(sb,
fs/nilfs2/super.c
430
int nilfs_resize_fs(struct super_block *sb, __u64 newsize)
fs/nilfs2/super.c
432
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
439
devsize = bdev_nr_bytes(sb->s_bdev);
fs/nilfs2/super.c
468
ret = nilfs_construct_segment(sb);
fs/nilfs2/super.c
473
nilfs_move_2nd_super(sb, sb2off);
fs/nilfs2/super.c
475
sbp = nilfs_prepare_super(sb, 0);
fs/nilfs2/super.c
489
ret = nilfs_commit_super(sb, NILFS_SB_COMMIT_ALL);
fs/nilfs2/super.c
505
static void nilfs_put_super(struct super_block *sb)
fs/nilfs2/super.c
507
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
509
nilfs_detach_log_writer(sb);
fs/nilfs2/super.c
511
if (!sb_rdonly(sb)) {
fs/nilfs2/super.c
513
nilfs_cleanup_super(sb);
fs/nilfs2/super.c
523
sb->s_fs_info = NULL;
fs/nilfs2/super.c
526
static int nilfs_sync_fs(struct super_block *sb, int wait)
fs/nilfs2/super.c
528
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
534
err = nilfs_construct_segment(sb);
fs/nilfs2/super.c
538
sbp = nilfs_prepare_super(sb, nilfs_sb_will_flip(nilfs));
fs/nilfs2/super.c
541
nilfs_commit_super(sb, NILFS_SB_COMMIT);
fs/nilfs2/super.c
552
int nilfs_attach_checkpoint(struct super_block *sb, __u64 cno, int curr_mnt,
fs/nilfs2/super.c
555
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
568
err = nilfs_ifile_read(sb, root, cno, nilfs->ns_inode_size);
fs/nilfs2/super.c
579
nilfs_err(sb, "Invalid checkpoint (checkpoint number=%llu)",
fs/nilfs2/super.c
586
static int nilfs_freeze(struct super_block *sb)
fs/nilfs2/super.c
588
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
591
if (sb_rdonly(sb))
fs/nilfs2/super.c
596
err = nilfs_cleanup_super(sb);
fs/nilfs2/super.c
601
static int nilfs_unfreeze(struct super_block *sb)
fs/nilfs2/super.c
603
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
605
if (sb_rdonly(sb))
fs/nilfs2/super.c
609
nilfs_setup_super(sb, false);
fs/nilfs2/super.c
616
struct super_block *sb = dentry->d_sb;
fs/nilfs2/super.c
619
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/nilfs2/super.c
63
static int nilfs_setup_super(struct super_block *sb, int is_mount);
fs/nilfs2/super.c
65
void __nilfs_msg(struct super_block *sb, const char *fmt, ...)
fs/nilfs2/super.c
652
nilfs_warn(sb, "failed to count free inodes: err=%d", err);
fs/nilfs2/super.c
668
buf->f_bsize = sb->s_blocksize;
fs/nilfs2/super.c
683
struct super_block *sb = dentry->d_sb;
fs/nilfs2/super.c
684
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
77
if (sb)
fs/nilfs2/super.c
779
struct super_block *sb = fc->root->d_sb;
fs/nilfs2/super.c
781
nilfs_err(sb,
fs/nilfs2/super.c
79
KERN_SOH_ASCII, level, sb->s_id, &vaf);
fs/nilfs2/super.c
809
static int nilfs_setup_super(struct super_block *sb, int is_mount)
fs/nilfs2/super.c
811
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
817
sbp = nilfs_prepare_super(sb, 0);
fs/nilfs2/super.c
828
nilfs_warn(sb, "mounting fs with errors");
fs/nilfs2/super.c
831
nilfs_warn(sb, "maximal mount count reached");
fs/nilfs2/super.c
846
return nilfs_commit_super(sb, NILFS_SB_COMMIT_ALL);
fs/nilfs2/super.c
849
struct nilfs_super_block *nilfs_read_super_block(struct super_block *sb,
fs/nilfs2/super.c
857
*pbh = sb_bread(sb, sb_index);
fs/nilfs2/super.c
863
int nilfs_store_magic(struct super_block *sb,
fs/nilfs2/super.c
866
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
868
sb->s_magic = le16_to_cpu(sbp->s_magic);
fs/nilfs2/super.c
87
static void nilfs_set_error(struct super_block *sb)
fs/nilfs2/super.c
872
sb->s_flags |= SB_NOATIME;
fs/nilfs2/super.c
883
int nilfs_check_feature_compatibility(struct super_block *sb,
fs/nilfs2/super.c
89
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/super.c
891
nilfs_err(sb,
fs/nilfs2/super.c
898
if (!sb_rdonly(sb) && features) {
fs/nilfs2/super.c
899
nilfs_err(sb,
fs/nilfs2/super.c
907
static int nilfs_get_root_dentry(struct super_block *sb,
fs/nilfs2/super.c
915
inode = nilfs_iget(sb, root, NILFS_ROOT_INO);
fs/nilfs2/super.c
918
nilfs_err(sb, "error %d getting root inode", ret);
fs/nilfs2/super.c
923
nilfs_err(sb, "corrupt root inode");
fs/nilfs2/super.c
95
sbp = nilfs_prepare_super(sb, 0);
fs/nilfs2/super.c
951
nilfs_err(sb, "error %d getting root dentry", ret);
fs/nilfs2/sysfs.c
1001
struct the_nilfs *nilfs = sb->s_fs_info;
fs/nilfs2/sysfs.c
1008
nilfs_err(sb, "unable to allocate memory for device group");
fs/nilfs2/sysfs.c
1015
"%s", sb->s_id);
fs/nilfs2/sysfs.c
999
int nilfs_sysfs_create_device_group(struct super_block *sb)
fs/nilfs2/the_nilfs.c
118
err = nilfs_dat_read(sb, dat_entry_size, rawi, &nilfs->ns_dat);
fs/nilfs2/the_nilfs.c
123
err = nilfs_cpfile_read(sb, checkpoint_size, rawi, &nilfs->ns_cpfile);
fs/nilfs2/the_nilfs.c
128
err = nilfs_sufile_read(sb, segment_usage_size, rawi,
fs/nilfs2/the_nilfs.c
208
static int nilfs_get_blocksize(struct super_block *sb,
fs/nilfs2/the_nilfs.c
215
nilfs_err(sb, "too large filesystem blocksize: 2 ^ %u KiB",
fs/nilfs2/the_nilfs.c
239
int load_nilfs(struct the_nilfs *nilfs, struct super_block *sb)
fs/nilfs2/the_nilfs.c
242
unsigned int s_flags = sb->s_flags;
fs/nilfs2/the_nilfs.c
248
nilfs_warn(sb, "mounting unchecked fs");
fs/nilfs2/the_nilfs.c
250
nilfs_info(sb,
fs/nilfs2/the_nilfs.c
252
nilfs_info(sb,
fs/nilfs2/the_nilfs.c
268
nilfs_warn(sb,
fs/nilfs2/the_nilfs.c
272
nilfs_info(sb, "trying rollback from an earlier position");
fs/nilfs2/the_nilfs.c
283
err = nilfs_get_blocksize(sb, sbp[0], &blocksize);
fs/nilfs2/the_nilfs.c
288
nilfs_warn(sb,
fs/nilfs2/the_nilfs.c
308
err = nilfs_load_super_root(nilfs, sb, ri.ri_super_root);
fs/nilfs2/the_nilfs.c
310
nilfs_err(sb, "error %d while loading super root", err);
fs/nilfs2/the_nilfs.c
314
err = nilfs_sysfs_create_device_group(sb);
fs/nilfs2/the_nilfs.c
325
nilfs_info(sb,
fs/nilfs2/the_nilfs.c
332
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
339
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
344
sb->s_flags &= ~SB_RDONLY;
fs/nilfs2/the_nilfs.c
346
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
352
err = nilfs_salvage_orphan_logs(nilfs, sb, &ri);
fs/nilfs2/the_nilfs.c
358
err = nilfs_cleanup_super(sb);
fs/nilfs2/the_nilfs.c
362
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
367
nilfs_info(sb, "recovery complete");
fs/nilfs2/the_nilfs.c
371
sb->s_flags = s_flags;
fs/nilfs2/the_nilfs.c
375
nilfs_err(sb, "error %d while searching super root", err);
fs/nilfs2/the_nilfs.c
388
sb->s_flags = s_flags;
fs/nilfs2/the_nilfs.c
55
struct the_nilfs *alloc_nilfs(struct super_block *sb)
fs/nilfs2/the_nilfs.c
606
struct super_block *sb, int blocksize,
fs/nilfs2/the_nilfs.c
615
nilfs_err(sb, "device size too small");
fs/nilfs2/the_nilfs.c
620
sbp[0] = nilfs_read_super_block(sb, NILFS_SB_OFFSET_BYTES, blocksize,
fs/nilfs2/the_nilfs.c
622
sbp[1] = nilfs_read_super_block(sb, sb2off, blocksize, &sbh[1]);
fs/nilfs2/the_nilfs.c
626
nilfs_err(sb, "unable to read superblock");
fs/nilfs2/the_nilfs.c
629
nilfs_warn(sb,
fs/nilfs2/the_nilfs.c
63
nilfs->ns_sb = sb;
fs/nilfs2/the_nilfs.c
633
nilfs_warn(sb,
fs/nilfs2/the_nilfs.c
64
nilfs->ns_bdev = sb->s_bdev;
fs/nilfs2/the_nilfs.c
657
nilfs_err(sb, "couldn't find nilfs on the device");
fs/nilfs2/the_nilfs.c
662
nilfs_warn(sb,
fs/nilfs2/the_nilfs.c
702
int init_nilfs(struct the_nilfs *nilfs, struct super_block *sb)
fs/nilfs2/the_nilfs.c
708
blocksize = sb_min_blocksize(sb, NILFS_MIN_BLOCK_SIZE);
fs/nilfs2/the_nilfs.c
710
nilfs_err(sb, "unable to set blocksize");
fs/nilfs2/the_nilfs.c
714
err = nilfs_load_super_block(nilfs, sb, blocksize, &sbp);
fs/nilfs2/the_nilfs.c
718
err = nilfs_store_magic(sb, sbp);
fs/nilfs2/the_nilfs.c
722
err = nilfs_check_feature_compatibility(sb, sbp);
fs/nilfs2/the_nilfs.c
726
err = nilfs_get_blocksize(sb, sbp, &blocksize);
fs/nilfs2/the_nilfs.c
731
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
737
if (sb->s_blocksize != blocksize) {
fs/nilfs2/the_nilfs.c
738
int hw_blocksize = bdev_logical_block_size(sb->s_bdev);
fs/nilfs2/the_nilfs.c
741
nilfs_err(sb,
fs/nilfs2/the_nilfs.c
748
if (!sb_set_blocksize(sb, blocksize)) {
fs/nilfs2/the_nilfs.c
749
nilfs_err(sb, "bad blocksize %d", blocksize);
fs/nilfs2/the_nilfs.c
754
err = nilfs_load_super_block(nilfs, sb, blocksize, &sbp);
fs/nilfs2/the_nilfs.c
762
nilfs->ns_blocksize_bits = sb->s_blocksize_bits;
fs/nilfs2/the_nilfs.c
769
sb->s_maxbytes = nilfs_max_size(sb->s_blocksize_bits);
fs/nilfs2/the_nilfs.c
95
struct super_block *sb, sector_t sr_block)
fs/nilfs2/the_nilfs.h
267
struct the_nilfs *alloc_nilfs(struct super_block *sb);
fs/nilfs2/the_nilfs.h
269
int init_nilfs(struct the_nilfs *nilfs, struct super_block *sb);
fs/nilfs2/the_nilfs.h
270
int load_nilfs(struct the_nilfs *nilfs, struct super_block *sb);
fs/notify/fanotify/fanotify_user.c
1362
struct super_block *sb;
fs/notify/fanotify/fanotify_user.c
1407
if (old_sb != fsid->sb)
fs/notify/fanotify/fanotify_user.c
1775
fsid->sb = dentry->d_sb;
fs/notify/fdinfo.c
126
struct super_block *sb = fsnotify_conn_sb(mark->connector);
fs/notify/fdinfo.c
129
sb->s_dev, mflags, mark->mask, mark->ignore_mask);
fs/notify/fsnotify.c
36
void fsnotify_sb_delete(struct super_block *sb)
fs/notify/fsnotify.c
38
struct fsnotify_sb_info *sbinfo = fsnotify_sb_info(sb);
fs/notify/fsnotify.c
45
fsnotify_clear_marks_by_sb(sb);
fs/notify/fsnotify.c
47
wait_var_event(fsnotify_sb_watched_objects(sb),
fs/notify/fsnotify.c
48
!atomic_long_read(fsnotify_sb_watched_objects(sb)));
fs/notify/fsnotify.c
49
WARN_ON(fsnotify_sb_has_priority_watchers(sb, FSNOTIFY_PRIO_CONTENT));
fs/notify/fsnotify.c
496
struct super_block *sb = fsnotify_data_sb(data, data_type);
fs/notify/fsnotify.c
498
struct fsnotify_sb_info *sbinfo = sb ? fsnotify_sb_info(sb) : NULL;
fs/notify/fsnotify.c
50
WARN_ON(fsnotify_sb_has_priority_watchers(sb,
fs/notify/fsnotify.c
54
void fsnotify_sb_free(struct super_block *sb)
fs/notify/fsnotify.c
542
if (sb)
fs/notify/fsnotify.c
543
marks_mask |= READ_ONCE(sb->s_fsnotify_mask);
fs/notify/fsnotify.c
56
if (sb->s_fsnotify_info) {
fs/notify/fsnotify.c
57
WARN_ON_ONCE(!list_empty(&sb->s_fsnotify_info->inode_conn_list));
fs/notify/fsnotify.c
58
kfree(sb->s_fsnotify_info);
fs/notify/fsnotify.c
618
struct super_block *sb = dentry->d_sb;
fs/notify/fsnotify.c
630
if (likely(!fsnotify_sb_has_priority_watchers(sb,
fs/notify/fsnotify.h
63
static inline fsnotify_connp_t *fsnotify_sb_marks(struct super_block *sb)
fs/notify/fsnotify.h
65
struct fsnotify_sb_info *sbinfo = fsnotify_sb_info(sb);
fs/notify/fsnotify.h
96
static inline void fsnotify_clear_marks_by_sb(struct super_block *sb)
fs/notify/fsnotify.h
98
fsnotify_destroy_marks(fsnotify_sb_marks(sb));
fs/notify/mark.c
139
static void fsnotify_get_sb_watched_objects(struct super_block *sb)
fs/notify/mark.c
141
atomic_long_inc(fsnotify_sb_watched_objects(sb));
fs/notify/mark.c
144
static void fsnotify_put_sb_watched_objects(struct super_block *sb)
fs/notify/mark.c
146
atomic_long_t *watched_objects = fsnotify_sb_watched_objects(sb);
fs/notify/mark.c
162
struct super_block *sb = inode->i_sb;
fs/notify/mark.c
165
fsnotify_put_sb_watched_objects(sb);
fs/notify/mark.c
172
static void fsnotify_update_sb_watchers(struct super_block *sb,
fs/notify/mark.c
175
struct fsnotify_sb_info *sbinfo = fsnotify_sb_info(sb);
fs/notify/mark.c
203
fsnotify_get_sb_watched_objects(sb);
fs/notify/mark.c
206
fsnotify_put_sb_watched_objects(sb);
fs/notify/mark.c
338
struct super_block *sb = fsnotify_connector_sb(conn);
fs/notify/mark.c
364
if (sb)
fs/notify/mark.c
365
fsnotify_update_sb_watchers(sb, conn);
fs/notify/mark.c
417
struct super_block *sb = fsnotify_connector_sb(conn);
fs/notify/mark.c
420
if (sb)
fs/notify/mark.c
421
fsnotify_update_sb_watchers(sb, conn);
fs/notify/mark.c
642
static int fsnotify_attach_info_to_sb(struct super_block *sb)
fs/notify/mark.c
657
if (cmpxchg(&sb->s_fsnotify_info, NULL, sbinfo)) {
fs/notify/mark.c
830
struct super_block *sb = fsnotify_object_sb(obj, obj_type);
fs/notify/mark.c
844
if (sb && !fsnotify_sb_info(sb)) {
fs/notify/mark.c
845
err = fsnotify_attach_info_to_sb(sb);
fs/notify/mark.c
890
if (sb)
fs/notify/mark.c
891
fsnotify_update_sb_watchers(sb, conn);
fs/nsfs.c
518
static struct dentry *nsfs_fh_to_dentry(struct super_block *sb, struct fid *fh,
fs/ntfs3/attrib.c
218
err = blkdev_issue_zeroout(sbi->sb->s_bdev,
fs/ntfs3/bitmap.c
1130
sb = wnd->sb;
fs/ntfs3/bitmap.c
1131
log2_bits = sb->s_blocksize_bits + 3;
fs/ntfs3/bitmap.c
1138
wbits = sb->s_blocksize * 8;
fs/ntfs3/bitmap.c
1326
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
1327
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/bitmap.c
1328
u32 blocksize = sb->s_blocksize;
fs/ntfs3/bitmap.c
1339
new_wnd = bytes_to_block(sb, ntfs3_bitmap_size(new_bits));
fs/ntfs3/bitmap.c
1360
for (iw = old_bits >> (sb->s_blocksize_bits + 3); bits; iw += 1) {
fs/ntfs3/bitmap.c
1376
bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
fs/ntfs3/bitmap.c
1423
struct super_block *sb = sbi->sb;
fs/ntfs3/bitmap.c
1425
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
1429
size_t iw = lcn_from >> (sb->s_blocksize_bits + 3);
fs/ntfs3/bitmap.c
501
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
502
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/bitmap.c
504
u32 blocksize = sb->s_blocksize;
fs/ntfs3/bitmap.c
506
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
512
struct address_space *mapping = sb->s_bdev->bd_mapping;
fs/ntfs3/bitmap.c
571
bh = ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
fs/ntfs3/bitmap.c
661
int wnd_init(struct wnd_bitmap *wnd, struct super_block *sb, size_t nbits)
fs/ntfs3/bitmap.c
664
u32 blocksize = sb->s_blocksize;
fs/ntfs3/bitmap.c
669
wnd->sb = sb;
fs/ntfs3/bitmap.c
674
wnd->nwnd = bytes_to_block(sb, ntfs3_bitmap_size(nbits));
fs/ntfs3/bitmap.c
701
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
706
sbi = sb->s_fs_info;
fs/ntfs3/bitmap.c
707
vbo = (u64)iw << sb->s_blocksize_bits;
fs/ntfs3/bitmap.c
716
bh = ntfs_bread(wnd->sb, lbo >> sb->s_blocksize_bits);
fs/ntfs3/bitmap.c
729
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
730
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
731
size_t iw = bit >> (sb->s_blocksize_bits + 3);
fs/ntfs3/bitmap.c
771
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
772
size_t iw = bit >> (sb->s_blocksize_bits + 3);
fs/ntfs3/bitmap.c
773
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
852
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
853
size_t iw = bit >> (sb->s_blocksize_bits + 3);
fs/ntfs3/bitmap.c
854
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
922
struct super_block *sb = wnd->sb;
fs/ntfs3/bitmap.c
923
size_t iw = bit >> (sb->s_blocksize_bits + 3);
fs/ntfs3/bitmap.c
924
u32 wbits = 8 * sb->s_blocksize;
fs/ntfs3/bitmap.c
979
struct super_block *sb;
fs/ntfs3/debug.h
24
void ntfs_printk(const struct super_block *sb, const char *fmt, ...);
fs/ntfs3/debug.h
29
void ntfs_printk(const struct super_block *sb, const char *fmt, ...)
fs/ntfs3/debug.h
43
#define ntfs_err(sb, fmt, ...) ntfs_printk(sb, KERN_ERR fmt, ##__VA_ARGS__)
fs/ntfs3/debug.h
44
#define ntfs_warn(sb, fmt, ...) ntfs_printk(sb, KERN_WARNING fmt, ##__VA_ARGS__)
fs/ntfs3/debug.h
45
#define ntfs_info(sb, fmt, ...) ntfs_printk(sb, KERN_INFO fmt, ##__VA_ARGS__)
fs/ntfs3/debug.h
46
#define ntfs_notice(sb, fmt, ...) \
fs/ntfs3/debug.h
47
ntfs_printk(sb, KERN_NOTICE fmt, ##__VA_ARGS__)
fs/ntfs3/dir.c
238
struct super_block *sb = dir->i_sb;
fs/ntfs3/dir.c
239
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/dir.c
265
inode = ntfs_iget5(sb, &e->ref, uni);
fs/ntfs3/dir.c
314
ntfs_warn(sbi->sb, "failed to convert name for inode %lx.",
fs/ntfs3/dir.c
337
struct inode *inode = ntfs_iget5(sbi->sb, &e->ref, NULL);
fs/ntfs3/dir.c
405
struct super_block *sb = dir->i_sb;
fs/ntfs3/dir.c
406
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/dir.c
45
ntfs_warn(sbi->sb,
fs/ntfs3/dir.c
69
ntfs_err(sbi->sb, "failed to convert \"%s\" to %s", dump,
fs/ntfs3/file.c
119
static int ntfs_force_shutdown(struct super_block *sb, u32 flags)
fs/ntfs3/file.c
122
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/file.c
124
if (unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/file.c
128
err = bdev_freeze(sb->s_bdev);
fs/ntfs3/file.c
132
bdev_thaw(sb->s_bdev);
fs/ntfs3/file.c
136
static int ntfs_ioctl_shutdown(struct super_block *sb, unsigned long arg)
fs/ntfs3/file.c
146
return ntfs_force_shutdown(sb, flags);
fs/ntfs3/file.c
1490
struct super_block *sb = inode->i_sb;
fs/ntfs3/file.c
1491
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/file.c
1494
if (unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/file.c
1512
err = sync_blockdev(sb->s_bdev);
fs/ntfs3/file.c
1516
blkdev_issue_flush(sb->s_bdev);
fs/ntfs3/file.c
155
struct super_block *sb = inode->i_sb;
fs/ntfs3/file.c
156
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/file.c
170
return ntfs_ioctl_shutdown(sb, arg);
fs/ntfs3/file.c
453
struct super_block *sb = inode->i_sb;
fs/ntfs3/file.c
454
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/file.c
72
dev = sbi->sb->s_bdev;
fs/ntfs3/frecord.c
2925
struct super_block *sb = sbi->sb;
fs/ntfs3/frecord.c
2945
dir = ntfs_iget5(sb, &fname->home, NULL);
fs/ntfs3/frecord.c
2977
struct super_block *sb = sbi->sb;
fs/ntfs3/frecord.c
3065
dir = ntfs_iget5(sb, &fname->home, NULL);
fs/ntfs3/frecord.c
3099
struct super_block *sb = inode->i_sb;
fs/ntfs3/frecord.c
3100
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/frecord.c
3106
if (is_bad_inode(inode) || sb_rdonly(sb))
fs/ntfs3/frecord.c
3113
if (unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/frecord.c
3172
&& (sb->s_flags & SB_ACTIVE)) {
fs/ntfs3/fslog.c
2172
if (sb_rdonly(log->ni->mi.sbi->sb)) {
fs/ntfs3/fslog.c
2231
if (sb_rdonly(log->ni->mi.sbi->sb)) {
fs/ntfs3/fslog.c
3113
inode = ilookup(sbi->sb, rno);
fs/ntfs3/fslog.c
3147
inode_parent = ntfs_iget5(sbi->sb, &rec->parent_ref, NULL);
fs/ntfs3/fslog.c
3776
bool is_ro = sb_rdonly(sbi->sb);
fs/ntfs3/fslog.c
4070
ntfs_warn(sbi->sb, "\x24LogFile version %d.%d is not supported",
fs/ntfs3/fslog.c
4733
inode = ntfs_iget5(sbi->sb, &oe->ref, NULL);
fs/ntfs3/fsntfs.c
1036
struct buffer_head *ntfs_bread(struct super_block *sb, sector_t block)
fs/ntfs3/fsntfs.c
1038
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/fsntfs.c
1043
ntfs_err(sb, "try to read out of volume at offset 0x%llx",
fs/ntfs3/fsntfs.c
1044
(u64)block << sb->s_blocksize_bits);
fs/ntfs3/fsntfs.c
1048
bh = sb_bread_unmovable(sb, block);
fs/ntfs3/fsntfs.c
1052
ntfs_err(sb, "failed to read volume at offset 0x%llx",
fs/ntfs3/fsntfs.c
1053
(u64)block << sb->s_blocksize_bits);
fs/ntfs3/fsntfs.c
1057
int ntfs_sb_write(struct super_block *sb, u64 lbo, size_t bytes,
fs/ntfs3/fsntfs.c
1060
u32 blocksize = sb->s_blocksize;
fs/ntfs3/fsntfs.c
1061
struct block_device *bdev = sb->s_bdev;
fs/ntfs3/fsntfs.c
1062
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1067
if (!wait && (sb->s_flags & SB_SYNCHRONOUS))
fs/ntfs3/fsntfs.c
1077
ntfs_err(sb, "failed to read block %llx",
fs/ntfs3/fsntfs.c
1105
sb,
fs/ntfs3/fsntfs.c
1123
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1141
int err = ntfs_sb_write(sb, lbo, op, buf, sync);
fs/ntfs3/fsntfs.c
1171
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1181
return ntfs_bread(sb, lbo >> sb->s_blocksize_bits);
fs/ntfs3/fsntfs.c
1189
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1190
struct address_space *mapping = sb->s_bdev->bd_mapping;
fs/ntfs3/fsntfs.c
1191
u32 blocksize = sb->s_blocksize;
fs/ntfs3/fsntfs.c
1235
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1252
bh = ntfs_bread(sb, block);
fs/ntfs3/fsntfs.c
1337
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1338
u32 blocksize = sb->s_blocksize;
fs/ntfs3/fsntfs.c
1362
sector_t block = lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
1378
bh = sb_getblk(sb, block);
fs/ntfs3/fsntfs.c
1391
bh = ntfs_bread(sb, block);
fs/ntfs3/fsntfs.c
1436
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1437
u32 block_size = sb->s_blocksize;
fs/ntfs3/fsntfs.c
1515
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1516
struct address_space *mapping = sb->s_bdev->bd_mapping;
fs/ntfs3/fsntfs.c
1601
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1602
struct block_device *bdev = sb->s_bdev;
fs/ntfs3/fsntfs.c
1689
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1690
struct inode *inode = new_inode(sb);
fs/ntfs3/fsntfs.c
1864
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
1884
inode = ntfs_iget5(sb, &ref, &NAME_SECURE);
fs/ntfs3/fsntfs.c
1887
ntfs_err(sb, "Failed to load $Secure (%d).", err);
fs/ntfs3/fsntfs.c
1905
ntfs_err(sb, "$Secure::$SDH is corrupted.");
fs/ntfs3/fsntfs.c
1912
ntfs_err(sb, "Failed to initialize $Secure::$SDH (%d).", err);
fs/ntfs3/fsntfs.c
1925
ntfs_err(sb, "$Secure::$SII is corrupted.");
fs/ntfs3/fsntfs.c
1932
ntfs_err(sb, "Failed to initialize $Secure::$SII (%d).", err);
fs/ntfs3/fsntfs.c
212
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
217
ntfs_notice(sb, "Skip $Extend 'cause NTFS version");
fs/ntfs3/fsntfs.c
224
inode = ntfs_iget5(sb, &ref, &NAME_EXTEND);
fs/ntfs3/fsntfs.c
227
ntfs_err(sb, "Failed to load $Extend (%d).", err);
fs/ntfs3/fsntfs.c
2476
ntfs_unmap_meta(sbi->sb, lcn, len);
fs/ntfs3/fsntfs.c
2670
ntfs_warn(sbi->sb, "new label is too long");
fs/ntfs3/fsntfs.c
279
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
286
ntfs_err(sb, "\x24LogFile is large than 4G.");
fs/ntfs3/fsntfs.c
297
inode = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
307
inode = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
315
ntfs_err(sb, "Failed to load $MFT.");
fs/ntfs3/fsntfs.c
329
sync_blockdev(sb->s_bdev);
fs/ntfs3/fsntfs.c
330
invalidate_bdev(sb->s_bdev);
fs/ntfs3/fsntfs.c
337
if (sb_rdonly(sb) || !initialized)
fs/ntfs3/fsntfs.c
358
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
375
ntfs_err(sbi->sb, "no free space to extend mft");
fs/ntfs3/fsntfs.c
437
ntfs_unmap_meta(sb, alcn, alen);
fs/ntfs3/fsntfs.c
586
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
642
i = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/fsntfs.c
646
sb,
fs/ntfs3/fsntfs.c
871
struct super_block *sb = sbi->sb;
fs/ntfs3/fsntfs.c
878
if (!sb || !(sbi->flags & NTFS_FLAGS_MFTMIRR) ||
fs/ntfs3/fsntfs.c
879
unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/fsntfs.c
882
blocksize = sb->s_blocksize;
fs/ntfs3/fsntfs.c
884
block1 = sbi->mft.lbo >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
885
block2 = sbi->mft.lbo2 >> sb->s_blocksize_bits;
fs/ntfs3/fsntfs.c
890
bh1 = sb_bread(sb, block1++);
fs/ntfs3/fsntfs.c
894
bh2 = sb_getblk(sb, block2++);
fs/ntfs3/fsntfs.c
958
if (sbi->volume.real_dirty || sb_rdonly(sbi->sb))
fs/ntfs3/fsntfs.c
988
ntfs_notice(sbi->sb, "Mark volume as dirty due to NTFS errors");
fs/ntfs3/index.c
213
struct super_block *sb;
fs/ntfs3/index.c
258
sb = sbi->sb;
fs/ntfs3/index.c
259
blocksize = sb->s_blocksize;
fs/ntfs3/index.c
365
struct super_block *sb = sbi->sb;
fs/ntfs3/index.c
368
u32 nbits = sb->s_blocksize * 8;
fs/ntfs3/index.c
369
u32 blocksize = sb->s_blocksize;
fs/ntfs3/index.c
372
sector_t eblock = bytes_to_block(sb, data_size);
fs/ntfs3/index.c
374
sector_t blk = (vbo & sbi->cluster_mask) >> sb->s_blocksize_bits;
fs/ntfs3/index.c
375
sector_t vblock = vbo >> sb->s_blocksize_bits;
fs/ntfs3/index.c
416
bh = ntfs_bread(sb, block + blk);
fs/ntfs3/index.c
420
vbo = (u64)vblock << sb->s_blocksize_bits;
fs/ntfs3/inode.c
1242
struct super_block *sb = dir->i_sb;
fs/ntfs3/inode.c
1243
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/inode.c
1351
if (unlikely(ntfs3_forced_shutdown(sb))) {
fs/ntfs3/inode.c
1649
inode->i_fop = unlikely(is_legacy_ntfs(sb)) ?
fs/ntfs3/inode.c
1660
inode->i_fop = unlikely(is_legacy_ntfs(sb)) ?
fs/ntfs3/inode.c
1672
if (!S_ISLNK(mode) && (sb->s_flags & SB_POSIXACL)) {
fs/ntfs3/inode.c
1877
static int ntfs_translate_junction(const struct super_block *sb,
fs/ntfs3/inode.c
1895
ntfs_err(sb, "Error getting link path");
fs/ntfs3/inode.c
1912
ntfs_err(sb,
fs/ntfs3/inode.c
1929
ntfs_err(sb, "Link target (%s) missing drive separator",
fs/ntfs3/inode.c
1940
ntfs_err(sb, "Link target %s too large for buffer (%d <= %d)",
fs/ntfs3/inode.c
1950
ntfs_err(sb, "Target path %s too large for buffer (%d <= %d)",
fs/ntfs3/inode.c
1970
struct super_block *sb = inode->i_sb;
fs/ntfs3/inode.c
1971
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/inode.c
2111
err = ntfs_translate_junction(sb, link_de, buffer, err, buflen);
fs/ntfs3/inode.c
297
mode = sb->s_root ?
fs/ntfs3/inode.c
30
struct super_block *sb = inode->i_sb;
fs/ntfs3/inode.c
31
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/inode.c
446
inode->i_fop = unlikely(is_legacy_ntfs(sb)) ?
fs/ntfs3/inode.c
458
inode->i_fop = unlikely(is_legacy_ntfs(sb)) ?
fs/ntfs3/inode.c
494
if (ino == MFT_REC_MFT && !sb->s_root)
fs/ntfs3/inode.c
502
if (ino == MFT_REC_MFT && !sb->s_root)
fs/ntfs3/inode.c
529
struct inode *ntfs_iget5(struct super_block *sb, const struct MFT_REF *ref,
fs/ntfs3/inode.c
534
inode = iget5_locked(sb, ino_get(ref), ntfs_test_inode, ntfs_set_inode,
fs/ntfs3/inode.c
552
ntfs_set_state(sb->s_fs_info, NTFS_DIRTY_ERROR);
fs/ntfs3/inode.c
59
if (!sbi->mft.ni && ino == MFT_REC_MFT && !sb->s_root) {
fs/ntfs3/inode.c
726
struct super_block *sb = inode->i_sb;
fs/ntfs3/inode.c
727
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/inode.c
790
if (unlikely(ntfs3_forced_shutdown(sbi->sb)))
fs/ntfs3/inode.c
82
ntfs_err(sb, "MFT: r=%lx, expect seq=%x instead of %x!", ino,
fs/ntfs3/inode.c
87
ntfs_err(sb, "Inode r=%x is not in use!", (u32)ino);
fs/ntfs3/namei.c
251
struct super_block *sb = dir->i_sb;
fs/ntfs3/namei.c
252
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/namei.c
275
if (unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/ntfs_fs.h
1056
static inline struct ntfs_sb_info *ntfs_sb(struct super_block *sb)
fs/ntfs3/ntfs_fs.h
1058
return sb->s_fs_info;
fs/ntfs3/ntfs_fs.h
1061
static inline int ntfs3_forced_shutdown(struct super_block *sb)
fs/ntfs3/ntfs_fs.h
1063
return test_bit(NTFS_FLAGS_SHUTDOWN_BIT, &ntfs_sb(sb)->flags);
fs/ntfs3/ntfs_fs.h
1107
static inline u64 ntfs_up_block(const struct super_block *sb, u64 size)
fs/ntfs3/ntfs_fs.h
1109
return (size + sb->s_blocksize - 1) & ~(u64)(sb->s_blocksize - 1);
fs/ntfs3/ntfs_fs.h
1117
static inline u64 bytes_to_block(const struct super_block *sb, u64 size)
fs/ntfs3/ntfs_fs.h
1119
return (size + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/ntfs3/ntfs_fs.h
1254
bool is_legacy_ntfs(struct super_block *sb);
fs/ntfs3/ntfs_fs.h
1256
static inline bool is_legacy_ntfs(struct super_block *sb)
fs/ntfs3/ntfs_fs.h
147
struct super_block *sb;
fs/ntfs3/ntfs_fs.h
215
struct super_block *sb;
fs/ntfs3/ntfs_fs.h
630
struct buffer_head *ntfs_bread(struct super_block *sb, sector_t block);
fs/ntfs3/ntfs_fs.h
655
int ntfs_sb_write(struct super_block *sb, u64 lbo, size_t bytes,
fs/ntfs3/ntfs_fs.h
774
struct inode *ntfs_iget5(struct super_block *sb, const struct MFT_REF *ref,
fs/ntfs3/ntfs_fs.h
894
void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len);
fs/ntfs3/ntfs_fs.h
905
int wnd_init(struct wnd_bitmap *wnd, struct super_block *sb, size_t nbits);
fs/ntfs3/ntfs_fs.h
970
return !!sbi->sb->s_root;
fs/ntfs3/record.c
184
ntfs_err(sbi->sb, "mft corrupted");
fs/ntfs3/run.c
1054
sbi->sb,
fs/ntfs3/run.c
1120
if (!sbi->used.bitmap.sb || !run || run == RUN_DEALLOCATE)
fs/ntfs3/super.c
1001
ntfs_err(sb, "%s signature is not NTFS.", hint);
fs/ntfs3/super.c
101
printk("%c%cntfs3(%s): %pV\n", KERN_SOH_ASCII, level, sb->s_id, &vaf);
fs/ntfs3/super.c
1014
ntfs_err(sb, "%s: invalid bytes per sector %u.", hint,
fs/ntfs3/super.c
1022
ntfs_err(sb, "%s: invalid sectors per cluster %u.", hint,
fs/ntfs3/super.c
1038
sb,
fs/ntfs3/super.c
1049
ntfs_err(sb, "%s: invalid record size %d.", hint,
fs/ntfs3/super.c
1060
ntfs_err(sb, "%s: invalid bytes per MFT record %u (%d).", hint,
fs/ntfs3/super.c
1066
ntfs_err(sb, "Unsupported bytes per MFT record %u.",
fs/ntfs3/super.c
1076
ntfs_err(sb, "%s: invalid index size %d.", hint,
fs/ntfs3/super.c
1083
ntfs_err(sb, "%s: invalid bytes per index %u(%d).", hint,
fs/ntfs3/super.c
1089
ntfs_err(sb, "%s: unsupported bytes per index %u.", hint,
fs/ntfs3/super.c
1105
sb,
fs/ntfs3/super.c
1117
ntfs_err(sb, "%s: invalid bytes per cluster (%u).", hint,
fs/ntfs3/super.c
1126
sb,
fs/ntfs3/super.c
1145
sb,
fs/ntfs3/super.c
1148
sb->s_flags |= SB_RDONLY;
fs/ntfs3/super.c
1156
sb,
fs/ntfs3/super.c
116
struct super_block *sb = inode->i_sb;
fs/ntfs3/super.c
117
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
1184
sb_set_blocksize(sb, min_t(u32, sbi->cluster_size, PAGE_SIZE));
fs/ntfs3/super.c
1186
sbi->block_mask = sb->s_blocksize - 1;
fs/ntfs3/super.c
1187
sbi->blocks_per_cluster = sbi->cluster_size >> sb->s_blocksize_bits;
fs/ntfs3/super.c
1188
sbi->volume.blocks = sbi->volume.size >> sb->s_blocksize_bits;
fs/ntfs3/super.c
1197
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/ntfs3/super.c
1201
sb->s_maxbytes = 0xFFFFFFFFull << cluster_bits;
fs/ntfs3/super.c
1213
if (bh->b_blocknr && !sb_rdonly(sb)) {
fs/ntfs3/super.c
1235
sb_set_blocksize(sb, block_size);
fs/ntfs3/super.c
1248
static int ntfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ntfs3/super.c
1251
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
1252
struct block_device *bdev = sb->s_bdev;
fs/ntfs3/super.c
1265
bool ro = sb_rdonly(sb);
fs/ntfs3/super.c
1270
sbi->sb = sb;
fs/ntfs3/super.c
1288
sb->s_flags |= SB_NODIRATIME;
fs/ntfs3/super.c
1289
sb->s_magic = 0x7366746e; // "ntfs"
fs/ntfs3/super.c
1290
sb->s_op = &ntfs_sops;
fs/ntfs3/super.c
1291
sb->s_export_op = &ntfs_export_ops;
fs/ntfs3/super.c
1292
sb->s_time_gran = NTFS_TIME_GRAN; // 100 nsec
fs/ntfs3/super.c
1293
sb->s_xattr = ntfs_xattr_handlers;
fs/ntfs3/super.c
1294
set_default_d_op(sb, options->nocase ? &ntfs_dentry_ops : NULL);
fs/ntfs3/super.c
1311
err = ntfs_init_from_boot(sb, bdev_logical_block_size(bdev),
fs/ntfs3/super.c
1322
inode = ntfs_iget5(sb, &ref, &NAME_VOLUME);
fs/ntfs3/super.c
1325
ntfs_err(sb, "Failed to load $Volume (%d).", err);
fs/ntfs3/super.c
1353
ntfs_err(sb, "$Volume is corrupted.");
fs/ntfs3/super.c
1364
ntfs_info(sb, "It is recommended to use chkdsk.");
fs/ntfs3/super.c
1370
inode = ntfs_iget5(sb, &ref, &NAME_MIRROR);
fs/ntfs3/super.c
1373
ntfs_err(sb, "Failed to load $MFTMirr (%d).", err);
fs/ntfs3/super.c
1385
inode = ntfs_iget5(sb, &ref, &NAME_LOGFILE);
fs/ntfs3/super.c
1388
ntfs_err(sb, "Failed to load \x24LogFile (%d).", err);
fs/ntfs3/super.c
1401
ntfs_warn(sb, "failed to replay log file. Can't mount rw!");
fs/ntfs3/super.c
1407
ntfs_warn(sb, "volume is dirty and \"force\" flag is not set!");
fs/ntfs3/super.c
1416
inode = ntfs_iget5(sb, &ref, &NAME_MFT);
fs/ntfs3/super.c
1419
ntfs_err(sb, "Failed to load $MFT (%d).", err);
fs/ntfs3/super.c
1429
err = wnd_init(&sbi->mft.bitmap, sb, tt);
fs/ntfs3/super.c
1435
ntfs_err(sb, "Failed to load $MFT's subrecords (%d).", err);
fs/ntfs3/super.c
1444
inode = ntfs_iget5(sb, &ref, &NAME_BITMAP);
fs/ntfs3/super.c
1447
ntfs_err(sb, "Failed to load $Bitmap (%d).", err);
fs/ntfs3/super.c
1461
ntfs_err(sb, "$Bitmap is corrupted.");
fs/ntfs3/super.c
1466
err = wnd_init(&sbi->used.bitmap, sb, tt);
fs/ntfs3/super.c
1468
ntfs_err(sb, "Failed to initialize $Bitmap (%d).", err);
fs/ntfs3/super.c
1477
ntfs_err(sb, "Failed to initialize MFT zone (%d).", err);
fs/ntfs3/super.c
1484
inode = ntfs_iget5(sb, &ref, &NAME_BADCLUS);
fs/ntfs3/super.c
1487
ntfs_err(sb, "Failed to load $BadClus (%d).", err);
fs/ntfs3/super.c
1513
ntfs_notice(sb,
fs/ntfs3/super.c
1522
inode = ntfs_iget5(sb, &ref, &NAME_ATTRDEF);
fs/ntfs3/super.c
1525
ntfs_err(sb, "Failed to load $AttrDef (%d)", err);
fs/ntfs3/super.c
1535
ntfs_err(sb, "Looks like $AttrDef is corrupted (size=%llu).",
fs/ntfs3/super.c
1551
ntfs_err(sb, "Failed to read $AttrDef (%d).", err);
fs/ntfs3/super.c
1556
ntfs_err(sb, "$AttrDef is corrupted.");
fs/ntfs3/super.c
157
sb->s_id, inode->i_ino, name ? name : "", &vaf);
fs/ntfs3/super.c
1586
inode = ntfs_iget5(sb, &ref, &NAME_UPCASE);
fs/ntfs3/super.c
1589
ntfs_err(sb, "Failed to load $UpCase (%d).", err);
fs/ntfs3/super.c
1595
ntfs_err(sb, "$UpCase is corrupted.");
fs/ntfs3/super.c
1602
ntfs_err(sb, "Failed to read $UpCase (%d).", err);
fs/ntfs3/super.c
1627
ntfs_err(sb, "Failed to initialize $Secure (%d).", err);
fs/ntfs3/super.c
1634
ntfs_warn(sb, "Failed to initialize $Extend.");
fs/ntfs3/super.c
1641
ntfs_warn(sb, "Failed to initialize $Extend/$Reparse.");
fs/ntfs3/super.c
1648
ntfs_warn(sb, "Failed to initialize $Extend/$ObjId.");
fs/ntfs3/super.c
1657
inode = ntfs_iget5(sb, &ref, &NAME_ROOT);
fs/ntfs3/super.c
1660
ntfs_err(sb, "Failed to load root (%d).", err);
fs/ntfs3/super.c
1669
ntfs_err(sb, "Failed to load root (%d).", err);
fs/ntfs3/super.c
1673
sb->s_root = d_make_root(inode);
fs/ntfs3/super.c
1674
if (!sb->s_root) {
fs/ntfs3/super.c
1684
struct buffer_head *bh0 = sb_getblk(sb, 0);
fs/ntfs3/super.c
1693
ntfs_warn(sb, "primary boot is updated");
fs/ntfs3/super.c
1700
ntfs_create_procdir(sb);
fs/ntfs3/super.c
1702
if (is_legacy_ntfs(sb))
fs/ntfs3/super.c
1703
sb->s_flags |= SB_RDONLY;
fs/ntfs3/super.c
1720
void ntfs_unmap_meta(struct super_block *sb, CLST lcn, CLST len)
fs/ntfs3/super.c
1722
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
1723
struct block_device *bdev = sb->s_bdev;
fs/ntfs3/super.c
1728
<< (PAGE_SHIFT - sb->s_blocksize_bits);
fs/ntfs3/super.c
1753
struct super_block *sb;
fs/ntfs3/super.c
1773
sb = sbi->sb;
fs/ntfs3/super.c
1777
err = blkdev_issue_discard(sb->s_bdev, start >> 9, (end - start) >> 9,
fs/ntfs3/super.c
1885
static void ntfs3_kill_sb(struct super_block *sb)
fs/ntfs3/super.c
1887
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
1889
kill_block_super(sb);
fs/ntfs3/super.c
1938
bool is_legacy_ntfs(struct super_block *sb)
fs/ntfs3/super.c
1940
return sb->s_type == &ntfs_legacy_fs_type;
fs/ntfs3/super.c
432
struct super_block *sb = fc->root->d_sb;
fs/ntfs3/super.c
433
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
438
if (is_legacy_ntfs(sb)) {
fs/ntfs3/super.c
443
ro_rw = sb_rdonly(sb) && !(fc->sb_flags & SB_RDONLY);
fs/ntfs3/super.c
470
sync_filesystem(sb);
fs/ntfs3/super.c
494
struct super_block *sb = m->private;
fs/ntfs3/super.c
495
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
516
struct super_block *sb = m->private;
fs/ntfs3/super.c
517
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
529
struct super_block *sb = pde_data(file_inode(file));
fs/ntfs3/super.c
533
if (sb_rdonly(sb))
fs/ntfs3/super.c
548
err = ntfs_set_label(sb->s_fs_info, label, ret);
fs/ntfs3/super.c
551
ntfs_err(sb, "failed (%d) to write label", err);
fs/ntfs3/super.c
583
static void ntfs_create_procdir(struct super_block *sb)
fs/ntfs3/super.c
590
e = proc_mkdir(sb->s_id, proc_info_root);
fs/ntfs3/super.c
592
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
594
proc_create_data("volinfo", 0444, e, &ntfs3_volinfo_fops, sb);
fs/ntfs3/super.c
595
proc_create_data("label", 0644, e, &ntfs3_label_fops, sb);
fs/ntfs3/super.c
600
static void ntfs_remove_procdir(struct super_block *sb)
fs/ntfs3/super.c
602
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
609
remove_proc_entry(sb->s_id, proc_info_root);
fs/ntfs3/super.c
627
static void ntfs_create_procdir(struct super_block *sb){}
fs/ntfs3/super.c
628
static void ntfs_remove_procdir(struct super_block *sb){}
fs/ntfs3/super.c
636
static struct inode *ntfs_alloc_inode(struct super_block *sb)
fs/ntfs3/super.c
638
struct ntfs_inode *ni = alloc_inode_sb(sb, ntfs_inode_cachep, GFP_NOFS);
fs/ntfs3/super.c
717
static void ntfs_put_super(struct super_block *sb)
fs/ntfs3/super.c
719
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
721
ntfs_remove_procdir(sb);
fs/ntfs3/super.c
736
struct super_block *sb = dentry->d_sb;
fs/ntfs3/super.c
737
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
741
buf->f_type = sb->s_magic;
fs/ntfs3/super.c
762
struct super_block *sb = root->d_sb;
fs/ntfs3/super.c
763
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
789
if (sb->s_flags & SB_POSIXACL)
fs/ntfs3/super.c
808
static void ntfs_shutdown(struct super_block *sb)
fs/ntfs3/super.c
810
set_bit(NTFS_FLAGS_SHUTDOWN_BIT, &ntfs_sb(sb)->flags);
fs/ntfs3/super.c
816
static int ntfs_sync_fs(struct super_block *sb, int wait)
fs/ntfs3/super.c
819
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
823
if (unlikely(ntfs3_forced_shutdown(sb)))
fs/ntfs3/super.c
85
void ntfs_printk(const struct super_block *sb, const char *fmt, ...)
fs/ntfs3/super.c
856
sync_blockdev(sb->s_bdev);
fs/ntfs3/super.c
857
blkdev_issue_flush(sb->s_bdev);
fs/ntfs3/super.c
875
static struct inode *ntfs_export_get_inode(struct super_block *sb, u64 ino,
fs/ntfs3/super.c
889
inode = ntfs_iget5(sb, &ref, NULL);
fs/ntfs3/super.c
898
static struct dentry *ntfs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/ntfs3/super.c
90
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
901
return generic_fh_to_dentry(sb, fid, fh_len, fh_type,
fs/ntfs3/super.c
905
static struct dentry *ntfs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/ntfs3/super.c
908
return generic_fh_to_parent(sb, fid, fh_len, fh_type,
fs/ntfs3/super.c
961
static int ntfs_init_from_boot(struct super_block *sb, u32 sector_size,
fs/ntfs3/super.c
964
struct ntfs_sb_info *sbi = sb->s_fs_info;
fs/ntfs3/super.c
983
if (!sb_min_blocksize(sb, PAGE_SIZE)) {
fs/ntfs3/super.c
988
bh = ntfs_bread(sb, boot_block);
fs/ntfs3/xattr.c
726
struct super_block *sb = inode->i_sb;
fs/ntfs3/xattr.c
728
if (!(sb->s_flags & SB_POSIXACL))
fs/ocfs2/alloc.c
1022
bhs[i] = sb_getblk(osb->sb, first_blkno);
fs/ocfs2/alloc.c
1038
memset(bhs[i]->b_data, 0, osb->sb->s_blocksize);
fs/ocfs2/alloc.c
1049
cpu_to_le16(ocfs2_extent_recs_per_eb(osb->sb));
fs/ocfs2/alloc.c
2224
int ocfs2_find_cpos_for_left_leaf(struct super_block *sb,
fs/ocfs2/alloc.c
2278
ocfs2_error(sb, "Invalid extent tree at extent block %llu\n",
fs/ocfs2/alloc.c
2386
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
2397
ret = ocfs2_find_cpos_for_left_leaf(sb, right_path, &cpos);
fs/ocfs2/alloc.c
2523
ret = ocfs2_find_cpos_for_left_leaf(sb, right_path, &cpos);
fs/ocfs2/alloc.c
2819
int ocfs2_find_cpos_for_right_leaf(struct super_block *sb,
fs/ocfs2/alloc.c
2874
ocfs2_error(sb, "Invalid extent tree at extent block %llu\n",
fs/ocfs2/alloc.c
2924
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
2931
ret = ocfs2_find_cpos_for_right_leaf(sb, path, &right_cpos);
fs/ocfs2/alloc.c
298
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
300
ocfs2_clusters_for_bytes(sb, OCFS2_MAX_XATTR_TREE_LEAF_SIZE);
fs/ocfs2/alloc.c
3015
ret = ocfs2_find_cpos_for_right_leaf(sb, left_path,
fs/ocfs2/alloc.c
3838
static void ocfs2_subtract_from_rec(struct super_block *sb,
fs/ocfs2/alloc.c
3845
len_blocks = ocfs2_clusters_to_blocks(sb,
fs/ocfs2/alloc.c
4377
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
4382
status = ocfs2_find_cpos_for_left_leaf(sb, path, &left_cpos);
fs/ocfs2/alloc.c
4405
status = ocfs2_error(sb,
fs/ocfs2/alloc.c
4435
status = ocfs2_find_cpos_for_right_leaf(sb, path, &right_cpos);
fs/ocfs2/alloc.c
4459
status = ocfs2_error(sb,
fs/ocfs2/alloc.c
4853
block = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/alloc.c
4885
ocfs2_clusters_to_blocks(osb->sb, bit_off),
fs/ocfs2/alloc.c
4896
static void ocfs2_make_right_split_rec(struct super_block *sb,
fs/ocfs2/alloc.c
4911
ocfs2_clusters_to_blocks(sb, cpos - rec_cpos));
fs/ocfs2/alloc.c
51
ocfs2_extent_rec_contig(struct super_block *sb,
fs/ocfs2/alloc.c
5171
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
5172
u64 start_blkno = ocfs2_clusters_to_blocks(sb, phys);
fs/ocfs2/alloc.c
5194
ocfs2_error(sb,
fs/ocfs2/alloc.c
5364
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/alloc.c
5418
ret = ocfs2_find_cpos_for_left_leaf(sb, path, &left_cpos);
fs/ocfs2/alloc.c
5483
le64_add_cpu(&rec->e_blkno, ocfs2_clusters_to_blocks(sb, len));
fs/ocfs2/alloc.c
5761
ocfs2_remove_extent_credits(osb->sb) + credits);
fs/ocfs2/alloc.c
5792
ocfs2_blocks_to_clusters(osb->sb,
fs/ocfs2/alloc.c
5865
start_cluster = ocfs2_blocks_to_clusters(osb->sb, start_blk);
fs/ocfs2/alloc.c
5876
mlog_bug_on_msg(tl_count > ocfs2_truncate_recs_per_inode(osb->sb) ||
fs/ocfs2/alloc.c
5881
ocfs2_truncate_recs_per_inode(osb->sb),
fs/ocfs2/alloc.c
6189
if (unlikely(tl_count > ocfs2_truncate_recs_per_inode(osb->sb) ||
fs/ocfs2/alloc.c
6255
ocfs2_compute_meta_ecc(osb->sb, tl_bh->b_data, &di->i_check);
fs/ocfs2/alloc.c
6317
start_blk = ocfs2_clusters_to_blocks(osb->sb, start_cluster);
fs/ocfs2/alloc.c
6722
new_eb_bh[i] = sb_getblk(osb->sb, bf->free_blk);
fs/ocfs2/alloc.c
6743
memset(new_eb_bh[i]->b_data, 0, osb->sb->s_blocksize);
fs/ocfs2/alloc.c
6756
cpu_to_le16(ocfs2_extent_recs_per_eb(osb->sb));
fs/ocfs2/alloc.c
6879
struct super_block *sb = inode->i_sb;
fs/ocfs2/alloc.c
6881
BUG_ON(!ocfs2_sparse_alloc(OCFS2_SB(sb)));
fs/ocfs2/alloc.c
6945
struct super_block *sb = inode->i_sb;
fs/ocfs2/alloc.c
6947
BUG_ON(start >> OCFS2_SB(sb)->s_clustersize_bits !=
fs/ocfs2/alloc.c
6948
(end - 1) >> OCFS2_SB(sb)->s_clustersize_bits);
fs/ocfs2/alloc.c
6968
struct super_block *sb = inode->i_sb;
fs/ocfs2/alloc.c
6974
if (!ocfs2_sparse_alloc(OCFS2_SB(sb)))
fs/ocfs2/alloc.c
6986
folios = kzalloc_objs(struct folio *, ocfs2_pages_per_cluster(sb),
fs/ocfs2/alloc.c
6995
range_start >> sb->s_blocksize_bits,
fs/ocfs2/alloc.c
7107
ocfs2_inline_to_extents_credits(osb->sb));
fs/ocfs2/alloc.c
7127
ocfs2_clusters_to_bytes(osb->sb, 1));
fs/ocfs2/alloc.c
7204
ocfs2_clusters_to_bytes(osb->sb, 1));
fs/ocfs2/alloc.c
7214
ocfs2_clusters_to_blocks(osb->sb, bit_off),
fs/ocfs2/alloc.c
7252
new_highest_cpos = ocfs2_clusters_for_bytes(osb->sb,
fs/ocfs2/alloc.c
7482
static int ocfs2_trim_extent(struct super_block *sb,
fs/ocfs2/alloc.c
7487
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/alloc.c
7489
bcount = ocfs2_clusters_to_blocks(sb, count);
fs/ocfs2/alloc.c
7490
discard = ocfs2_clusters_to_blocks(sb, start);
fs/ocfs2/alloc.c
7503
trace_ocfs2_trim_extent(sb, (unsigned long long)discard, bcount);
fs/ocfs2/alloc.c
7505
return sb_issue_discard(sb, discard, bcount, GFP_NOFS, 0);
fs/ocfs2/alloc.c
7508
static int ocfs2_trim_group(struct super_block *sb,
fs/ocfs2/alloc.c
7528
ret = ocfs2_trim_extent(sb, gd, group,
fs/ocfs2/alloc.c
7554
int ocfs2_trim_mainbm(struct super_block *sb, struct fstrim_range *range)
fs/ocfs2/alloc.c
7556
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/alloc.c
7570
if (minlen >= osb->bitmap_cpg || range->len < sb->s_blocksize)
fs/ocfs2/alloc.c
7614
first_bit = start - ocfs2_blocks_to_clusters(sb,
fs/ocfs2/alloc.c
7636
cnt = ocfs2_trim_group(sb, gd, group,
fs/ocfs2/alloc.c
7650
group = ocfs2_clusters_to_blocks(sb, osb->bitmap_cpg);
fs/ocfs2/alloc.c
7652
group += ocfs2_clusters_to_blocks(sb, osb->bitmap_cpg);
fs/ocfs2/alloc.c
7677
int ocfs2_trim_fs(struct super_block *sb, struct fstrim_range *range)
fs/ocfs2/alloc.c
7680
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/alloc.c
7723
ret = ocfs2_trim_mainbm(sb, range);
fs/ocfs2/alloc.c
795
static int ocfs2_block_extent_contig(struct super_block *sb,
fs/ocfs2/alloc.c
801
blk_end += ocfs2_clusters_to_blocks(sb,
fs/ocfs2/alloc.c
819
ocfs2_extent_rec_contig(struct super_block *sb,
fs/ocfs2/alloc.c
834
ocfs2_block_extent_contig(sb, ext, blkno))
fs/ocfs2/alloc.c
839
ocfs2_block_extent_contig(sb, insert_rec, blkno))
fs/ocfs2/alloc.c
877
static int ocfs2_validate_extent_block(struct super_block *sb,
fs/ocfs2/alloc.c
893
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &eb->h_check);
fs/ocfs2/alloc.c
905
rc = ocfs2_error(sb,
fs/ocfs2/alloc.c
913
rc = ocfs2_error(sb,
fs/ocfs2/alloc.c
920
if (le32_to_cpu(eb->h_fs_generation) != OCFS2_SB(sb)->fs_generation)
fs/ocfs2/alloc.c
921
rc = ocfs2_error(sb,
fs/ocfs2/alloc.h
228
int ocfs2_trim_fs(struct super_block *sb, struct fstrim_range *range);
fs/ocfs2/alloc.h
301
int ocfs2_find_cpos_for_right_leaf(struct super_block *sb,
fs/ocfs2/alloc.h
303
int ocfs2_find_cpos_for_left_leaf(struct super_block *sb,
fs/ocfs2/aops.c
1759
ocfs2_clusters_to_bytes(osb->sb, clusters_to_alloc));
fs/ocfs2/aops.c
1815
ocfs2_clusters_to_bytes(osb->sb, clusters_to_alloc));
fs/ocfs2/aops.c
82
buffer_cache_bh = sb_getblk(osb->sb, blkno);
fs/ocfs2/blockcheck.c
564
void ocfs2_compute_meta_ecc(struct super_block *sb, void *data,
fs/ocfs2/blockcheck.c
567
if (ocfs2_meta_ecc(OCFS2_SB(sb)))
fs/ocfs2/blockcheck.c
568
ocfs2_block_check_compute(data, sb->s_blocksize, bc);
fs/ocfs2/blockcheck.c
571
int ocfs2_validate_meta_ecc(struct super_block *sb, void *data,
fs/ocfs2/blockcheck.c
575
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/blockcheck.c
578
rc = ocfs2_block_check_validate(data, sb->s_blocksize, bc,
fs/ocfs2/blockcheck.c
584
void ocfs2_compute_meta_ecc_bhs(struct super_block *sb,
fs/ocfs2/blockcheck.c
588
if (ocfs2_meta_ecc(OCFS2_SB(sb)))
fs/ocfs2/blockcheck.c
592
int ocfs2_validate_meta_ecc_bhs(struct super_block *sb,
fs/ocfs2/blockcheck.c
597
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/blockcheck.h
30
void ocfs2_compute_meta_ecc(struct super_block *sb, void *data,
fs/ocfs2/blockcheck.h
32
int ocfs2_validate_meta_ecc(struct super_block *sb, void *data,
fs/ocfs2/blockcheck.h
34
void ocfs2_compute_meta_ecc_bhs(struct super_block *sb,
fs/ocfs2/blockcheck.h
37
int ocfs2_validate_meta_ecc_bhs(struct super_block *sb,
fs/ocfs2/buffer_head_io.c
109
bhs[i] = sb_getblk(osb->sb, block++);
fs/ocfs2/buffer_head_io.c
195
int (*validate)(struct super_block *sb,
fs/ocfs2/buffer_head_io.c
201
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/buffer_head_io.c
236
bhs[i] = sb_getblk(sb, block++);
fs/ocfs2/buffer_head_io.c
382
status = validate(sb, bh);
fs/ocfs2/buffer_head_io.c
405
static void ocfs2_check_super_or_backup(struct super_block *sb,
fs/ocfs2/buffer_head_io.c
415
backup_blkno = ocfs2_backup_super_blkno(sb, i);
fs/ocfs2/buffer_head_io.c
435
ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr);
fs/ocfs2/buffer_head_io.c
451
ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check);
fs/ocfs2/buffer_head_io.h
30
int (*validate)(struct super_block *sb,
fs/ocfs2/buffer_head_io.h
41
int (*validate)(struct super_block *sb,
fs/ocfs2/dir.c
1231
static int ocfs2_find_max_rec_len(struct super_block *sb,
fs/ocfs2/dir.c
1238
trailer = (char *)ocfs2_trailer_from_bh(dirblock_bh, sb);
fs/ocfs2/dir.c
1239
size = ocfs2_dir_trailer_blk_off(sb);
fs/ocfs2/dir.c
1638
struct super_block *sb = dir->i_sb;
fs/ocfs2/dir.c
1640
unsigned int size = sb->s_blocksize;
fs/ocfs2/dir.c
1859
struct super_block * sb = inode->i_sb;
fs/ocfs2/dir.c
1865
offset = ctx->pos & (sb->s_blocksize - 1);
fs/ocfs2/dir.c
1868
blk = ctx->pos >> sb->s_blocksize_bits;
fs/ocfs2/dir.c
1871
ctx->pos += sb->s_blocksize - offset;
fs/ocfs2/dir.c
1882
for (i = ra_sectors >> (sb->s_blocksize_bits - 9);
fs/ocfs2/dir.c
1898
for (i = 0; i < sb->s_blocksize && i < offset; ) {
fs/ocfs2/dir.c
1912
ctx->pos = (ctx->pos & ~(sb->s_blocksize - 1))
fs/ocfs2/dir.c
1918
&& offset < sb->s_blocksize) {
fs/ocfs2/dir.c
1921
sb->s_blocksize, offset)) {
fs/ocfs2/dir.c
1924
ctx->pos = (ctx->pos | (sb->s_blocksize - 1)) + 1;
fs/ocfs2/dir.c
2275
unsigned int size = osb->sb->s_blocksize;
fs/ocfs2/dir.c
2282
status = ocfs2_do_extend_dir(osb->sb, handle, inode, fe_bh,
fs/ocfs2/dir.c
2297
memset(new_bh->b_data, 0, osb->sb->s_blocksize);
fs/ocfs2/dir.c
2365
dx_root_bh = sb_getblk(osb->sb, dr_blkno);
fs/ocfs2/dir.c
2380
memset(dx_root, 0, osb->sb->s_blocksize);
fs/ocfs2/dir.c
2397
cpu_to_le16(ocfs2_dx_entries_per_root(osb->sb));
fs/ocfs2/dir.c
2400
cpu_to_le16(ocfs2_extent_recs_per_dx_root(osb->sb));
fs/ocfs2/dir.c
2438
bh = sb_getblk(osb->sb, start_blk + i);
fs/ocfs2/dir.c
2456
memset(dx_leaf, 0, osb->sb->s_blocksize);
fs/ocfs2/dir.c
2461
cpu_to_le16(ocfs2_dx_entries_per_leaf(osb->sb));
fs/ocfs2/dir.c
2508
phys_blkno = ocfs2_clusters_to_blocks(osb->sb, phys);
fs/ocfs2/dir.c
2547
static struct buffer_head **ocfs2_dx_dir_kmalloc_leaves(struct super_block *sb,
fs/ocfs2/dir.c
2550
int num_dx_leaves = ocfs2_clusters_to_blocks(sb, 1);
fs/ocfs2/dir.c
2770
struct super_block *sb = dir->i_sb;
fs/ocfs2/dir.c
2774
unsigned int new_size = sb->s_blocksize;
fs/ocfs2/dir.c
2779
new_size = ocfs2_dir_trailer_blk_off(sb);
fs/ocfs2/dir.c
2823
struct super_block *sb = dir->i_sb;
fs/ocfs2/dir.c
2825
credits = ocfs2_inline_to_extents_credits(sb);
fs/ocfs2/dir.c
2827
bytes = blocks_wanted << sb->s_blocksize_bits;
fs/ocfs2/dir.c
2843
alloc = ocfs2_clusters_for_bytes(sb, bytes);
fs/ocfs2/dir.c
2849
credits += ocfs2_add_dir_index_credits(sb);
fs/ocfs2/dir.c
2855
dx_leaves = ocfs2_dx_dir_kmalloc_leaves(sb,
fs/ocfs2/dir.c
2903
ocfs2_clusters_to_bytes(osb->sb, alloc + dx_alloc));
fs/ocfs2/dir.c
2944
dirdata_bh = sb_getblk(sb, blkno);
fs/ocfs2/dir.c
2962
sb->s_blocksize - i_size_read(dir));
fs/ocfs2/dir.c
3017
i_size_write(dir, sb->s_blocksize);
fs/ocfs2/dir.c
3020
di->i_size = cpu_to_le64(sb->s_blocksize);
fs/ocfs2/dir.c
3134
static int ocfs2_do_extend_dir(struct super_block *sb,
fs/ocfs2/dir.c
3147
extend = (i_size_read(dir) == ocfs2_clusters_to_bytes(sb, OCFS2_I(dir)->ip_clusters));
fs/ocfs2/dir.c
3154
ocfs2_clusters_to_bytes(sb, 1));
fs/ocfs2/dir.c
3159
status = ocfs2_add_inode_data(OCFS2_SB(sb), dir, &offset,
fs/ocfs2/dir.c
3169
v_blkno = ocfs2_blocks_for_bytes(sb, i_size_read(dir));
fs/ocfs2/dir.c
3176
*new_bh = sb_getblk(sb, p_blkno);
fs/ocfs2/dir.c
3185
dquot_free_space_nodirty(dir, ocfs2_clusters_to_bytes(sb, 1));
fs/ocfs2/dir.c
3216
struct super_block *sb = osb->sb;
fs/ocfs2/dir.c
3273
if (dir_i_size == ocfs2_clusters_to_bytes(sb, OCFS2_I(dir)->ip_clusters)) {
fs/ocfs2/dir.c
3303
credits = ocfs2_calc_extend_credits(sb, el);
fs/ocfs2/dir.c
3322
status = ocfs2_do_extend_dir(osb->sb, handle, dir, parent_fe_bh,
fs/ocfs2/dir.c
3337
memset(new_bh->b_data, 0, sb->s_blocksize);
fs/ocfs2/dir.c
3342
de->rec_len = cpu_to_le16(ocfs2_dir_trailer_blk_off(sb));
fs/ocfs2/dir.c
3355
de->rec_len = cpu_to_le16(sb->s_blocksize);
fs/ocfs2/dir.c
3394
struct super_block *sb = dir->i_sb;
fs/ocfs2/dir.c
3406
free_space = ocfs2_dir_trailer_blk_off(sb) - i_size_read(dir);
fs/ocfs2/dir.c
3447
ret = ocfs2_error(sb, "Directory entry (#%llu: size=%lld) "
fs/ocfs2/dir.c
3476
struct super_block *sb = dir->i_sb;
fs/ocfs2/dir.c
3488
if ((char *)de >= sb->s_blocksize + bh->b_data) {
fs/ocfs2/dir.c
3501
offset >> sb->s_blocksize_bits,
fs/ocfs2/dir.c
3727
int credits = ocfs2_clusters_to_blocks(osb->sb, 3);
fs/ocfs2/dir.c
3729
credits += ocfs2_calc_extend_credits(osb->sb, &dx_root->dr_list);
fs/ocfs2/dir.c
3730
credits += ocfs2_quota_trans_credits(osb->sb);
fs/ocfs2/dir.c
3780
orig_dx_leaves = ocfs2_dx_dir_kmalloc_leaves(osb->sb, &num_dx_leaves);
fs/ocfs2/dir.c
3787
new_dx_leaves = ocfs2_dx_dir_kmalloc_leaves(osb->sb, NULL);
fs/ocfs2/dir.c
3859
tmp_dx_leaf = kmalloc(osb->sb->s_blocksize, GFP_NOFS);
fs/ocfs2/dir.c
4075
dx_leaves = ocfs2_dx_dir_kmalloc_leaves(osb->sb, &num_dx_leaves);
fs/ocfs2/dir.c
4082
handle = ocfs2_start_trans(osb, ocfs2_calc_dxi_expand_credits(osb->sb));
fs/ocfs2/dir.c
4090
ocfs2_clusters_to_bytes(osb->sb, 1));
fs/ocfs2/dir.c
4138
cpu_to_le16(ocfs2_extent_recs_per_dx_root(osb->sb));
fs/ocfs2/dir.c
4141
osb->sb->s_blocksize -
fs/ocfs2/dir.c
437
static int ocfs2_validate_dir_block(struct super_block *sb,
fs/ocfs2/dir.c
442
ocfs2_trailer_from_bh(bh, sb);
fs/ocfs2/dir.c
462
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &trailer->db_check);
fs/ocfs2/dir.c
56
static int ocfs2_do_extend_dir(struct super_block *sb,
fs/ocfs2/dir.c
581
static int ocfs2_validate_dx_root(struct super_block *sb,
fs/ocfs2/dir.c
591
ret = ocfs2_validate_meta_ecc(sb, bh->b_data, &dx_root->dr_check);
fs/ocfs2/dir.c
600
ret = ocfs2_error(sb,
fs/ocfs2/dir.c
626
static int ocfs2_validate_dx_leaf(struct super_block *sb,
fs/ocfs2/dir.c
634
ret = ocfs2_validate_meta_ecc(sb, bh->b_data, &dx_leaf->dl_check);
fs/ocfs2/dir.c
643
ret = ocfs2_error(sb, "Dir Index Leaf has bad signature %.*s\n",
fs/ocfs2/dir.c
687
struct super_block *sb;
fs/ocfs2/dir.c
698
sb = dir->i_sb;
fs/ocfs2/dir.c
700
nblocks = i_size_read(dir) >> sb->s_blocksize_bits;
fs/ocfs2/dir.c
745
block << sb->s_blocksize_bits,
fs/ocfs2/dir.c
746
bh->b_data, sb->s_blocksize,
fs/ocfs2/dir.c
767
nblocks = i_size_read(dir) >> sb->s_blocksize_bits;
fs/ocfs2/dir.c
95
static inline unsigned int ocfs2_dir_trailer_blk_off(struct super_block *sb)
fs/ocfs2/dir.c
97
return sb->s_blocksize - sizeof(struct ocfs2_dir_block_trailer);
fs/ocfs2/dlmfs/dlmfs.c
279
static struct inode *dlmfs_alloc_inode(struct super_block *sb)
fs/ocfs2/dlmfs/dlmfs.c
283
ip = alloc_inode_sb(sb, dlmfs_inode_cache, GFP_NOFS);
fs/ocfs2/dlmfs/dlmfs.c
332
static struct inode *dlmfs_get_root_inode(struct super_block *sb)
fs/ocfs2/dlmfs/dlmfs.c
334
struct inode *inode = new_inode(sb);
fs/ocfs2/dlmfs/dlmfs.c
354
struct super_block *sb = parent->i_sb;
fs/ocfs2/dlmfs/dlmfs.c
355
struct inode * inode = new_inode(sb);
fs/ocfs2/dlmfs/dlmfs.c
508
static int dlmfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ocfs2/dlmfs/dlmfs.c
510
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/ocfs2/dlmfs/dlmfs.c
511
sb->s_blocksize = PAGE_SIZE;
fs/ocfs2/dlmfs/dlmfs.c
512
sb->s_blocksize_bits = PAGE_SHIFT;
fs/ocfs2/dlmfs/dlmfs.c
513
sb->s_magic = DLMFS_MAGIC;
fs/ocfs2/dlmfs/dlmfs.c
514
sb->s_op = &dlmfs_ops;
fs/ocfs2/dlmfs/dlmfs.c
515
sb->s_root = d_make_root(dlmfs_get_root_inode(sb));
fs/ocfs2/dlmfs/dlmfs.c
516
if (!sb->s_root)
fs/ocfs2/export.c
252
static struct dentry *ocfs2_fh_to_dentry(struct super_block *sb,
fs/ocfs2/export.c
263
return ocfs2_get_dentry(sb, &handle);
fs/ocfs2/export.c
266
static struct dentry *ocfs2_fh_to_parent(struct super_block *sb,
fs/ocfs2/export.c
277
return ocfs2_get_dentry(sb, &parent);
fs/ocfs2/export.c
34
static struct dentry *ocfs2_get_dentry(struct super_block *sb,
fs/ocfs2/export.c
38
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/export.c
43
trace_ocfs2_get_dentry_begin(sb, handle, (unsigned long long)blkno);
fs/ocfs2/export.c
50
inode = ocfs2_ilookup(sb, blkno);
fs/ocfs2/extent_map.c
525
static void ocfs2_relative_extent_offsets(struct super_block *sb,
fs/ocfs2/extent_map.c
533
*p_cluster = ocfs2_blocks_to_clusters(sb, le64_to_cpu(rec->e_blkno));
fs/ocfs2/extent_map.c
809
phys_bytes = le64_to_cpu(rec.e_blkno) << osb->sb->s_blocksize_bits;
fs/ocfs2/extent_map.c
976
int (*validate)(struct super_block *sb,
fs/ocfs2/extent_map.h
53
int (*validate)(struct super_block *sb,
fs/ocfs2/extent_map.h
62
int (*validate)(struct super_block *sb,
fs/ocfs2/file.c
1121
struct super_block *sb = inode->i_sb;
fs/ocfs2/file.c
1122
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/file.c
1238
&& OCFS2_HAS_RO_COMPAT_FEATURE(sb,
fs/ocfs2/file.c
1240
transfer_to[USRQUOTA] = dqget(sb, make_kqid_uid(attr->ia_uid));
fs/ocfs2/file.c
1248
&& OCFS2_HAS_RO_COMPAT_FEATURE(sb,
fs/ocfs2/file.c
1250
transfer_to[GRPQUOTA] = dqget(sb, make_kqid_gid(attr->ia_gid));
fs/ocfs2/file.c
1259
2 * ocfs2_quota_trans_credits(sb));
fs/ocfs2/file.c
1319
struct super_block *sb = path->dentry->d_sb;
fs/ocfs2/file.c
1320
struct ocfs2_super *osb = sb->s_fs_info;
fs/ocfs2/file.c
1557
struct super_block *sb = inode->i_sb;
fs/ocfs2/file.c
1558
u64 end = ocfs2_align_bytes_to_clusters(sb, start);
fs/ocfs2/file.c
1563
start_block = ocfs2_blocks_for_bytes(sb, start);
fs/ocfs2/file.c
1564
end_block = ocfs2_blocks_for_bytes(sb, end);
fs/ocfs2/file.c
1569
cluster = ocfs2_bytes_to_clusters(sb, start);
fs/ocfs2/file.c
1577
offset = start_block - ocfs2_clusters_to_blocks(sb, cluster);
fs/ocfs2/file.c
1578
p_block = ocfs2_clusters_to_blocks(sb, p_cluster) + offset;
fs/ocfs2/file.c
1579
return sb_issue_zeroout(sb, p_block, nr_blocks, GFP_NOFS);
fs/ocfs2/file.c
1842
trunc_start = ocfs2_clusters_for_bytes(osb->sb, byte_start);
fs/ocfs2/file.c
2155
struct super_block *sb = inode->i_sb;
fs/ocfs2/file.c
2162
cpos = pos >> OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/file.c
2163
clusters = ocfs2_clusters_for_bytes(sb, pos + count) - cpos;
fs/ocfs2/file.c
597
credits = ocfs2_calc_extend_credits(osb->sb, &fe->id2.i_list);
fs/ocfs2/file.c
614
ocfs2_clusters_to_bytes(osb->sb, clusters_to_add));
fs/ocfs2/file.c
654
ocfs2_clusters_to_bytes(osb->sb, clusters_to_add));
fs/ocfs2/file.c
685
ocfs2_clusters_to_bytes(osb->sb, clusters_to_add));
fs/ocfs2/file.c
972
struct super_block *sb = inode->i_sb;
fs/ocfs2/file.c
974
zero_start = ocfs2_align_bytes_to_blocks(sb, i_size_read(inode));
fs/ocfs2/inode.c
108
struct inode *ocfs2_ilookup(struct super_block *sb, u64 blkno)
fs/ocfs2/inode.c
114
args.fi_ino = ino_from_blkno(sb, blkno);
fs/ocfs2/inode.c
117
return ilookup5(sb, blkno, ocfs2_find_actor, &args);
fs/ocfs2/inode.c
124
struct super_block *sb = osb->sb;
fs/ocfs2/inode.c
142
args.fi_ino = ino_from_blkno(sb, blkno);
fs/ocfs2/inode.c
1423
int ocfs2_validate_inode_block(struct super_block *sb,
fs/ocfs2/inode.c
1438
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check);
fs/ocfs2/inode.c
145
inode = iget5_locked(sb, args.fi_ino, ocfs2_find_actor,
fs/ocfs2/inode.c
1460
rc = ocfs2_error(sb, "Invalid dinode #%llu: signature = %.*s\n",
fs/ocfs2/inode.c
1467
rc = ocfs2_error(sb, "Invalid dinode #%llu: i_blkno is %llu\n",
fs/ocfs2/inode.c
1474
rc = ocfs2_error(sb,
fs/ocfs2/inode.c
1481
OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/inode.c
1482
rc = ocfs2_error(sb,
fs/ocfs2/inode.c
1490
(u32)le16_to_cpu(di->i_suballoc_slot) > OCFS2_SB(sb)->max_slots - 1) {
fs/ocfs2/inode.c
1491
rc = ocfs2_error(sb, "Invalid dinode %llu: suballoc slot %u\n",
fs/ocfs2/inode.c
1501
rc = ocfs2_error(sb,
fs/ocfs2/inode.c
1509
ocfs2_max_inline_data_with_xattr(sb, di)) {
fs/ocfs2/inode.c
1510
rc = ocfs2_error(sb,
fs/ocfs2/inode.c
1514
ocfs2_max_inline_data_with_xattr(sb, di));
fs/ocfs2/inode.c
1519
rc = ocfs2_error(sb,
fs/ocfs2/inode.c
1530
u16 bpc = 1 << (OCFS2_SB(sb)->s_clustersize_bits -
fs/ocfs2/inode.c
1531
sb->s_blocksize_bits);
fs/ocfs2/inode.c
1533
if (le16_to_cpu(cl->cl_count) != ocfs2_chain_recs_per_inode(sb)) {
fs/ocfs2/inode.c
1534
rc = ocfs2_error(sb, "Invalid dinode %llu: chain list count %u\n",
fs/ocfs2/inode.c
1540
rc = ocfs2_error(sb, "Invalid dinode %llu: chain list index %u\n",
fs/ocfs2/inode.c
1545
if (OCFS2_SB(sb)->bitmap_blkno &&
fs/ocfs2/inode.c
1546
OCFS2_SB(sb)->bitmap_blkno != le64_to_cpu(di->i_blkno) &&
fs/ocfs2/inode.c
1548
rc = ocfs2_error(sb, "Invalid dinode %llu: bits per cluster %u\n",
fs/ocfs2/inode.c
1557
rc = ocfs2_error(sb, "Inode #%llu has refcount flag but no i_refcount_loc\n",
fs/ocfs2/inode.c
1568
static int ocfs2_filecheck_validate_inode_block(struct super_block *sb,
fs/ocfs2/inode.c
1585
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check);
fs/ocfs2/inode.c
1621
OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/inode.c
1633
static int ocfs2_filecheck_repair_inode_block(struct super_block *sb,
fs/ocfs2/inode.c
1639
if (!ocfs2_filecheck_validate_inode_block(sb, bh))
fs/ocfs2/inode.c
1645
if (unlikely(ocfs2_emergency_state(OCFS2_SB(sb)))) {
fs/ocfs2/inode.c
1683
OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/inode.c
1684
di->i_fs_generation = cpu_to_le32(OCFS2_SB(sb)->fs_generation);
fs/ocfs2/inode.c
1702
if (changed || ocfs2_validate_meta_ecc(sb, bh->b_data, &di->i_check)) {
fs/ocfs2/inode.c
1703
ocfs2_compute_meta_ecc(sb, bh->b_data, &di->i_check);
fs/ocfs2/inode.c
341
struct super_block *sb;
fs/ocfs2/inode.c
345
sb = inode->i_sb;
fs/ocfs2/inode.c
346
osb = OCFS2_SB(sb);
fs/ocfs2/inode.c
482
struct super_block *sb;
fs/ocfs2/inode.c
490
sb = inode->i_sb;
fs/ocfs2/inode.c
491
osb = OCFS2_SB(sb);
fs/ocfs2/inode.c
589
osb->sb, bh);
fs/ocfs2/inode.c
592
osb->sb, bh);
fs/ocfs2/inode.c
595
osb->sb, bh);
fs/ocfs2/inode.c
64
static int ocfs2_filecheck_validate_inode_block(struct super_block *sb,
fs/ocfs2/inode.c
651
void ocfs2_sync_blockdev(struct super_block *sb)
fs/ocfs2/inode.c
653
sync_blockdev(sb->s_bdev);
fs/ocfs2/inode.c
66
static int ocfs2_filecheck_repair_inode_block(struct super_block *sb,
fs/ocfs2/inode.h
126
struct inode *ocfs2_ilookup(struct super_block *sb, u64 feoff);
fs/ocfs2/inode.h
132
void ocfs2_sync_blockdev(struct super_block *sb);
fs/ocfs2/inode.h
150
int ocfs2_validate_inode_block(struct super_block *sb,
fs/ocfs2/ioctl.c
917
struct super_block *sb = inode->i_sb;
fs/ocfs2/ioctl.c
924
if (!bdev_max_discard_sectors(sb->s_bdev))
fs/ocfs2/ioctl.c
930
range.minlen = max_t(u64, bdev_discard_granularity(sb->s_bdev),
fs/ocfs2/ioctl.c
932
ret = ocfs2_trim_fs(sb, &range);
fs/ocfs2/journal.c
1036
ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &fe->i_check);
fs/ocfs2/journal.c
1128
static void ocfs2_clear_journal_error(struct super_block *sb,
fs/ocfs2/journal.c
1139
sb->s_id);
fs/ocfs2/journal.c
1161
ocfs2_clear_journal_error(osb->sb, journal->j_journal, osb->slot_num);
fs/ocfs2/journal.c
1264
bh = __find_get_block_nonatomic(osb->sb->s_bdev, p_blkno,
fs/ocfs2/journal.c
1265
osb->sb->s_blocksize);
fs/ocfs2/journal.c
1469
int quota_enabled = OCFS2_HAS_RO_COMPAT_FEATURE(osb->sb,
fs/ocfs2/journal.c
1471
|| OCFS2_HAS_RO_COMPAT_FEATURE(osb->sb,
fs/ocfs2/journal.c
1549
MAJOR(osb->sb->s_dev), MINOR(osb->sb->s_dev));
fs/ocfs2/journal.c
1741
"device (%u,%u)\n", node_num, slot_num, MAJOR(osb->sb->s_dev),
fs/ocfs2/journal.c
1742
MINOR(osb->sb->s_dev));
fs/ocfs2/journal.c
1767
ocfs2_clear_journal_error(osb->sb, journal, slot_num);
fs/ocfs2/journal.c
1786
ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &fe->i_check);
fs/ocfs2/journal.c
1796
"device (%u,%u)\n", node_num, slot_num, MAJOR(osb->sb->s_dev),
fs/ocfs2/journal.c
1797
MINOR(osb->sb->s_dev));
fs/ocfs2/journal.c
367
sb_start_intwrite(osb->sb);
fs/ocfs2/journal.c
374
sb_end_intwrite(osb->sb);
fs/ocfs2/journal.c
379
ocfs2_abort(osb->sb, "Detected aborted journal\n");
fs/ocfs2/journal.c
405
sb_end_intwrite(osb->sb);
fs/ocfs2/journal.c
589
ocfs2_error(ot->sb,
fs/ocfs2/journal.c
593
static void ocfs2_setup_csum_triggers(struct super_block *sb,
fs/ocfs2/journal.c
640
ot->sb = sb;
fs/ocfs2/journal.c
643
void ocfs2_initialize_journal_triggers(struct super_block *sb,
fs/ocfs2/journal.c
649
ocfs2_setup_csum_triggers(sb, type, &triggers[type]);
fs/ocfs2/journal.c
689
return ocfs2_error(osb->sb, "A previous attempt to "
fs/ocfs2/journal.h
366
static inline int ocfs2_quota_trans_credits(struct super_block *sb)
fs/ocfs2/journal.h
370
if (OCFS2_HAS_RO_COMPAT_FEATURE(sb, OCFS2_FEATURE_RO_COMPAT_USRQUOTA))
fs/ocfs2/journal.h
372
if (OCFS2_HAS_RO_COMPAT_FEATURE(sb, OCFS2_FEATURE_RO_COMPAT_GRPQUOTA))
fs/ocfs2/journal.h
387
static inline int ocfs2_inline_to_extents_credits(struct super_block *sb)
fs/ocfs2/journal.h
390
ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
400
static inline int ocfs2_remove_extent_credits(struct super_block *sb)
fs/ocfs2/journal.h
403
ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
410
static inline int ocfs2_add_dir_index_credits(struct super_block *sb)
fs/ocfs2/journal.h
415
ocfs2_clusters_to_blocks(sb, 1);
fs/ocfs2/journal.h
421
static inline int ocfs2_mknod_credits(struct super_block *sb, int is_dir,
fs/ocfs2/journal.h
427
dir_credits += ocfs2_add_dir_index_credits(sb);
fs/ocfs2/journal.h
430
ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
444
static inline int ocfs2_link_credits(struct super_block *sb)
fs/ocfs2/journal.h
447
ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
452
static inline int ocfs2_unlink_credits(struct super_block *sb)
fs/ocfs2/journal.h
455
return 2 * OCFS2_INODE_UPDATE_CREDITS + 3 + ocfs2_link_credits(sb);
fs/ocfs2/journal.h
471
static inline int ocfs2_rename_credits(struct super_block *sb)
fs/ocfs2/journal.h
473
return 3 * OCFS2_INODE_UPDATE_CREDITS + 6 + ocfs2_unlink_credits(sb);
fs/ocfs2/journal.h
487
static inline int ocfs2_calc_dxi_expand_credits(struct super_block *sb)
fs/ocfs2/journal.h
491
credits += ocfs2_clusters_to_blocks(sb, 1);
fs/ocfs2/journal.h
492
credits += ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
519
static inline int ocfs2_calc_extend_credits(struct super_block *sb,
fs/ocfs2/journal.h
543
ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
546
static inline int ocfs2_calc_symlink_credits(struct super_block *sb)
fs/ocfs2/journal.h
548
int blocks = ocfs2_mknod_credits(sb, 0, 0);
fs/ocfs2/journal.h
552
blocks += ocfs2_clusters_to_blocks(sb, 1);
fs/ocfs2/journal.h
554
return blocks + ocfs2_quota_trans_credits(sb);
fs/ocfs2/journal.h
557
static inline int ocfs2_calc_group_alloc_credits(struct super_block *sb,
fs/ocfs2/journal.h
575
static inline int ocfs2_calc_bg_discontig_credits(struct super_block *sb)
fs/ocfs2/journal.h
577
return ocfs2_extent_recs_per_gd(sb);
fs/ocfs2/localalloc.c
1056
if (bits > ocfs2_megabytes_to_clusters(osb->sb, 1)) {
fs/ocfs2/localalloc.c
107
struct super_block *sb = osb->sb;
fs/ocfs2/localalloc.c
109
gd_mb = ocfs2_clusters_to_megabytes(osb->sb,
fs/ocfs2/localalloc.c
110
8 * ocfs2_group_bitmap_size(sb, 0, osb->s_feature_incompat));
fs/ocfs2/localalloc.c
117
if ((sb->s_blocksize == 512 && osb->s_clustersize <= 8192)
fs/ocfs2/localalloc.c
118
|| (sb->s_blocksize == 1024 && osb->s_clustersize == 4096))
fs/ocfs2/localalloc.c
165
megs_per_slot = ocfs2_clusters_to_megabytes(osb->sb, megs_per_slot);
fs/ocfs2/localalloc.c
171
la_max_mb = ocfs2_clusters_to_megabytes(osb->sb,
fs/ocfs2/localalloc.c
172
ocfs2_local_alloc_size(sb) * 8);
fs/ocfs2/localalloc.c
181
struct super_block *sb = osb->sb;
fs/ocfs2/localalloc.c
185
la_max_mb = ocfs2_clusters_to_megabytes(sb,
fs/ocfs2/localalloc.c
186
ocfs2_local_alloc_size(sb) * 8);
fs/ocfs2/localalloc.c
193
ocfs2_megabytes_to_clusters(sb, la_default_mb);
fs/ocfs2/localalloc.c
197
ocfs2_megabytes_to_clusters(sb, la_max_mb);
fs/ocfs2/localalloc.c
200
ocfs2_megabytes_to_clusters(sb, requested_mb);
fs/ocfs2/localalloc.c
286
ocfs2_megabytes_to_clusters(osb->sb,
fs/ocfs2/localalloc.c
518
ocfs2_compute_meta_ecc(osb->sb, alloc_bh->b_data, &alloc->i_check);
fs/ocfs2/localalloc.c
656
status = ocfs2_error(osb->sb, "local alloc inode %llu says it has %u used bits, but a count shows %u\n",
fs/ocfs2/localalloc.c
966
la_start_blk = ocfs2_clusters_to_blocks(osb->sb,
fs/ocfs2/localalloc.c
981
ocfs2_clusters_to_blocks(osb->sb,
fs/ocfs2/move_extents.c
130
ocfs2_blocks_to_clusters(osb->sb,
fs/ocfs2/move_extents.c
179
*credits += ocfs2_calc_extend_credits(osb->sb, et->et_root_el);
fs/ocfs2/move_extents.c
338
ocfs2_clusters_to_blocks(osb->sb, new_phys_cpos),
fs/ocfs2/move_extents.c
804
move_start = ocfs2_clusters_for_bytes(osb->sb, range->me_start);
fs/ocfs2/move_extents.c
897
range->me_moved_len = ocfs2_clusters_to_bytes(osb->sb,
fs/ocfs2/move_extents.c
899
range->me_new_offset = ocfs2_clusters_to_bytes(osb->sb,
fs/ocfs2/namei.c
1492
handle = ocfs2_start_trans(osb, ocfs2_rename_credits(osb->sb));
fs/ocfs2/namei.c
1577
ocfs2_error(osb->sb, "new entry %.*s is added, but old entry %.*s "
fs/ocfs2/namei.c
1589
ocfs2_error(osb->sb, "new entry %.*s is added, but old entry %.*s "
fs/ocfs2/namei.c
1719
struct super_block *sb = osb->sb;
fs/ocfs2/namei.c
1726
blocks = (bytes_left + sb->s_blocksize - 1) >> sb->s_blocksize_bits;
fs/ocfs2/namei.c
1733
ocfs2_clusters_to_bytes(sb, OCFS2_I(inode)->ip_clusters)) {
fs/ocfs2/namei.c
1756
if ((p_blocks << sb->s_blocksize_bits) < bytes_left) {
fs/ocfs2/namei.c
1764
c = &symname[virtual * sb->s_blocksize];
fs/ocfs2/namei.c
1766
bhs[virtual] = sb_getblk(sb, p_blkno);
fs/ocfs2/namei.c
1783
memset(bhs[virtual]->b_data, 0, sb->s_blocksize);
fs/ocfs2/namei.c
1786
(bytes_left > sb->s_blocksize) ? sb->s_blocksize :
fs/ocfs2/namei.c
1793
bytes_left -= sb->s_blocksize;
fs/ocfs2/namei.c
1819
struct super_block *sb;
fs/ocfs2/namei.c
1849
sb = dir->i_sb;
fs/ocfs2/namei.c
1850
osb = OCFS2_SB(sb);
fs/ocfs2/namei.c
1854
credits = ocfs2_calc_symlink_credits(sb);
fs/ocfs2/namei.c
1921
if (l > ocfs2_fast_symlink_chars(sb))
fs/ocfs2/namei.c
1966
if (l > ocfs2_fast_symlink_chars(sb)) {
fs/ocfs2/namei.c
1970
ocfs2_clusters_to_bytes(osb->sb, 1));
fs/ocfs2/namei.c
2048
ocfs2_clusters_to_bytes(osb->sb, 1));
fs/ocfs2/namei.c
2558
handle = ocfs2_start_trans(osb, ocfs2_mknod_credits(osb->sb, 0, 0));
fs/ocfs2/namei.c
2860
handle = ocfs2_start_trans(osb, ocfs2_rename_credits(osb->sb));
fs/ocfs2/namei.c
365
handle = ocfs2_start_trans(osb, ocfs2_mknod_credits(osb->sb,
fs/ocfs2/namei.c
532
inode->i_ino = ino_from_blkno(osb->sb, fe_blkno);
fs/ocfs2/namei.c
538
*new_fe_bh = sb_getblk(osb->sb, fe_blkno);
fs/ocfs2/namei.c
555
memset(fe, 0, osb->sb->s_blocksize);
fs/ocfs2/namei.c
590
ocfs2_max_inline_data_with_xattr(osb->sb, fe));
fs/ocfs2/namei.c
595
fel->l_count = cpu_to_le16(ocfs2_extent_recs_per_inode(osb->sb));
fs/ocfs2/namei.c
782
handle = ocfs2_start_trans(osb, ocfs2_link_credits(osb->sb));
fs/ocfs2/namei.c
974
handle = ocfs2_start_trans(osb, ocfs2_unlink_credits(osb->sb));
fs/ocfs2/ocfs2.h
290
struct super_block *sb;
fs/ocfs2/ocfs2.h
308
void ocfs2_initialize_journal_triggers(struct super_block *sb,
fs/ocfs2/ocfs2.h
334
struct super_block *sb;
fs/ocfs2/ocfs2.h
514
#define OCFS2_SB(sb) ((struct ocfs2_super *)(sb)->s_fs_info)
fs/ocfs2/ocfs2.h
769
static inline unsigned long ino_from_blkno(struct super_block *sb,
fs/ocfs2/ocfs2.h
775
static inline u64 ocfs2_clusters_to_blocks(struct super_block *sb,
fs/ocfs2/ocfs2.h
778
int c_to_b_bits = OCFS2_SB(sb)->s_clustersize_bits -
fs/ocfs2/ocfs2.h
779
sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
784
static inline u32 ocfs2_clusters_for_blocks(struct super_block *sb,
fs/ocfs2/ocfs2.h
787
int b_to_c_bits = OCFS2_SB(sb)->s_clustersize_bits -
fs/ocfs2/ocfs2.h
788
sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
794
static inline u32 ocfs2_blocks_to_clusters(struct super_block *sb,
fs/ocfs2/ocfs2.h
797
int b_to_c_bits = OCFS2_SB(sb)->s_clustersize_bits -
fs/ocfs2/ocfs2.h
798
sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
803
static inline unsigned int ocfs2_clusters_for_bytes(struct super_block *sb,
fs/ocfs2/ocfs2.h
806
int cl_bits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
809
bytes += OCFS2_SB(sb)->s_clustersize - 1;
fs/ocfs2/ocfs2.h
816
static inline unsigned int ocfs2_bytes_to_clusters(struct super_block *sb,
fs/ocfs2/ocfs2.h
819
int cl_bits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
826
static inline u64 ocfs2_blocks_for_bytes(struct super_block *sb,
fs/ocfs2/ocfs2.h
829
bytes += sb->s_blocksize - 1;
fs/ocfs2/ocfs2.h
830
return bytes >> sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
833
static inline u64 ocfs2_clusters_to_bytes(struct super_block *sb,
fs/ocfs2/ocfs2.h
836
return (u64)clusters << OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
839
static inline u64 ocfs2_block_to_cluster_start(struct super_block *sb,
fs/ocfs2/ocfs2.h
842
int bits = OCFS2_SB(sb)->s_clustersize_bits - sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
845
clusters = ocfs2_blocks_to_clusters(sb, blocks);
fs/ocfs2/ocfs2.h
849
static inline u64 ocfs2_align_bytes_to_clusters(struct super_block *sb,
fs/ocfs2/ocfs2.h
852
int cl_bits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
855
clusters = ocfs2_clusters_for_bytes(sb, bytes);
fs/ocfs2/ocfs2.h
859
static inline u64 ocfs2_align_bytes_to_blocks(struct super_block *sb,
fs/ocfs2/ocfs2.h
864
blocks = ocfs2_blocks_for_bytes(sb, bytes);
fs/ocfs2/ocfs2.h
865
return blocks << sb->s_blocksize_bits;
fs/ocfs2/ocfs2.h
873
static inline unsigned int ocfs2_page_index_to_clusters(struct super_block *sb,
fs/ocfs2/ocfs2.h
877
unsigned int cbits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
890
static inline pgoff_t ocfs2_align_clusters_to_page_index(struct super_block *sb,
fs/ocfs2/ocfs2.h
893
unsigned int cbits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
905
static inline unsigned int ocfs2_pages_per_cluster(struct super_block *sb)
fs/ocfs2/ocfs2.h
907
unsigned int cbits = OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/ocfs2.h
916
static inline unsigned int ocfs2_megabytes_to_clusters(struct super_block *sb,
fs/ocfs2/ocfs2.h
921
return megs << (20 - OCFS2_SB(sb)->s_clustersize_bits);
fs/ocfs2/ocfs2.h
924
static inline unsigned int ocfs2_clusters_to_megabytes(struct super_block *sb,
fs/ocfs2/ocfs2.h
927
return clusters >> (20 - OCFS2_SB(sb)->s_clustersize_bits);
fs/ocfs2/ocfs2_fs.h
1246
static inline int ocfs2_fast_symlink_chars(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1248
return sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1252
static inline int ocfs2_max_inline_data_with_xattr(struct super_block *sb,
fs/ocfs2/ocfs2_fs.h
1258
return sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1262
return sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1266
static inline int ocfs2_extent_recs_per_inode(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1270
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1277
struct super_block *sb,
fs/ocfs2/ocfs2_fs.h
1284
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1288
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1294
static inline int ocfs2_extent_recs_per_dx_root(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1298
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1304
static inline int ocfs2_chain_recs_per_inode(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1308
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1314
static inline u16 ocfs2_extent_recs_per_eb(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1318
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1324
static inline u16 ocfs2_extent_recs_per_gd(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1328
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1334
static inline int ocfs2_dx_entries_per_leaf(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1338
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1344
static inline int ocfs2_dx_entries_per_root(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1348
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1354
static inline u16 ocfs2_local_alloc_size(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1358
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1364
static inline int ocfs2_group_bitmap_size(struct super_block *sb,
fs/ocfs2/ocfs2_fs.h
1368
int size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1384
static inline int ocfs2_truncate_recs_per_inode(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1388
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1394
static inline u64 ocfs2_backup_super_blkno(struct super_block *sb, int index)
fs/ocfs2/ocfs2_fs.h
1400
offset >>= sb->s_blocksize_bits;
fs/ocfs2/ocfs2_fs.h
1408
static inline u16 ocfs2_xattr_recs_per_xb(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1412
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1419
static inline u16 ocfs2_extent_recs_per_rb(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1423
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1429
static inline u16 ocfs2_refcount_recs_per_rb(struct super_block *sb)
fs/ocfs2/ocfs2_fs.h
1433
size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
1516
int size = sb->s_blocksize -
fs/ocfs2/ocfs2_fs.h
58
#define OCFS2_HAS_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
59
( OCFS2_SB(sb)->s_feature_compat & (mask) )
fs/ocfs2/ocfs2_fs.h
60
#define OCFS2_HAS_RO_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
61
( OCFS2_SB(sb)->s_feature_ro_compat & (mask) )
fs/ocfs2/ocfs2_fs.h
62
#define OCFS2_HAS_INCOMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
63
( OCFS2_SB(sb)->s_feature_incompat & (mask) )
fs/ocfs2/ocfs2_fs.h
64
#define OCFS2_SET_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
65
OCFS2_SB(sb)->s_feature_compat |= (mask)
fs/ocfs2/ocfs2_fs.h
66
#define OCFS2_SET_RO_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
67
OCFS2_SB(sb)->s_feature_ro_compat |= (mask)
fs/ocfs2/ocfs2_fs.h
68
#define OCFS2_SET_INCOMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
69
OCFS2_SB(sb)->s_feature_incompat |= (mask)
fs/ocfs2/ocfs2_fs.h
70
#define OCFS2_CLEAR_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
71
OCFS2_SB(sb)->s_feature_compat &= ~(mask)
fs/ocfs2/ocfs2_fs.h
72
#define OCFS2_CLEAR_RO_COMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
73
OCFS2_SB(sb)->s_feature_ro_compat &= ~(mask)
fs/ocfs2/ocfs2_fs.h
74
#define OCFS2_CLEAR_INCOMPAT_FEATURE(sb,mask) \
fs/ocfs2/ocfs2_fs.h
75
OCFS2_SB(sb)->s_feature_incompat &= ~(mask)
fs/ocfs2/ocfs2_trace.h
1659
TP_PROTO(void *sb, void *fc, int silent),
fs/ocfs2/ocfs2_trace.h
1660
TP_ARGS(sb, fc, silent),
fs/ocfs2/ocfs2_trace.h
1662
__field(void *, sb)
fs/ocfs2/ocfs2_trace.h
1667
__entry->sb = sb;
fs/ocfs2/ocfs2_trace.h
1671
TP_printk("%p %p %d", __entry->sb,
fs/ocfs2/ocfs2_trace.h
1692
TP_PROTO(void *sb, void *buf),
fs/ocfs2/ocfs2_trace.h
1693
TP_ARGS(sb, buf),
fs/ocfs2/ocfs2_trace.h
1695
__field(void *, sb)
fs/ocfs2/ocfs2_trace.h
1699
__entry->sb = sb;
fs/ocfs2/ocfs2_trace.h
1702
TP_printk("%p %p", __entry->sb, __entry->buf)
fs/ocfs2/ocfs2_trace.h
2492
TP_PROTO(void *sb, void *handle, unsigned long long blkno),
fs/ocfs2/ocfs2_trace.h
2493
TP_ARGS(sb, handle, blkno),
fs/ocfs2/ocfs2_trace.h
2495
__field(void *, sb)
fs/ocfs2/ocfs2_trace.h
2500
__entry->sb = sb;
fs/ocfs2/ocfs2_trace.h
2504
TP_printk("%p %p %llu", __entry->sb, __entry->handle, __entry->blkno)
fs/ocfs2/ocfs2_trace.h
693
TP_PROTO(struct super_block *sb, unsigned long long blk,
fs/ocfs2/ocfs2_trace.h
695
TP_ARGS(sb, blk, count),
fs/ocfs2/ocfs2_trace.h
703
__entry->dev_major = MAJOR(sb->s_dev);
fs/ocfs2/ocfs2_trace.h
704
__entry->dev_minor = MINOR(sb->s_dev);
fs/ocfs2/quota.h
112
int ocfs2_validate_quota_block(struct super_block *sb, struct buffer_head *bh);
fs/ocfs2/quota.h
94
ssize_t ocfs2_quota_read(struct super_block *sb, int type, char *data,
fs/ocfs2/quota.h
96
ssize_t ocfs2_quota_write(struct super_block *sb, int type,
fs/ocfs2/quota.h
98
int ocfs2_global_read_info(struct super_block *sb, int type);
fs/ocfs2/quota.h
99
int ocfs2_global_write_info(struct super_block *sb, int type);
fs/ocfs2/quota_global.c
1003
status = dquot_commit_info(sb, type);
fs/ocfs2/quota_global.c
1004
ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_global.c
1013
static struct dquot *ocfs2_alloc_dquot(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
134
int ocfs2_validate_quota_block(struct super_block *sb, struct buffer_head *bh)
fs/ocfs2/quota_global.c
137
ocfs2_block_dqtrailer(sb->s_blocksize, bh->b_data);
fs/ocfs2/quota_global.c
148
return ocfs2_validate_meta_ecc(sb, bh->b_data, &dqt->dq_check);
fs/ocfs2/quota_global.c
167
ssize_t ocfs2_quota_read(struct super_block *sb, int type, char *data,
fs/ocfs2/quota_global.c
170
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
173
int offset = off & (sb->s_blocksize - 1);
fs/ocfs2/quota_global.c
174
sector_t blk = off >> sb->s_blocksize_bits;
fs/ocfs2/quota_global.c
186
tocopy = min_t(size_t, (sb->s_blocksize - offset), toread);
fs/ocfs2/quota_global.c
216
ssize_t ocfs2_quota_write(struct super_block *sb, int type,
fs/ocfs2/quota_global.c
219
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_global.c
222
int offset = off & (sb->s_blocksize - 1);
fs/ocfs2/quota_global.c
223
sector_t blk = off >> sb->s_blocksize_bits;
fs/ocfs2/quota_global.c
235
if (len > sb->s_blocksize - OCFS2_QBLK_RESERVED_SPACE - offset) {
fs/ocfs2/quota_global.c
237
len = sb->s_blocksize - OCFS2_QBLK_RESERVED_SPACE - offset;
fs/ocfs2/quota_global.c
242
ocfs2_align_bytes_to_blocks(sb, off + len);
fs/ocfs2/quota_global.c
258
if ((offset || len < sb->s_blocksize - OCFS2_QBLK_RESERVED_SPACE) &&
fs/ocfs2/quota_global.c
263
bh = sb_getblk(sb, pblock);
fs/ocfs2/quota_global.c
274
memset(bh->b_data, 0, sb->s_blocksize);
fs/ocfs2/quota_global.c
338
int ocfs2_global_read_info(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
343
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_global.c
348
oinfo->dqi_gi.dqi_sb = sb;
fs/ocfs2/quota_global.c
357
oinfo->dqi_gqinode = ocfs2_get_system_file_inode(OCFS2_SB(sb), ino[type],
fs/ocfs2/quota_global.c
384
status = sb->s_op->quota_read(sb, type, (char *)&dinfo,
fs/ocfs2/quota_global.c
403
oinfo->dqi_gi.dqi_blocksize_bits = sb->s_blocksize_bits;
fs/ocfs2/quota_global.c
404
oinfo->dqi_gi.dqi_usable_bs = sb->s_blocksize -
fs/ocfs2/quota_global.c
420
static int __ocfs2_global_write_info(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
422
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_global.c
436
size = sb->s_op->quota_write(sb, type, (char *)&dinfo,
fs/ocfs2/quota_global.c
448
int ocfs2_global_write_info(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
451
struct quota_info *dqopt = sb_dqopt(sb);
fs/ocfs2/quota_global.c
460
err = __ocfs2_global_write_info(sb, type);
fs/ocfs2/quota_global.c
468
static int ocfs2_global_qinit_alloc(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
470
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
479
static int ocfs2_calc_global_qinit_credits(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
483
return (ocfs2_global_qinit_alloc(sb, type) + 2) *
fs/ocfs2/quota_global.c
493
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_global.c
495
struct ocfs2_mem_dqinfo *info = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
500
err = sb->s_op->quota_read(sb, type, (char *)&dqblk,
fs/ocfs2/quota_global.c
586
if (info_dirty(sb_dqinfo(sb, type))) {
fs/ocfs2/quota_global.c
587
err2 = __ocfs2_global_write_info(sb, type);
fs/ocfs2/quota_global.c
606
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_global.c
607
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
608
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/quota_global.c
614
type, sb->s_id);
fs/ocfs2/quota_global.c
627
down_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_global.c
637
up_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_global.c
650
struct super_block *sb = oinfo->dqi_gqinode->i_sb;
fs/ocfs2/quota_global.c
657
if (down_read_trylock(&sb->s_umount)) {
fs/ocfs2/quota_global.c
658
dquot_scan_active(sb, ocfs2_sync_dquot_helper, oinfo->dqi_type);
fs/ocfs2/quota_global.c
659
up_read(&sb->s_umount);
fs/ocfs2/quota_global.c
695
static int ocfs2_calc_qdel_credits(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
697
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
813
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_global.c
814
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/quota_global.c
816
struct ocfs2_mem_dqinfo *info = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
818
int need_alloc = ocfs2_global_qinit_alloc(sb, type);
fs/ocfs2/quota_global.c
854
i_size_read(gqinode) + (need_alloc << sb->s_blocksize_bits),
fs/ocfs2/quota_global.c
861
ocfs2_calc_global_qinit_credits(sb, type));
fs/ocfs2/quota_global.c
870
if (ex && info_dirty(sb_dqinfo(sb, type))) {
fs/ocfs2/quota_global.c
871
err = __ocfs2_global_write_info(sb, type);
fs/ocfs2/quota_global.c
894
static int ocfs2_get_next_id(struct super_block *sb, struct kqid *qid)
fs/ocfs2/quota_global.c
897
struct ocfs2_mem_dqinfo *info = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
901
if (!sb_has_quota_active(sb, type)) {
fs/ocfs2/quota_global.c
935
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_global.c
937
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
939
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/quota_global.c
966
down_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_global.c
977
up_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_global.c
988
static int ocfs2_write_info(struct super_block *sb, int type)
fs/ocfs2/quota_global.c
992
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_global.c
997
handle = ocfs2_start_trans(OCFS2_SB(sb), OCFS2_QINFO_WRITE_CREDITS);
fs/ocfs2/quota_local.c
101
handle = ocfs2_start_trans(OCFS2_SB(sb),
fs/ocfs2/quota_local.c
1013
bh = sb_getblk(sb, p_blkno);
fs/ocfs2/quota_local.c
1028
dchunk->dqc_free = cpu_to_le32(ol_quota_entries_per_block(sb));
fs/ocfs2/quota_local.c
1030
sb->s_blocksize - sizeof(struct ocfs2_local_disk_chunk) -
fs/ocfs2/quota_local.c
1042
dbh = sb_getblk(sb, p_blkno);
fs/ocfs2/quota_local.c
1056
memset(dbh->b_data, 0, sb->s_blocksize - OCFS2_QBLK_RESERVED_SPACE);
fs/ocfs2/quota_local.c
1063
status = ocfs2_local_write_info(sb, type);
fs/ocfs2/quota_local.c
1068
status = ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
1082
ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
1092
struct super_block *sb,
fs/ocfs2/quota_local.c
1096
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
1099
struct inode *lqinode = sb_dqopt(sb)->files[type];
fs/ocfs2/quota_local.c
1101
int epb = ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
1109
return ocfs2_local_quota_add_chunk(sb, type, offset);
fs/ocfs2/quota_local.c
1114
ol_quota_chunk_block(sb, chunk->qc_num) - 1;
fs/ocfs2/quota_local.c
1115
if (ol_chunk_blocks(sb) == chunk_blocks)
fs/ocfs2/quota_local.c
1116
return ocfs2_local_quota_add_chunk(sb, type, offset);
fs/ocfs2/quota_local.c
112
ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
1120
i_size_read(lqinode) + sb->s_blocksize,
fs/ocfs2/quota_local.c
1127
i_size_read(lqinode) + sb->s_blocksize);
fs/ocfs2/quota_local.c
1140
bh = sb_getblk(sb, p_blkno);
fs/ocfs2/quota_local.c
1149
handle = ocfs2_start_trans(OCFS2_SB(sb),
fs/ocfs2/quota_local.c
1165
memset(bh->b_data, 0, sb->s_blocksize);
fs/ocfs2/quota_local.c
1180
le32_add_cpu(&dchunk->dqc_free, ol_quota_entries_per_block(sb));
fs/ocfs2/quota_local.c
1186
status = ocfs2_local_write_info(sb, type);
fs/ocfs2/quota_local.c
1192
status = ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
120
status = ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
1200
ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
1218
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_local.c
1220
struct inode *lqinode = sb_dqopt(sb)->files[type];
fs/ocfs2/quota_local.c
1228
chunk = ocfs2_find_free_entry(sb, type, &offset);
fs/ocfs2/quota_local.c
1230
chunk = ocfs2_extend_local_quota_file(sb, type, &offset);
fs/ocfs2/quota_local.c
1239
od->dq_local_off = ol_dqblk_off(sb, chunk->qc_num, offset);
fs/ocfs2/quota_local.c
1242
ol_dqblk_block(sb, chunk->qc_num, offset),
fs/ocfs2/quota_local.c
1279
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_local.c
1284
INODE_CACHE(sb_dqopt(sb)->files[type]),
fs/ocfs2/quota_local.c
1290
offset = ol_dqblk_chunk_off(sb, od->dq_chunk->qc_num,
fs/ocfs2/quota_local.c
160
static int ocfs2_local_check_quota_file(struct super_block *sb, int type)
fs/ocfs2/quota_local.c
169
struct inode *linode = sb_dqopt(sb)->files[type];
fs/ocfs2/quota_local.c
199
ginode = ocfs2_get_system_file_inode(OCFS2_SB(sb), ino[type],
fs/ocfs2/quota_local.c
29
static inline unsigned int ol_quota_entries_per_block(struct super_block *sb)
fs/ocfs2/quota_local.c
294
static int ocfs2_add_recovery_chunk(struct super_block *sb,
fs/ocfs2/quota_local.c
305
rc->rc_bitmap = kmalloc(sb->s_blocksize, GFP_NOFS);
fs/ocfs2/quota_local.c
31
return ((sb->s_blocksize - OCFS2_QBLK_RESERVED_SPACE) /
fs/ocfs2/quota_local.c
311
(ol_chunk_entries(sb) + 7) >> 3);
fs/ocfs2/quota_local.c
343
struct super_block *sb = lqinode->i_sb;
fs/ocfs2/quota_local.c
352
ol_quota_chunk_block(sb, i),
fs/ocfs2/quota_local.c
359
if (le32_to_cpu(dchunk->dqc_free) < ol_chunk_entries(sb))
fs/ocfs2/quota_local.c
36
static inline unsigned int ol_chunk_blocks(struct super_block *sb)
fs/ocfs2/quota_local.c
360
status = ocfs2_add_recovery_chunk(sb, dchunk, i, head);
fs/ocfs2/quota_local.c
38
return ((sb->s_blocksize - sizeof(struct ocfs2_local_disk_chunk) -
fs/ocfs2/quota_local.c
393
struct super_block *sb = osb->sb;
fs/ocfs2/quota_local.c
40
ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
410
if (!OCFS2_HAS_RO_COMPAT_FEATURE(sb, feature[type]))
fs/ocfs2/quota_local.c
44
static unsigned int ol_chunk_entries(struct super_block *sb)
fs/ocfs2/quota_local.c
46
return ol_chunk_blocks(sb) * ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
461
struct super_block *sb = lqinode->i_sb;
fs/ocfs2/quota_local.c
462
struct ocfs2_mem_dqinfo *oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/quota_local.c
480
ol_quota_chunk_block(sb, chunk),
fs/ocfs2/quota_local.c
487
for_each_set_bit(bit, rchunk->rc_bitmap, ol_chunk_entries(sb)) {
fs/ocfs2/quota_local.c
490
ol_dqblk_block(sb, chunk, bit),
fs/ocfs2/quota_local.c
497
ol_dqblk_block_off(sb, chunk, bit));
fs/ocfs2/quota_local.c
498
dquot = dqget(sb,
fs/ocfs2/quota_local.c
50
static unsigned int ol_quota_chunk_block(struct super_block *sb, int c)
fs/ocfs2/quota_local.c
516
handle = ocfs2_start_trans(OCFS2_SB(sb),
fs/ocfs2/quota_local.c
523
down_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_local.c
53
return 1 + (ol_chunk_blocks(sb) + 1) * c;
fs/ocfs2/quota_local.c
558
up_write(&sb_dqopt(sb)->dqio_sem);
fs/ocfs2/quota_local.c
559
ocfs2_commit_trans(OCFS2_SB(sb), handle);
fs/ocfs2/quota_local.c
56
static unsigned int ol_dqblk_block(struct super_block *sb, int c, int off)
fs/ocfs2/quota_local.c
58
int epb = ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
60
return ol_quota_chunk_block(sb, c) + 1 + off / epb;
fs/ocfs2/quota_local.c
63
static unsigned int ol_dqblk_block_off(struct super_block *sb, int c, int off)
fs/ocfs2/quota_local.c
65
int epb = ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
682
static int ocfs2_local_read_info(struct super_block *sb, int type)
fs/ocfs2/quota_local.c
685
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
687
struct inode *lqinode = sb_dqopt(sb)->files[type];
fs/ocfs2/quota_local.c
709
status = ocfs2_global_read_info(sb, type);
fs/ocfs2/quota_local.c
71
static loff_t ol_dqblk_off(struct super_block *sb, int c, int off)
fs/ocfs2/quota_local.c
73
return (ol_dqblk_block(sb, c, off) << sb->s_blocksize_bits) +
fs/ocfs2/quota_local.c
738
rec = OCFS2_SB(sb)->quota_rec;
fs/ocfs2/quota_local.c
74
ol_dqblk_block_off(sb, c, off);
fs/ocfs2/quota_local.c
746
OCFS2_SB(sb)->quota_rec = rec;
fs/ocfs2/quota_local.c
77
static inline unsigned int ol_dqblk_block_offset(struct super_block *sb, loff_t off)
fs/ocfs2/quota_local.c
777
ocfs2_simple_drop_lockres(OCFS2_SB(sb), &oinfo->dqi_gqlock);
fs/ocfs2/quota_local.c
79
return off & ((1 << sb->s_blocksize_bits) - 1);
fs/ocfs2/quota_local.c
792
static int ocfs2_local_write_info(struct super_block *sb, int type)
fs/ocfs2/quota_local.c
794
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
799
status = ocfs2_modify_bh(sb_dqopt(sb)->files[type], bh, olq_update_info,
fs/ocfs2/quota_local.c
810
static int ocfs2_local_free_info(struct super_block *sb, int type)
fs/ocfs2/quota_local.c
812
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
820
ocfs2_simple_drop_lockres(OCFS2_SB(sb), &oinfo->dqi_gqlock);
fs/ocfs2/quota_local.c
826
len = ol_chunk_entries(sb);
fs/ocfs2/quota_local.c
829
ol_quota_chunk_block(sb, chunk->qc_num) - 1)
fs/ocfs2/quota_local.c
83
static int ol_dqblk_chunk_off(struct super_block *sb, int c, loff_t off)
fs/ocfs2/quota_local.c
830
* ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
85
int epb = ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
854
status = ocfs2_modify_bh(sb_dqopt(sb)->files[type],
fs/ocfs2/quota_local.c
861
ocfs2_inode_unlock(sb_dqopt(sb)->files[type], 1);
fs/ocfs2/quota_local.c
87
return ((off >> sb->s_blocksize_bits) -
fs/ocfs2/quota_local.c
873
struct super_block *sb = od->dq_dquot.dq_sb;
fs/ocfs2/quota_local.c
876
+ ol_dqblk_block_offset(sb, od->dq_local_off));
fs/ocfs2/quota_local.c
88
ol_quota_chunk_block(sb, c) - 1) * epb
fs/ocfs2/quota_local.c
89
+ ((unsigned int)(off & ((1 << sb->s_blocksize_bits) - 1))) /
fs/ocfs2/quota_local.c
895
struct super_block *sb = dquot->dq_sb;
fs/ocfs2/quota_local.c
898
struct inode *lqinode = sb_dqopt(sb)->files[dquot->dq_id.type];
fs/ocfs2/quota_local.c
918
static struct ocfs2_quota_chunk *ocfs2_find_free_entry(struct super_block *sb,
fs/ocfs2/quota_local.c
922
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
940
len = ol_chunk_entries(sb);
fs/ocfs2/quota_local.c
943
ol_quota_chunk_block(sb, chunk->qc_num) - 1)
fs/ocfs2/quota_local.c
944
* ol_quota_entries_per_block(sb);
fs/ocfs2/quota_local.c
961
struct super_block *sb,
fs/ocfs2/quota_local.c
965
struct mem_dqinfo *info = sb_dqinfo(sb, type);
fs/ocfs2/quota_local.c
967
struct inode *lqinode = sb_dqopt(sb)->files[type];
fs/ocfs2/quota_local.c
97
struct super_block *sb = inode->i_sb;
fs/ocfs2/quota_local.c
977
i_size_read(lqinode) + 2 * sb->s_blocksize,
fs/ocfs2/quota_local.c
984
i_size_read(lqinode) + 2 * sb->s_blocksize);
fs/ocfs2/quota_local.c
997
handle = ocfs2_start_trans(OCFS2_SB(sb),
fs/ocfs2/refcounttree.c
1019
ret = ocfs2_find_cpos_for_right_leaf(sb, left_path, &cpos);
fs/ocfs2/refcounttree.c
106
rc = ocfs2_error(sb,
fs/ocfs2/refcounttree.c
1071
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
1097
ret = ocfs2_error(sb,
fs/ocfs2/refcounttree.c
113
if (le32_to_cpu(rb->rf_fs_generation) != OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/refcounttree.c
114
rc = ocfs2_error(sb,
fs/ocfs2/refcounttree.c
1286
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
1307
new_bh = sb_getblk(sb, blkno);
fs/ocfs2/refcounttree.c
1327
memcpy(new_bh->b_data, ref_root_bh->b_data, sb->s_blocksize);
fs/ocfs2/refcounttree.c
1340
memset(&root_rb->rf_list, 0, sb->s_blocksize -
fs/ocfs2/refcounttree.c
1342
root_rb->rf_list.l_count = cpu_to_le16(ocfs2_extent_recs_per_rb(sb));
fs/ocfs2/refcounttree.c
1517
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
1548
new_bh = sb_getblk(sb, blkno);
fs/ocfs2/refcounttree.c
1565
memset(new_rb, 0, sb->s_blocksize);
fs/ocfs2/refcounttree.c
1570
new_rb->rf_fs_generation = cpu_to_le32(OCFS2_SB(sb)->fs_generation);
fs/ocfs2/refcounttree.c
1575
cpu_to_le16(ocfs2_refcount_recs_per_rb(sb));
fs/ocfs2/refcounttree.c
2075
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
2135
memset(&rb->rf_records, 0, sb->s_blocksize -
fs/ocfs2/refcounttree.c
2138
cpu_to_le16(ocfs2_refcount_recs_per_rb(sb));
fs/ocfs2/refcounttree.c
2226
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
2261
ocfs2_clusters_to_blocks(sb, cpos),
fs/ocfs2/refcounttree.c
2363
static int ocfs2_calc_refcount_meta_credits(struct super_block *sb,
fs/ocfs2/refcounttree.c
2485
*credits += ocfs2_calc_extend_credits(sb,
fs/ocfs2/refcounttree.c
2567
static inline unsigned int ocfs2_cow_contig_clusters(struct super_block *sb)
fs/ocfs2/refcounttree.c
2569
return ocfs2_clusters_for_bytes(sb, MAX_CONTIG_BYTES);
fs/ocfs2/refcounttree.c
2572
static inline unsigned int ocfs2_cow_contig_mask(struct super_block *sb)
fs/ocfs2/refcounttree.c
2574
return ~(ocfs2_cow_contig_clusters(sb) - 1);
fs/ocfs2/refcounttree.c
2584
static inline unsigned int ocfs2_cow_align_start(struct super_block *sb,
fs/ocfs2/refcounttree.c
2590
return start + ((cpos - start) & ocfs2_cow_contig_mask(sb));
fs/ocfs2/refcounttree.c
2597
static inline unsigned int ocfs2_cow_align_length(struct super_block *sb,
fs/ocfs2/refcounttree.c
2601
(len + (ocfs2_cow_contig_clusters(sb) - 1)) &
fs/ocfs2/refcounttree.c
2602
ocfs2_cow_contig_mask(sb);
fs/ocfs2/refcounttree.c
2831
static int ocfs2_lock_refcount_allocators(struct super_block *sb,
fs/ocfs2/refcounttree.c
2853
*credits += ocfs2_calc_extend_credits(sb, et->et_root_el);
fs/ocfs2/refcounttree.c
2855
ret = ocfs2_calc_refcount_meta_credits(sb, ref_ci, ref_root_bh,
fs/ocfs2/refcounttree.c
2864
ret = ocfs2_reserve_new_metadata_blocks(OCFS2_SB(sb), meta_add,
fs/ocfs2/refcounttree.c
2872
ret = ocfs2_reserve_clusters(OCFS2_SB(sb), num_clusters,
fs/ocfs2/refcounttree.c
2904
struct super_block *sb = inode->i_sb;
fs/ocfs2/refcounttree.c
2905
u64 new_block = ocfs2_clusters_to_blocks(sb, new_cluster);
fs/ocfs2/refcounttree.c
2914
offset = ((loff_t)cpos) << OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/refcounttree.c
2915
end = offset + (new_len << OCFS2_SB(sb)->s_clustersize_bits);
fs/ocfs2/refcounttree.c
292
struct super_block *sb)
fs/ocfs2/refcounttree.c
2949
if (PAGE_SIZE <= OCFS2_SB(sb)->s_clustersize) {
fs/ocfs2/refcounttree.c
296
new->rf_sb = sb;
fs/ocfs2/refcounttree.c
2999
struct super_block *sb = inode->i_sb;
fs/ocfs2/refcounttree.c
3001
int i, blocks = ocfs2_clusters_to_blocks(sb, new_len);
fs/ocfs2/refcounttree.c
3002
u64 old_block = ocfs2_clusters_to_blocks(sb, old_cluster);
fs/ocfs2/refcounttree.c
3003
u64 new_block = ocfs2_clusters_to_blocks(sb, new_cluster);
fs/ocfs2/refcounttree.c
3004
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/refcounttree.c
3012
new_bh = sb_getblk(osb->sb, new_block);
fs/ocfs2/refcounttree.c
3034
memcpy(new_bh->b_data, old_bh->b_data, sb->s_blocksize);
fs/ocfs2/refcounttree.c
3059
struct super_block *sb = ocfs2_metadata_cache_get_super(et->et_ci);
fs/ocfs2/refcounttree.c
3068
replace_rec.e_blkno = cpu_to_le64(ocfs2_clusters_to_blocks(sb,
fs/ocfs2/refcounttree.c
3090
ret = ocfs2_error(sb,
fs/ocfs2/refcounttree.c
3138
int ocfs2_cow_sync_writeback(struct super_block *sb,
fs/ocfs2/refcounttree.c
3148
start = ((loff_t)cpos) << OCFS2_SB(sb)->s_clustersize_bits;
fs/ocfs2/refcounttree.c
3149
end = start + (num_clusters << OCFS2_SB(sb)->s_clustersize_bits) - 1;
fs/ocfs2/refcounttree.c
3167
static int ocfs2_make_clusters_writable(struct super_block *sb,
fs/ocfs2/refcounttree.c
3175
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/refcounttree.c
3184
ret = ocfs2_lock_refcount_allocators(sb, p_cluster, num_clusters,
fs/ocfs2/refcounttree.c
320
ocfs2_init_refcount_tree_ci(new, osb->sb);
fs/ocfs2/refcounttree.c
3295
ret = ocfs2_cow_sync_writeback(sb, context->inode, cpos,
fs/ocfs2/refcounttree.c
3907
ocfs2_clusters_to_bytes(osb->sb, num_clusters));
fs/ocfs2/refcounttree.c
633
cpu_to_le16(ocfs2_refcount_recs_per_rb(osb->sb));
fs/ocfs2/refcounttree.c
73
static int ocfs2_validate_refcount_block(struct super_block *sb,
fs/ocfs2/refcounttree.c
89
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &rb->rf_check);
fs/ocfs2/refcounttree.c
967
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/refcounttree.c
98
rc = ocfs2_error(sb,
fs/ocfs2/refcounttree.h
85
int ocfs2_cow_sync_writeback(struct super_block *sb,
fs/ocfs2/resize.c
119
if (OCFS2_HAS_COMPAT_FEATURE(osb->sb,
fs/ocfs2/resize.c
247
if (OCFS2_HAS_COMPAT_FEATURE(osb->sb, OCFS2_FEATURE_COMPAT_BACKUP_SB))
fs/ocfs2/resize.c
311
ocfs2_group_bitmap_size(osb->sb, 0,
fs/ocfs2/resize.c
492
ocfs2_group_bitmap_size(osb->sb, 0,
fs/ocfs2/slot_map.c
339
static int ocfs2_validate_slot_map_block(struct super_block *sb,
fs/ocfs2/slot_map.c
347
rc = ocfs2_error(sb,
fs/ocfs2/slot_map.c
378
(osb->sb->s_blocksize /
fs/ocfs2/slot_map.c
381
si->si_slots_per_block = osb->sb->s_blocksize / sizeof(__le16);
fs/ocfs2/slot_map.c
47
static int ocfs2_validate_slot_map_block(struct super_block *sb,
fs/ocfs2/suballoc.c
159
return ocfs2_error(sb, fmt, ##__VA_ARGS__); \
fs/ocfs2/suballoc.c
162
static int ocfs2_validate_gd_self(struct super_block *sb,
fs/ocfs2/suballoc.c
180
if (le32_to_cpu(gd->bg_generation) != OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/suballoc.c
203
static int ocfs2_validate_gd_parent(struct super_block *sb,
fs/ocfs2/suballoc.c
244
int ocfs2_check_group_descriptor(struct super_block *sb,
fs/ocfs2/suballoc.c
2452
u32 data_cluster = ocfs2_blocks_to_clusters(osb->sb, data_blkno);
fs/ocfs2/suballoc.c
2462
*bg_bit_off = (u16) ocfs2_blocks_to_clusters(osb->sb,
fs/ocfs2/suballoc.c
258
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &gd->bg_check);
fs/ocfs2/suballoc.c
264
rc = ocfs2_validate_gd_self(sb, bh, 1);
fs/ocfs2/suballoc.c
266
rc = ocfs2_validate_gd_parent(sb, di, bh, 1);
fs/ocfs2/suballoc.c
2673
ocfs2_calc_group_alloc_credits(osb->sb,
fs/ocfs2/suballoc.c
271
static int ocfs2_validate_group_descriptor(struct super_block *sb,
fs/ocfs2/suballoc.c
287
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &gd->bg_check);
fs/ocfs2/suballoc.c
295
return ocfs2_validate_gd_self(sb, bh, 0);
fs/ocfs2/suballoc.c
401
el->l_count = cpu_to_le16(ocfs2_extent_recs_per_gd(osb->sb));
fs/ocfs2/suballoc.c
424
struct super_block * sb = alloc_inode->i_sb;
fs/ocfs2/suballoc.c
443
memset(bg, 0, sb->s_blocksize);
fs/ocfs2/suballoc.c
446
bg->bg_size = cpu_to_le16(ocfs2_group_bitmap_size(sb, 1,
fs/ocfs2/suballoc.c
511
bg_blkno = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/suballoc.c
515
bg_bh = sb_getblk(osb->sb, bg_blkno);
fs/ocfs2/suballoc.c
592
p_blkno = ocfs2_clusters_to_blocks(osb->sb, p_cpos);
fs/ocfs2/suballoc.c
666
ocfs2_calc_bg_discontig_credits(osb->sb));
fs/ocfs2/suballoc.c
691
bg_blkno = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/suballoc.c
695
bg_bh = sb_getblk(osb->sb, bg_blkno);
fs/ocfs2/suballoc.c
752
credits = ocfs2_calc_group_alloc_credits(osb->sb,
fs/ocfs2/suballoc.h
165
return ocfs2_blocks_to_clusters(osb->sb, bg_blkno);
fs/ocfs2/suballoc.h
191
int ocfs2_check_group_descriptor(struct super_block *sb,
fs/ocfs2/super.c
1001
if (!ocfs2_check_set_options(sb, parsed_options)) {
fs/ocfs2/super.c
1022
sb->s_magic = OCFS2_SUPER_MAGIC;
fs/ocfs2/super.c
1024
sb->s_flags = (sb->s_flags & ~(SB_POSIXACL | SB_NOSEC)) |
fs/ocfs2/super.c
1029
if (bdev_read_only(sb->s_bdev)) {
fs/ocfs2/super.c
1030
if (!sb_rdonly(sb)) {
fs/ocfs2/super.c
1063
if (sb_rdonly(sb))
fs/ocfs2/super.c
1078
ocfs2_initialize_journal_triggers(sb, osb->s_journal_triggers);
fs/ocfs2/super.c
1083
status = ocfs2_mount_volume(sb);
fs/ocfs2/super.c
109
static int ocfs2_initialize_super(struct super_block *sb,
fs/ocfs2/super.c
1095
osb->osb_dev_kset = kset_create_and_add(sb->s_id, NULL,
fs/ocfs2/super.c
1099
mlog(ML_ERROR, "Unable to create device kset %s.\n", sb->s_id);
fs/ocfs2/super.c
1108
"/sys/fs/ocfs2/%s/filecheck.\n", sb->s_id);
fs/ocfs2/super.c
1118
sb->s_root = root;
fs/ocfs2/super.c
113
static int ocfs2_get_sector(struct super_block *sb,
fs/ocfs2/super.c
1139
if (!sb_rdonly(sb)) {
fs/ocfs2/super.c
1166
ocfs2_dismount_volume(sb, 1);
fs/ocfs2/super.c
117
static struct inode *ocfs2_alloc_inode(struct super_block *sb);
fs/ocfs2/super.c
1233
static int ocfs2_check_set_options(struct super_block *sb,
fs/ocfs2/super.c
1249
!OCFS2_HAS_RO_COMPAT_FEATURE(sb,
fs/ocfs2/super.c
1256
!OCFS2_HAS_RO_COMPAT_FEATURE(sb,
fs/ocfs2/super.c
1263
!OCFS2_HAS_INCOMPAT_FEATURE(sb, OCFS2_FEATURE_INCOMPAT_XATTR)) {
fs/ocfs2/super.c
1271
if (OCFS2_HAS_INCOMPAT_FEATURE(sb, OCFS2_FEATURE_INCOMPAT_XATTR))
fs/ocfs2/super.c
1548
static void ocfs2_put_super(struct super_block *sb)
fs/ocfs2/super.c
1550
trace_ocfs2_put_super(sb);
fs/ocfs2/super.c
1552
ocfs2_sync_blockdev(sb);
fs/ocfs2/super.c
1553
ocfs2_dismount_volume(sb, 0);
fs/ocfs2/super.c
1593
(osb->s_clustersize >> osb->sb->s_blocksize_bits);
fs/ocfs2/super.c
1595
(osb->s_clustersize >> osb->sb->s_blocksize_bits);
fs/ocfs2/super.c
1694
static int ocfs2_get_sector(struct super_block *sb,
fs/ocfs2/super.c
1699
if (!sb_set_blocksize(sb, sect_size)) {
fs/ocfs2/super.c
1704
*bh = sb_getblk(sb, block);
fs/ocfs2/super.c
1723
static int ocfs2_mount_volume(struct super_block *sb)
fs/ocfs2/super.c
1726
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/super.c
1794
static void ocfs2_dismount_volume(struct super_block *sb, int mnt_err)
fs/ocfs2/super.c
1800
trace_ocfs2_dismount_volume(sb);
fs/ocfs2/super.c
1802
BUG_ON(!sb);
fs/ocfs2/super.c
1803
osb = OCFS2_SB(sb);
fs/ocfs2/super.c
1832
ocfs2_sync_blockdev(sb);
fs/ocfs2/super.c
1886
sb->s_dev = 0;
fs/ocfs2/super.c
1887
sb->s_fs_info = NULL;
fs/ocfs2/super.c
1921
ocfs2_clusters_to_blocks(osb->sb,
fs/ocfs2/super.c
1930
if (!(OCFS2_HAS_COMPAT_FEATURE(osb->sb,
fs/ocfs2/super.c
1944
static int ocfs2_initialize_super(struct super_block *sb,
fs/ocfs2/super.c
1963
sb->s_fs_info = osb;
fs/ocfs2/super.c
1964
sb->s_op = &ocfs2_sops;
fs/ocfs2/super.c
1965
set_default_d_op(sb, &ocfs2_dentry_ops);
fs/ocfs2/super.c
1966
sb->s_export_op = &ocfs2_export_ops;
fs/ocfs2/super.c
1967
sb->s_qcop = &dquot_quotactl_sysfile_ops;
fs/ocfs2/super.c
1968
sb->dq_op = &ocfs2_quota_operations;
fs/ocfs2/super.c
1969
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
fs/ocfs2/super.c
1970
sb->s_xattr = ocfs2_xattr_handlers;
fs/ocfs2/super.c
1971
sb->s_time_gran = 1;
fs/ocfs2/super.c
1972
sb->s_flags |= SB_NOATIME;
fs/ocfs2/super.c
1976
sb->s_maxbytes = ocfs2_max_file_offset(bbits, cbits);
fs/ocfs2/super.c
1977
super_set_uuid(sb, di->id2.i_super.s_uuid,
fs/ocfs2/super.c
1986
osb->sb = sb;
fs/ocfs2/super.c
2014
MAJOR(osb->sb->s_dev), MINOR(osb->sb->s_dev));
fs/ocfs2/super.c
2085
if ((i = OCFS2_HAS_INCOMPAT_FEATURE(osb->sb, ~OCFS2_FEATURE_INCOMPAT_SUPP))) {
fs/ocfs2/super.c
2091
if (!sb_rdonly(osb->sb) && (i = OCFS2_HAS_RO_COMPAT_FEATURE(osb->sb, ~OCFS2_FEATURE_RO_COMPAT_SUPP))) {
fs/ocfs2/super.c
2154
total_blocks = ocfs2_clusters_to_blocks(osb->sb,
fs/ocfs2/super.c
2157
status = generic_check_addressable(osb->sb->s_blocksize_bits,
fs/ocfs2/super.c
2217
osb->bitmap_cpg = ocfs2_group_bitmap_size(sb, 0,
fs/ocfs2/super.c
2255
sb->s_fs_info = NULL;
fs/ocfs2/super.c
235
osb->sb->s_blocksize, osb->s_clustersize);
fs/ocfs2/super.c
2473
static int ocfs2_handle_error(struct super_block *sb)
fs/ocfs2/super.c
2475
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/super.c
2484
sb->s_id);
fs/ocfs2/super.c
2490
if (sb_rdonly(sb) && ocfs2_emergency_state(osb))
fs/ocfs2/super.c
2494
sb->s_flags |= SB_RDONLY;
fs/ocfs2/super.c
2501
int __ocfs2_error(struct super_block *sb, const char *function,
fs/ocfs2/super.c
2514
sb->s_id, function, &vaf);
fs/ocfs2/super.c
2518
return ocfs2_handle_error(sb);
fs/ocfs2/super.c
2524
void __ocfs2_abort(struct super_block *sb, const char *function,
fs/ocfs2/super.c
2536
sb->s_id, function, &vaf);
fs/ocfs2/super.c
2551
if (!ocfs2_mount_local(OCFS2_SB(sb)))
fs/ocfs2/super.c
2552
OCFS2_SB(sb)->s_mount_opt |= OCFS2_MOUNT_ERRORS_PANIC;
fs/ocfs2/super.c
2553
ocfs2_handle_error(sb);
fs/ocfs2/super.c
396
static int ocfs2_sync_fs(struct super_block *sb, int wait)
fs/ocfs2/super.c
400
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/super.c
424
if (!OCFS2_HAS_RO_COMPAT_FEATURE(osb->sb, OCFS2_FEATURE_RO_COMPAT_USRQUOTA)
fs/ocfs2/super.c
428
if (!OCFS2_HAS_RO_COMPAT_FEATURE(osb->sb, OCFS2_FEATURE_RO_COMPAT_GRPQUOTA)
fs/ocfs2/super.c
549
static struct inode *ocfs2_alloc_inode(struct super_block *sb)
fs/ocfs2/super.c
553
oi = alloc_inode_sb(sb, ocfs2_inode_cachep, GFP_NOFS);
fs/ocfs2/super.c
612
struct super_block *sb = fc->root->d_sb;
fs/ocfs2/super.c
613
struct ocfs2_super *osb = OCFS2_SB(sb);
fs/ocfs2/super.c
616
sync_filesystem(sb);
fs/ocfs2/super.c
618
if (!ocfs2_check_set_options(sb, parsed_options)) {
fs/ocfs2/super.c
648
if ((bool)(fc->sb_flags & SB_RDONLY) != sb_rdonly(sb)) {
fs/ocfs2/super.c
665
sb->s_flags |= SB_RDONLY;
fs/ocfs2/super.c
674
incompat_features = OCFS2_HAS_RO_COMPAT_FEATURE(sb, ~OCFS2_FEATURE_RO_COMPAT_SUPP);
fs/ocfs2/super.c
682
sb->s_flags &= ~SB_RDONLY;
fs/ocfs2/super.c
685
trace_ocfs2_remount(sb->s_flags, osb->osb_flags, fc->sb_flags);
fs/ocfs2/super.c
690
if (sb_any_quota_suspended(sb))
fs/ocfs2/super.c
697
sb->s_flags |= SB_RDONLY;
fs/ocfs2/super.c
717
sb->s_flags = (sb->s_flags & ~SB_POSIXACL) |
fs/ocfs2/super.c
725
static int ocfs2_sb_probe(struct super_block *sb,
fs/ocfs2/super.c
738
*sector_size = bdev_logical_block_size(sb->s_bdev);
fs/ocfs2/super.c
751
status = ocfs2_get_sector(sb, bh, 0, *sector_size);
fs/ocfs2/super.c
785
tmpstat = ocfs2_get_sector(sb, bh,
fs/ocfs2/super.c
87
static int ocfs2_check_set_options(struct super_block *sb,
fs/ocfs2/super.c
879
struct super_block *sb = osb->sb;
fs/ocfs2/super.c
886
if (!OCFS2_HAS_RO_COMPAT_FEATURE(sb, feature[type]))
fs/ocfs2/super.c
889
status = dquot_resume(sb, type);
fs/ocfs2/super.c
894
oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/super.c
896
status = dquot_suspend(sb, type);
fs/ocfs2/super.c
90
static void ocfs2_put_super(struct super_block *sb);
fs/ocfs2/super.c
91
static int ocfs2_mount_volume(struct super_block *sb);
fs/ocfs2/super.c
910
struct super_block *sb = osb->sb;
fs/ocfs2/super.c
92
static void ocfs2_dismount_volume(struct super_block *sb, int mnt_err);
fs/ocfs2/super.c
920
sb_dqopt(sb)->flags |= DQUOT_QUOTA_SYS_FILE | DQUOT_NEGATIVE_USAGE;
fs/ocfs2/super.c
922
if (!OCFS2_HAS_RO_COMPAT_FEATURE(sb, feature[type]))
fs/ocfs2/super.c
951
struct super_block *sb = osb->sb;
fs/ocfs2/super.c
957
if (!sb_has_quota_loaded(sb, type))
fs/ocfs2/super.c
959
if (!sb_has_quota_suspended(sb, type)) {
fs/ocfs2/super.c
960
oinfo = sb_dqinfo(sb, type)->dqi_priv;
fs/ocfs2/super.c
963
inode = igrab(sb->s_dquot.files[type]);
fs/ocfs2/super.c
967
dquot_disable(sb, type, DQUOT_USAGE_ENABLED |
fs/ocfs2/super.c
973
static int ocfs2_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ocfs2/super.c
984
trace_ocfs2_fill_super(sb, fc, fc->sb_flags & SB_SILENT);
fs/ocfs2/super.c
987
status = ocfs2_sb_probe(sb, &bh, §or_size, &stats);
fs/ocfs2/super.c
99
static int ocfs2_sync_fs(struct super_block *sb, int wait);
fs/ocfs2/super.c
993
status = ocfs2_initialize_super(sb, bh, sector_size, &stats);
fs/ocfs2/super.c
999
osb = OCFS2_SB(sb);
fs/ocfs2/super.h
14
int __ocfs2_error(struct super_block *sb, const char *function,
fs/ocfs2/super.h
17
#define ocfs2_error(sb, fmt, ...) \
fs/ocfs2/super.h
18
__ocfs2_error(sb, __PRETTY_FUNCTION__, fmt, ##__VA_ARGS__)
fs/ocfs2/super.h
21
void __ocfs2_abort(struct super_block *sb, const char *function,
fs/ocfs2/super.h
24
#define ocfs2_abort(sb, fmt, ...) \
fs/ocfs2/super.h
25
__ocfs2_abort(sb, __PRETTY_FUNCTION__, fmt, ##__VA_ARGS__)
fs/ocfs2/uptodate.c
600
struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
fs/ocfs2/uptodate.c
601
unsigned int i, b_len = ocfs2_clusters_to_blocks(sb, 1) * c_len;
fs/ocfs2/xattr.c
1779
static int ocfs2_bucket_align_free_start(struct super_block *sb,
fs/ocfs2/xattr.c
1786
if (((free_start - size) >> sb->s_blocksize_bits) !=
fs/ocfs2/xattr.c
1787
((free_start - 1) >> sb->s_blocksize_bits))
fs/ocfs2/xattr.c
1788
free_start -= free_start % sb->s_blocksize;
fs/ocfs2/xattr.c
1801
struct super_block *sb = loc->xl_inode->i_sb;
fs/ocfs2/xattr.c
1832
free_start = ocfs2_bucket_align_free_start(sb, free_start,
fs/ocfs2/xattr.c
1883
struct super_block *sb = loc->xl_inode->i_sb;
fs/ocfs2/xattr.c
1886
free_start = ocfs2_bucket_align_free_start(sb, free_start, size);
fs/ocfs2/xattr.c
1898
struct super_block *sb = loc->xl_inode->i_sb;
fs/ocfs2/xattr.c
1901
int block_offset = nameval_offset >> sb->s_blocksize_bits;
fs/ocfs2/xattr.c
1905
((nameval_offset + size - 1) >> sb->s_blocksize_bits));
fs/ocfs2/xattr.c
241
static int ocfs2_xattr_bucket_get_name_value(struct super_block *sb,
fs/ocfs2/xattr.c
2425
ocfs2_remove_extent_credits(osb->sb));
fs/ocfs2/xattr.c
300
static int ocfs2_get_xattr_tree_value_root(struct super_block *sb,
fs/ocfs2/xattr.c
311
static inline u16 ocfs2_blocks_per_xattr_bucket(struct super_block *sb)
fs/ocfs2/xattr.c
313
return OCFS2_XATTR_BUCKET_SIZE / (1 << sb->s_blocksize_bits);
fs/ocfs2/xattr.c
4059
static int ocfs2_xattr_bucket_get_name_value(struct super_block *sb,
fs/ocfs2/xattr.c
4072
*block_off = name_offset >> sb->s_blocksize_bits;
fs/ocfs2/xattr.c
4073
*new_offset = name_offset % sb->s_blocksize;
fs/ocfs2/xattr.c
4540
struct super_block *sb = inode->i_sb;
fs/ocfs2/xattr.c
4541
int blks_per_bucket = ocfs2_blocks_per_xattr_bucket(sb);
fs/ocfs2/xattr.c
4542
int num_buckets = ocfs2_xattr_buckets_per_cluster(OCFS2_SB(sb));
fs/ocfs2/xattr.c
4546
((num_clusters - 1) * ocfs2_clusters_to_blocks(sb, 1));
fs/ocfs2/xattr.c
4549
BUG_ON(OCFS2_XATTR_BUCKET_SIZE == OCFS2_SB(sb)->s_clustersize);
fs/ocfs2/xattr.c
467
static int ocfs2_validate_xattr_block(struct super_block *sb,
fs/ocfs2/xattr.c
483
rc = ocfs2_validate_meta_ecc(sb, bh->b_data, &xb->xb_check);
fs/ocfs2/xattr.c
492
return ocfs2_error(sb,
fs/ocfs2/xattr.c
499
return ocfs2_error(sb,
fs/ocfs2/xattr.c
505
if (le32_to_cpu(xb->xb_fs_generation) != OCFS2_SB(sb)->fs_generation) {
fs/ocfs2/xattr.c
506
return ocfs2_error(sb,
fs/ocfs2/xattr.c
5161
block = ocfs2_clusters_to_blocks(osb->sb, bit_off);
fs/ocfs2/xattr.c
5492
handle = ocfs2_start_trans(osb, ocfs2_remove_extent_credits(osb->sb));
fs/ocfs2/xattr.c
5687
int credits = ocfs2_remove_extent_credits(osb->sb) +
fs/ocfs2/xattr.c
6024
static int ocfs2_get_xattr_tree_value_root(struct super_block *sb,
fs/ocfs2/xattr.c
6035
ret = ocfs2_xattr_bucket_get_name_value(sb,
fs/ocfs2/xattr.c
6225
typedef int (get_xattr_value_root)(struct super_block *sb,
fs/ocfs2/xattr.c
6238
static int ocfs2_value_metas_in_xattr_header(struct super_block *sb,
fs/ocfs2/xattr.c
6255
ret = func(sb, bh, xh, i, &xv, NULL, para);
fs/ocfs2/xattr.c
6264
*credits += ocfs2_calc_extend_credits(sb,
fs/ocfs2/xattr.c
6274
*num_recs += ocfs2_clusters_for_bytes(sb,
fs/ocfs2/xattr.c
6282
static int ocfs2_get_xattr_value_root(struct super_block *sb,
fs/ocfs2/xattr.c
6318
ret = ocfs2_value_metas_in_xattr_header(osb->sb, NULL, xh,
fs/ocfs2/xattr.c
6334
num_recs = num_recs / ocfs2_refcount_recs_per_rb(osb->sb) * 2;
fs/ocfs2/xattr.c
6374
struct super_block *sb = args->old_inode->i_sb;
fs/ocfs2/xattr.c
6415
ret = func(sb, old_bh, xh, i, &xv, NULL, para);
fs/ocfs2/xattr.c
6421
ret = func(sb, new_bh, new_xh, j, &new_xv, &value_bh, para);
fs/ocfs2/xattr.c
6499
int header_off = osb->sb->s_blocksize - inline_size;
fs/ocfs2/xattr.c
6657
osb->sb->s_blocksize - header_off);
fs/ocfs2/xattr.c
6701
static int ocfs2_get_reflink_xattr_value_root(struct super_block *sb,
fs/ocfs2/xattr.c
6718
return ocfs2_get_xattr_tree_value_root(sb, bucket, offset,
fs/ocfs2/xattr.c
6728
static int ocfs2_value_tree_metas_in_bucket(struct super_block *sb,
fs/ocfs2/xattr.c
6739
return ocfs2_get_xattr_tree_value_root(sb, bucket, offset,
fs/ocfs2/xattr.c
6802
(metas.num_recs + ocfs2_refcount_recs_per_rb(osb->sb) - 1) /
fs/ocfs2/xattr.c
6803
ocfs2_refcount_recs_per_rb(osb->sb) * 2;
fs/ocfs2/xattr.c
6824
*credits += ocfs2_calc_extend_credits(osb->sb,
fs/ocfs2/xattr.c
6860
struct super_block *sb = args->reflink->old_inode->i_sb;
fs/ocfs2/xattr.c
6890
sb->s_blocksize);
fs/ocfs2/xattr.c
879
static int ocfs2_xattr_list_entry(struct super_block *sb,
fs/ocfs2/xattr.c
891
if (OCFS2_SB(sb)->s_mount_opt & OCFS2_MOUNT_NOUSERXATTR)
fs/ocfs2/xattr.c
897
if (!(sb->s_flags & SB_POSIXACL))
fs/omfs/bitmap.c
109
bh = sb_bread(sb, clus_to_blk(sbi, sbi->s_bitmap_ino) + map);
fs/omfs/bitmap.c
12
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/bitmap.c
13
int nbits = sb->s_blocksize * 8;
fs/omfs/bitmap.c
130
int omfs_allocate_range(struct super_block *sb,
fs/omfs/bitmap.c
136
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/bitmap.c
137
int bits_per_entry = 8 * sb->s_blocksize;
fs/omfs/bitmap.c
165
ret = set_run(sb, i, bits_per_entry, bit, run, 1);
fs/omfs/bitmap.c
175
int omfs_clear_range(struct super_block *sb, u64 block, int count)
fs/omfs/bitmap.c
177
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/bitmap.c
178
int bits_per_entry = 8 * sb->s_blocksize;
fs/omfs/bitmap.c
191
ret = set_run(sb, map, bits_per_entry, bit, count, 0);
fs/omfs/bitmap.c
48
static int set_run(struct super_block *sb, int map,
fs/omfs/bitmap.c
54
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/bitmap.c
57
bh = sb_bread(sb, clus_to_blk(sbi, sbi->s_bitmap_ino) + map);
fs/omfs/bitmap.c
68
bh = sb_bread(sb,
fs/omfs/bitmap.c
8
unsigned long omfs_count_free(struct super_block *sb)
fs/omfs/bitmap.c
91
int omfs_allocate_block(struct super_block *sb, u64 block)
fs/omfs/bitmap.c
94
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/bitmap.c
95
int bits_per_entry = 8 * sb->s_blocksize;
fs/omfs/dir.c
84
int omfs_make_empty(struct inode *inode, struct super_block *sb)
fs/omfs/dir.c
86
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/dir.c
90
bh = omfs_bread(sb, inode->i_ino);
fs/omfs/inode.c
204
struct inode *omfs_iget(struct super_block *sb, ino_t ino)
fs/omfs/inode.c
206
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/inode.c
213
inode = iget_locked(sb, ino);
fs/omfs/inode.c
26
struct buffer_head *omfs_bread(struct super_block *sb, sector_t block)
fs/omfs/inode.c
265
static void omfs_put_super(struct super_block *sb)
fs/omfs/inode.c
267
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/inode.c
270
sb->s_fs_info = NULL;
fs/omfs/inode.c
28
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/inode.c
32
return sb_bread(sb, clus_to_blk(sbi, block));
fs/omfs/inode.c
335
static int omfs_get_imap(struct super_block *sb)
fs/omfs/inode.c
339
struct omfs_sb_info *sbi = OMFS_SB(sb);
fs/omfs/inode.c
345
array_size = DIV_ROUND_UP(bitmap_size, sb->s_blocksize);
fs/omfs/inode.c
360
for (count = bitmap_size; count > 0; count -= sb->s_blocksize) {
fs/omfs/inode.c
361
bh = sb_bread(sb, block++);
fs/omfs/inode.c
364
*ptr = kmemdup(bh->b_data, sb->s_blocksize, GFP_KERNEL);
fs/omfs/inode.c
369
if (count < sb->s_blocksize)
fs/omfs/inode.c
371
sb->s_blocksize - count);
fs/omfs/inode.c
456
static int omfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/omfs/inode.c
471
sb->s_fs_info = sbi;
fs/omfs/inode.c
475
sb->s_maxbytes = 0xffffffff;
fs/omfs/inode.c
477
sb->s_time_gran = NSEC_PER_MSEC;
fs/omfs/inode.c
478
sb->s_time_min = 0;
fs/omfs/inode.c
479
sb->s_time_max = U64_MAX / MSEC_PER_SEC;
fs/omfs/inode.c
481
sb_set_blocksize(sb, 0x200);
fs/omfs/inode.c
483
bh = sb_bread(sb, 0);
fs/omfs/inode.c
495
sb->s_magic = OMFS_MAGIC;
fs/omfs/inode.c
527
sb_set_blocksize(sb, sbi->s_sys_blocksize);
fs/omfs/inode.c
536
bh2 = omfs_bread(sb, be64_to_cpu(omfs_sb->s_root_block));
fs/omfs/inode.c
568
ret = omfs_get_imap(sb);
fs/omfs/inode.c
572
sb->s_op = &omfs_sops;
fs/omfs/inode.c
574
root = omfs_iget(sb, be64_to_cpu(omfs_rb->r_root_dir));
fs/omfs/inode.c
580
sb->s_root = d_make_root(root);
fs/omfs/inode.c
581
if (!sb->s_root) {
fs/omfs/omfs.h
35
static inline struct omfs_sb_info *OMFS_SB(struct super_block *sb)
fs/omfs/omfs.h
37
return sb->s_fs_info;
fs/omfs/omfs.h
41
extern unsigned long omfs_count_free(struct super_block *sb);
fs/omfs/omfs.h
42
extern int omfs_allocate_block(struct super_block *sb, u64 block);
fs/omfs/omfs.h
43
extern int omfs_allocate_range(struct super_block *sb, int min_request,
fs/omfs/omfs.h
45
extern int omfs_clear_range(struct super_block *sb, u64 block, int count);
fs/omfs/omfs.h
50
extern int omfs_make_empty(struct inode *inode, struct super_block *sb);
fs/omfs/omfs.h
62
extern struct buffer_head *omfs_bread(struct super_block *sb, sector_t block);
fs/omfs/omfs.h
63
extern struct inode *omfs_iget(struct super_block *sb, ino_t inode);
fs/omfs/omfs.h
65
extern int omfs_reserve_block(struct super_block *sb, sector_t block);
fs/omfs/omfs.h
66
extern int omfs_find_empty_block(struct super_block *sb, int mode, ino_t *ino);
fs/openpromfs/inode.c
334
static struct inode *openprom_alloc_inode(struct super_block *sb)
fs/openpromfs/inode.c
338
oi = alloc_inode_sb(sb, op_inode_cachep, GFP_KERNEL);
fs/openpromfs/inode.c
350
static struct inode *openprom_iget(struct super_block *sb, ino_t ino)
fs/openpromfs/inode.c
352
struct inode *inode = iget_locked(sb, ino);
fs/openpromfs/inode.c
43
static struct inode *openprom_iget(struct super_block *sb, ino_t ino);
fs/orangefs/inode.c
1034
struct inode *orangefs_iget(struct super_block *sb,
fs/orangefs/inode.c
1042
inode = iget5_locked(sb,
fs/orangefs/inode.c
1077
struct inode *orangefs_new_inode(struct super_block *sb, struct inode *dir,
fs/orangefs/inode.c
1088
sb,
fs/orangefs/inode.c
1093
inode = new_inode(sb);
fs/orangefs/orangefs-kernel.h
208
struct super_block *sb;
fs/orangefs/orangefs-kernel.h
247
static inline struct orangefs_sb_info_s *ORANGEFS_SB(struct super_block *sb)
fs/orangefs/orangefs-kernel.h
249
return (struct orangefs_sb_info_s *) sb->s_fs_info;
fs/orangefs/orangefs-kernel.h
337
void orangefs_kill_sb(struct super_block *sb);
fs/orangefs/orangefs-kernel.h
347
struct inode *orangefs_new_inode(struct super_block *sb,
fs/orangefs/orangefs-kernel.h
374
struct inode *orangefs_iget(struct super_block *sb,
fs/orangefs/super.c
148
struct super_block *sb = NULL;
fs/orangefs/super.c
150
sb = dentry->d_sb;
fs/orangefs/super.c
155
sb,
fs/orangefs/super.c
156
(int)(ORANGEFS_SB(sb)->fs_id));
fs/orangefs/super.c
161
new_op->upcall.req.statfs.fs_id = ORANGEFS_SB(sb)->fs_id;
fs/orangefs/super.c
163
if (ORANGEFS_SB(sb)->flags & ORANGEFS_OPT_INTR)
fs/orangefs/super.c
182
buf->f_type = sb->s_magic;
fs/orangefs/super.c
183
buf->f_fsid.val[0] = ORANGEFS_SB(sb)->fs_id;
fs/orangefs/super.c
184
buf->f_fsid.val[1] = ORANGEFS_SB(sb)->id;
fs/orangefs/super.c
207
struct super_block *sb = fc->root->d_sb;
fs/orangefs/super.c
208
struct orangefs_sb_info_s *orangefs_sb = ORANGEFS_SB(sb);
fs/orangefs/super.c
314
static struct dentry *orangefs_fh_to_dentry(struct super_block *sb,
fs/orangefs/super.c
331
return d_obtain_alias(orangefs_iget(sb, &refn));
fs/orangefs/super.c
400
static int orangefs_fill_sb(struct super_block *sb,
fs/orangefs/super.c
409
ORANGEFS_SB(sb)->sb = sb;
fs/orangefs/super.c
411
ORANGEFS_SB(sb)->root_khandle = fs_mount->root_khandle;
fs/orangefs/super.c
412
ORANGEFS_SB(sb)->fs_id = fs_mount->fs_id;
fs/orangefs/super.c
413
ORANGEFS_SB(sb)->id = fs_mount->id;
fs/orangefs/super.c
416
sb->s_xattr = orangefs_xattr_handlers;
fs/orangefs/super.c
417
sb->s_magic = ORANGEFS_SUPER_MAGIC;
fs/orangefs/super.c
418
sb->s_op = &orangefs_s_ops;
fs/orangefs/super.c
419
set_default_d_op(sb, &orangefs_dentry_operations);
fs/orangefs/super.c
421
sb->s_blocksize = PAGE_SIZE;
fs/orangefs/super.c
422
sb->s_blocksize_bits = PAGE_SHIFT;
fs/orangefs/super.c
423
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/orangefs/super.c
425
ret = super_setup_bdi(sb);
fs/orangefs/super.c
429
root_object.khandle = ORANGEFS_SB(sb)->root_khandle;
fs/orangefs/super.c
430
root_object.fs_id = ORANGEFS_SB(sb)->fs_id;
fs/orangefs/super.c
436
root = orangefs_iget(sb, &root_object);
fs/orangefs/super.c
450
sb->s_export_op = &orangefs_export_ops;
fs/orangefs/super.c
451
sb->s_root = root_dentry;
fs/orangefs/super.c
458
struct super_block *sb = ERR_PTR(-EINVAL);
fs/orangefs/super.c
490
sb = sget_fc(fc, NULL, set_anon_super_fc);
fs/orangefs/super.c
492
if (IS_ERR(sb)) {
fs/orangefs/super.c
493
ret = PTR_ERR(sb);
fs/orangefs/super.c
501
ret = orangefs_fill_sb(sb, fc, &new_op->downcall.resp.fs_mount);
fs/orangefs/super.c
510
strscpy(ORANGEFS_SB(sb)->devname, fc->source);
fs/orangefs/super.c
513
ORANGEFS_SB(sb)->mount_pending = 0;
fs/orangefs/super.c
521
ORANGEFS_SB(sb));
fs/orangefs/super.c
523
list_add_tail(&ORANGEFS_SB(sb)->list, &orangefs_superblocks);
fs/orangefs/super.c
528
ORANGEFS_SB(sb)->no_list = 0;
fs/orangefs/super.c
542
fc->root = dget(sb->s_root);
fs/orangefs/super.c
547
ORANGEFS_SB(sb)->no_list = 1;
fs/orangefs/super.c
549
deactivate_locked_super(sb);
fs/orangefs/super.c
598
void orangefs_kill_sb(struct super_block *sb)
fs/orangefs/super.c
604
kill_anon_super(sb);
fs/orangefs/super.c
606
if (!ORANGEFS_SB(sb)) {
fs/orangefs/super.c
615
r = orangefs_unmount(ORANGEFS_SB(sb)->id, ORANGEFS_SB(sb)->fs_id,
fs/orangefs/super.c
616
ORANGEFS_SB(sb)->devname);
fs/orangefs/super.c
618
ORANGEFS_SB(sb)->mount_pending = 1;
fs/orangefs/super.c
620
if (!ORANGEFS_SB(sb)->no_list) {
fs/orangefs/super.c
624
__list_del_entry(&ORANGEFS_SB(sb)->list);
fs/orangefs/super.c
625
ORANGEFS_SB(sb)->list.prev = NULL;
fs/orangefs/super.c
637
kfree(ORANGEFS_SB(sb));
fs/orangefs/super.c
85
static struct inode *orangefs_alloc_inode(struct super_block *sb)
fs/orangefs/super.c
89
orangefs_inode = alloc_inode_sb(sb, orangefs_inode_cache, GFP_KERNEL);
fs/overlayfs/copy_up.c
113
if (ovl_is_private_xattr(sb, name))
fs/overlayfs/copy_up.c
125
error = ovl_copy_acl(OVL_FS(sb), oldpath, new, name);
fs/overlayfs/copy_up.c
156
error = ovl_do_setxattr(OVL_FS(sb), new, name, value, size, 0);
fs/overlayfs/copy_up.c
75
int ovl_copy_xattr(struct super_block *sb, const struct path *oldpath, struct dentry *new)
fs/overlayfs/export.c
294
static struct dentry *ovl_obtain_alias(struct super_block *sb,
fs/overlayfs/export.c
321
inode = ovl_get_inode(sb, &oip);
fs/overlayfs/export.c
417
static struct dentry *ovl_lookup_real(struct super_block *sb,
fs/overlayfs/export.c
424
static struct dentry *ovl_lookup_real_inode(struct super_block *sb,
fs/overlayfs/export.c
428
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/export.c
437
inode = ovl_lookup_inode(sb, real, !layer->idx);
fs/overlayfs/export.c
449
if (!this && layer->idx && ovl_indexdir(sb) && !WARN_ON(!d_is_dir(real))) {
fs/overlayfs/export.c
470
this = ovl_lookup_real(sb, upper, &ofs->layers[0]);
fs/overlayfs/export.c
489
static struct dentry *ovl_lookup_real_ancestor(struct super_block *sb,
fs/overlayfs/export.c
497
return dget(sb->s_root);
fs/overlayfs/export.c
508
ancestor = ovl_lookup_real_inode(sb, next, layer);
fs/overlayfs/export.c
513
ancestor = dget(sb->s_root);
fs/overlayfs/export.c
542
static struct dentry *ovl_lookup_real(struct super_block *sb,
fs/overlayfs/export.c
549
connected = ovl_lookup_real_ancestor(sb, real, layer);
fs/overlayfs/export.c
580
connected = dget(sb->s_root);
fs/overlayfs/export.c
615
this = ovl_lookup_real_ancestor(sb, real,
fs/overlayfs/export.c
644
static struct dentry *ovl_get_dentry(struct super_block *sb,
fs/overlayfs/export.c
649
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/export.c
658
return ovl_obtain_alias(sb, upper, lowerpath, index);
fs/overlayfs/export.c
668
return ovl_lookup_real(sb, real, layer);
fs/overlayfs/export.c
671
static struct dentry *ovl_upper_fh_to_d(struct super_block *sb,
fs/overlayfs/export.c
674
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/export.c
685
dentry = ovl_get_dentry(sb, upper, NULL, NULL);
fs/overlayfs/export.c
691
static struct dentry *ovl_lower_fh_to_d(struct super_block *sb,
fs/overlayfs/export.c
694
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/export.c
709
inode = ovl_lookup_inode(sb, origin.dentry, false);
fs/overlayfs/export.c
722
if (ovl_indexdir(sb)) {
fs/overlayfs/export.c
739
dentry = ovl_get_dentry(sb, upper, NULL, NULL);
fs/overlayfs/export.c
759
dentry = ovl_get_dentry(sb, NULL, &origin, index);
fs/overlayfs/export.c
794
static struct dentry *ovl_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/overlayfs/export.c
814
ovl_upper_fh_to_d(sb, fh) :
fs/overlayfs/export.c
815
ovl_lower_fh_to_d(sb, fh);
fs/overlayfs/export.c
834
static struct dentry *ovl_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/overlayfs/inode.c
1025
struct inode *ovl_new_inode(struct super_block *sb, umode_t mode, dev_t rdev)
fs/overlayfs/inode.c
1029
inode = new_inode(sb);
fs/overlayfs/inode.c
1085
struct inode *ovl_lookup_inode(struct super_block *sb, struct dentry *real,
fs/overlayfs/inode.c
1090
inode = ilookup5(sb, (unsigned long) key, ovl_inode_test, key);
fs/overlayfs/inode.c
1103
bool ovl_lookup_trap_inode(struct super_block *sb, struct dentry *dir)
fs/overlayfs/inode.c
1109
trap = ilookup5(sb, (unsigned long) key, ovl_inode_test, key);
fs/overlayfs/inode.c
1125
struct inode *ovl_get_trap_inode(struct super_block *sb, struct dentry *dir)
fs/overlayfs/inode.c
1133
trap = iget5_locked(sb, (unsigned long) key, ovl_inode_test,
fs/overlayfs/inode.c
1154
static bool ovl_hash_bylower(struct super_block *sb, struct dentry *upper,
fs/overlayfs/inode.c
1157
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/inode.c
1172
if ((upper || !ovl_indexdir(sb)) &&
fs/overlayfs/inode.c
1184
static struct inode *ovl_iget5(struct super_block *sb, struct inode *newinode,
fs/overlayfs/inode.c
1189
iget5_locked(sb, (unsigned long) key,
fs/overlayfs/inode.c
1193
struct inode *ovl_get_inode(struct super_block *sb,
fs/overlayfs/inode.c
1196
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/inode.c
1206
bool bylower = ovl_hash_bylower(sb, upperdentry, lowerdentry,
fs/overlayfs/inode.c
1226
inode = ovl_iget5(sb, oip->newinode, key);
fs/overlayfs/inode.c
1256
inode = new_inode(sb);
fs/overlayfs/inode.c
1268
if (upperdentry && ovl_is_impuredir(sb, upperdentry))
fs/overlayfs/inode.c
154
static inline int ovl_real_getattr_nosec(struct super_block *sb,
fs/overlayfs/inode.c
159
with_ovl_creds(sb)
fs/overlayfs/inode.c
167
struct super_block *sb = dentry->d_sb;
fs/overlayfs/inode.c
179
err = ovl_real_getattr_nosec(sb, &realpath, stat, request_mask, flags);
fs/overlayfs/inode.c
204
err = ovl_real_getattr_nosec(sb, &realpath, &lowerstat, lowermask, flags);
fs/overlayfs/inode.c
256
err = ovl_real_getattr_nosec(sb, &realpath, &lowerdatastat,
fs/overlayfs/namei.c
1066
struct ovl_fs *ofs = OVL_FS(d->sb);
fs/overlayfs/namei.c
1393
.sb = dentry->d_sb,
fs/overlayfs/namei.c
162
bool ovl_uuid_match(struct ovl_fs *ofs, const struct super_block *sb,
fs/overlayfs/namei.c
170
return ovl_origin_uuid(ofs) ? uuid_equal(uuid, &sb->s_uuid) :
fs/overlayfs/namei.c
19
struct super_block *sb;
fs/overlayfs/namei.c
238
struct ovl_fs *ofs = OVL_FS(d->sb);
fs/overlayfs/namei.c
311
if (ovl_lookup_trap_inode(d->sb, this)) {
fs/overlayfs/namei.c
40
struct ovl_fs *ofs = OVL_FS(d->sb);
fs/overlayfs/overlayfs.h
463
const struct cred *ovl_override_creds(struct super_block *sb);
fs/overlayfs/overlayfs.h
465
EXTEND_CLASS(override_creds, _ovl, ovl_override_creds(sb), struct super_block *sb)
fs/overlayfs/overlayfs.h
467
#define with_ovl_creds(sb) \
fs/overlayfs/overlayfs.h
468
scoped_class(override_creds_ovl, __UNIQUE_ID(label), sb)
fs/overlayfs/overlayfs.h
470
static inline const struct cred *ovl_creds(struct super_block *sb)
fs/overlayfs/overlayfs.h
472
return OVL_FS(sb)->creator_cred;
fs/overlayfs/overlayfs.h
475
int ovl_can_decode_fh(struct super_block *sb);
fs/overlayfs/overlayfs.h
476
struct dentry *ovl_indexdir(struct super_block *sb);
fs/overlayfs/overlayfs.h
477
bool ovl_index_all(struct super_block *sb);
fs/overlayfs/overlayfs.h
478
bool ovl_verify_lower(struct super_block *sb);
fs/overlayfs/overlayfs.h
552
bool ovl_init_uuid_xattr(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/overlayfs.h
616
static inline bool ovl_is_impuredir(struct super_block *sb,
fs/overlayfs/overlayfs.h
619
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/overlayfs.h
734
bool ovl_uuid_match(struct ovl_fs *ofs, const struct super_block *sb,
fs/overlayfs/overlayfs.h
848
bool ovl_is_private_xattr(struct super_block *sb, const char *name);
fs/overlayfs/overlayfs.h
860
struct inode *ovl_new_inode(struct super_block *sb, umode_t mode, dev_t rdev);
fs/overlayfs/overlayfs.h
861
struct inode *ovl_lookup_inode(struct super_block *sb, struct dentry *real,
fs/overlayfs/overlayfs.h
863
bool ovl_lookup_trap_inode(struct super_block *sb, struct dentry *dir);
fs/overlayfs/overlayfs.h
864
struct inode *ovl_get_trap_inode(struct super_block *sb, struct dentry *dir);
fs/overlayfs/overlayfs.h
865
struct inode *ovl_get_inode(struct super_block *sb,
fs/overlayfs/overlayfs.h
934
int ovl_copy_xattr(struct super_block *sb, const struct path *path, struct dentry *new);
fs/overlayfs/overlayfs.h
947
int ovl_fill_super(struct super_block *sb, struct fs_context *fc);
fs/overlayfs/ovl_entry.h
115
static inline struct ovl_fs *OVL_FS(struct super_block *sb)
fs/overlayfs/ovl_entry.h
118
WARN_ON_ONCE(sb->s_type != &ovl_fs_type);
fs/overlayfs/ovl_entry.h
120
return (struct ovl_fs *)sb->s_fs_info;
fs/overlayfs/ovl_entry.h
25
struct super_block *sb;
fs/overlayfs/params.c
1051
struct super_block *sb = dentry->d_sb;
fs/overlayfs/params.c
1052
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/params.c
762
struct super_block *sb = fc->root->d_sb;
fs/overlayfs/params.c
763
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/params.c
770
if (fc->sb_flags & SB_RDONLY && !sb_rdonly(sb)) {
fs/overlayfs/super.c
1010
static int ovl_set_encoding(struct super_block *sb, struct super_block *fs_sb)
fs/overlayfs/super.c
1021
sb->s_encoding = fs_sb->s_encoding;
fs/overlayfs/super.c
1022
sb->s_encoding_flags = fs_sb->s_encoding_flags;
fs/overlayfs/super.c
1027
static int ovl_get_layers(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
1057
ofs->fs[0].sb = ovl_upper_mnt(ofs)->mnt_sb;
fs/overlayfs/super.c
1061
err = ovl_set_encoding(sb, ofs->fs[0].sb);
fs/overlayfs/super.c
1088
err = ovl_setup_trap(sb, l->path.dentry, &trap, "lowerdir");
fs/overlayfs/super.c
1126
if (!ovl_upper_mnt(ofs) && !sb_has_encoding(sb)) {
fs/overlayfs/super.c
1127
err = ovl_set_encoding(sb, ofs->fs[fsid].sb);
fs/overlayfs/super.c
1132
if (!sb_same_encoding(sb, mnt->mnt_sb)) {
fs/overlayfs/super.c
1174
static struct ovl_entry *ovl_get_lowerstack(struct super_block *sb,
fs/overlayfs/super.c
1201
err = ovl_lower_dir(l->name, &l->path, ofs, &sb->s_stack_depth);
fs/overlayfs/super.c
1207
sb->s_stack_depth++;
fs/overlayfs/super.c
1208
if (sb->s_stack_depth > FILESYSTEM_MAX_STACK_DEPTH) {
fs/overlayfs/super.c
1213
err = ovl_get_layers(sb, ofs, ctx, layers);
fs/overlayfs/super.c
1240
static int ovl_check_layer(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
1254
if (is_lower && ovl_lookup_trap_inode(sb, parent)) {
fs/overlayfs/super.c
1273
static int ovl_check_overlapping_layers(struct super_block *sb,
fs/overlayfs/super.c
1279
err = ovl_check_layer(sb, ofs, ovl_upper_mnt(ofs)->mnt_root,
fs/overlayfs/super.c
1291
err = ovl_check_layer(sb, ofs, ofs->workbasedir, "workdir",
fs/overlayfs/super.c
1298
err = ovl_check_layer(sb, ofs,
fs/overlayfs/super.c
1308
static struct dentry *ovl_get_root(struct super_block *sb,
fs/overlayfs/super.c
1313
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/super.c
1322
root = d_make_root(ovl_new_inode(sb, S_IFDIR, 0));
fs/overlayfs/super.c
1331
if (ovl_is_impuredir(sb, upperdentry))
fs/overlayfs/super.c
1362
static void ovl_set_d_op(struct super_block *sb)
fs/overlayfs/super.c
1365
struct ovl_fs *ofs = sb->s_fs_info;
fs/overlayfs/super.c
1368
set_default_d_op(sb, &ovl_dentry_ci_operations);
fs/overlayfs/super.c
1372
set_default_d_op(sb, &ovl_dentry_operations);
fs/overlayfs/super.c
1375
static int ovl_fill_super_creds(struct fs_context *fc, struct super_block *sb)
fs/overlayfs/super.c
1377
struct ovl_fs *ofs = sb->s_fs_info;
fs/overlayfs/super.c
1415
sb->s_stack_depth = 0;
fs/overlayfs/super.c
1416
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/overlayfs/super.c
1428
sb->s_op = &ovl_super_operations;
fs/overlayfs/super.c
1439
err = ovl_get_upper(sb, ofs, &layers[0], &ctx->upper);
fs/overlayfs/super.c
1453
err = ovl_get_workdir(sb, ofs, &ctx->upper, &ctx->work);
fs/overlayfs/super.c
1458
sb->s_flags |= SB_RDONLY;
fs/overlayfs/super.c
1460
sb->s_stack_depth = upper_sb->s_stack_depth;
fs/overlayfs/super.c
1461
sb->s_time_gran = upper_sb->s_time_gran;
fs/overlayfs/super.c
1463
oe = ovl_get_lowerstack(sb, ctx, ofs, layers);
fs/overlayfs/super.c
1470
sb->s_flags |= SB_RDONLY;
fs/overlayfs/super.c
1474
ovl_init_uuid_xattr(sb, ofs, &ctx->upper);
fs/overlayfs/super.c
1478
err = ovl_get_indexdir(sb, ofs, oe, &ctx->upper);
fs/overlayfs/super.c
1484
sb->s_flags |= SB_RDONLY;
fs/overlayfs/super.c
1487
err = ovl_check_overlapping_layers(sb, ofs);
fs/overlayfs/super.c
1511
sb->s_export_op = &ovl_export_operations;
fs/overlayfs/super.c
1513
sb->s_export_op = &ovl_export_fid_operations;
fs/overlayfs/super.c
1518
sb->s_magic = OVERLAYFS_SUPER_MAGIC;
fs/overlayfs/super.c
1519
sb->s_xattr = ovl_xattr_handlers(ofs);
fs/overlayfs/super.c
1520
sb->s_fs_info = ofs;
fs/overlayfs/super.c
1522
sb->s_flags |= SB_POSIXACL;
fs/overlayfs/super.c
1524
sb->s_iflags |= SB_I_SKIP_SYNC;
fs/overlayfs/super.c
1530
sb->s_iflags |= SB_I_NOUMASK;
fs/overlayfs/super.c
1531
sb->s_iflags |= SB_I_EVM_HMAC_UNSUPPORTED;
fs/overlayfs/super.c
1534
sb->s_root = ovl_get_root(sb, ctx->upper.dentry, oe);
fs/overlayfs/super.c
1535
if (!sb->s_root)
fs/overlayfs/super.c
1545
int ovl_fill_super(struct super_block *sb, struct fs_context *fc)
fs/overlayfs/super.c
1547
struct ovl_fs *ofs = sb->s_fs_info;
fs/overlayfs/super.c
1554
ovl_set_d_op(sb);
fs/overlayfs/super.c
1563
with_ovl_creds(sb)
fs/overlayfs/super.c
1564
err = ovl_fill_super_creds(fc, sb);
fs/overlayfs/super.c
1569
sb->s_fs_info = NULL;
fs/overlayfs/super.c
184
static struct inode *ovl_alloc_inode(struct super_block *sb)
fs/overlayfs/super.c
186
struct ovl_inode *oi = alloc_inode_sb(sb, ovl_inode_cachep, GFP_KERNEL);
fs/overlayfs/super.c
225
static void ovl_put_super(struct super_block *sb)
fs/overlayfs/super.c
227
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/super.c
234
static int ovl_sync_fs(struct super_block *sb, int wait)
fs/overlayfs/super.c
236
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/super.c
278
struct super_block *sb = dentry->d_sb;
fs/overlayfs/super.c
279
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/super.c
280
struct dentry *root_dentry = sb->s_root;
fs/overlayfs/super.c
291
buf->f_fsid = uuid_to_fsid(sb->s_uuid.b);
fs/overlayfs/super.c
468
static int ovl_setup_trap(struct super_block *sb, struct dentry *dir,
fs/overlayfs/super.c
474
trap = ovl_get_trap_inode(sb, dir);
fs/overlayfs/super.c
505
static int ovl_get_upper(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
523
err = ovl_setup_trap(sb, upperpath->dentry, &upper_layer->trap,
fs/overlayfs/super.c
551
sb->s_flags |= SB_NOSEC;
fs/overlayfs/super.c
667
static int ovl_make_workdir(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
689
err = ovl_setup_trap(sb, ofs->workdir, &ofs->workdir_trap, "workdir");
fs/overlayfs/super.c
809
static int ovl_get_workdir(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
835
err = ovl_setup_trap(sb, ofs->workbasedir, &ofs->workbasedir_trap,
fs/overlayfs/super.c
840
return ovl_make_workdir(sb, ofs, workpath);
fs/overlayfs/super.c
843
static int ovl_get_indexdir(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/super.c
877
err = ovl_setup_trap(sb, indexdir, &ofs->workdir_trap,
fs/overlayfs/super.c
943
ovl_uuid_match(ofs, ofs->fs[i].sb, uuid)) {
fs/overlayfs/super.c
954
struct super_block *sb = path->mnt->mnt_sb;
fs/overlayfs/super.c
955
const uuid_t *uuid = ovl_origin_uuid(ofs) ? &sb->s_uuid : &uuid_null;
fs/overlayfs/super.c
963
if (ofs->fs[i].sb == sb)
fs/overlayfs/super.c
991
ofs->fs[ofs->numfs].sb = sb;
fs/overlayfs/util.c
102
bool ovl_index_all(struct super_block *sb)
fs/overlayfs/util.c
104
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/util.c
110
bool ovl_verify_lower(struct super_block *sb)
fs/overlayfs/util.c
112
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/util.c
65
const struct cred *ovl_override_creds(struct super_block *sb)
fs/overlayfs/util.c
67
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/util.c
789
bool ovl_init_uuid_xattr(struct super_block *sb, struct ovl_fs *ofs,
fs/overlayfs/util.c
80
int ovl_can_decode_fh(struct super_block *sb)
fs/overlayfs/util.c
838
super_set_uuid(sb, uuid.b, sizeof(uuid));
fs/overlayfs/util.c
85
if (!exportfs_can_decode_fh(sb->s_export_op))
fs/overlayfs/util.c
88
if (sb->s_export_op->encode_fh == generic_encode_ino32_fh)
fs/overlayfs/util.c
94
struct dentry *ovl_indexdir(struct super_block *sb)
fs/overlayfs/util.c
96
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/xattrs.c
19
static bool ovl_is_own_xattr(struct super_block *sb, const char *name)
fs/overlayfs/xattrs.c
21
struct ovl_fs *ofs = OVL_FS(sb);
fs/overlayfs/xattrs.c
31
bool ovl_is_private_xattr(struct super_block *sb, const char *name)
fs/overlayfs/xattrs.c
33
return ovl_is_own_xattr(sb, name) && !ovl_is_escaped_xattr(sb, name);
fs/overlayfs/xattrs.c
7
static bool ovl_is_escaped_xattr(struct super_block *sb, const char *name)
fs/overlayfs/xattrs.c
9
struct ovl_fs *ofs = sb->s_fs_info;
fs/overlayfs/xattrs.c
91
static bool ovl_can_list(struct super_block *sb, const char *s)
fs/overlayfs/xattrs.c
94
if (ovl_is_private_xattr(sb, s))
fs/pidfs.c
844
static struct dentry *pidfs_fh_to_dentry(struct super_block *sb,
fs/proc/base.c
1944
struct inode *proc_pid_make_inode(struct super_block *sb,
fs/proc/base.c
1953
inode = new_inode(sb);
fs/proc/base.c
1997
static struct inode *proc_pid_make_base_inode(struct super_block *sb,
fs/proc/base.c
2004
inode = proc_pid_make_inode(sb, task, mode);
fs/proc/inode.c
116
struct super_block *sb;
fs/proc/inode.c
124
sb = inode->i_sb;
fs/proc/inode.c
125
if ((sb != old_sb) && !atomic_inc_not_zero(&sb->s_active))
fs/proc/inode.c
129
if (sb != old_sb) {
fs/proc/inode.c
132
old_sb = sb;
fs/proc/inode.c
53
static struct inode *proc_alloc_inode(struct super_block *sb)
fs/proc/inode.c
57
ei = alloc_inode_sb(sb, proc_inode_cachep, GFP_KERNEL);
fs/proc/inode.c
628
struct inode *proc_get_inode(struct super_block *sb, struct proc_dir_entry *de)
fs/proc/inode.c
630
struct inode *inode = new_inode(sb);
fs/proc/proc_sysctl.c
449
static struct inode *proc_sys_make_inode(struct super_block *sb,
fs/proc/proc_sysctl.c
456
inode = new_inode(sb);
fs/proc/root.c
301
struct super_block *sb = fc->root->d_sb;
fs/proc/root.c
302
struct proc_fs_info *fs_info = proc_sb_info(sb);
fs/proc/root.c
304
sync_filesystem(sb);
fs/proc/root.c
346
static void proc_kill_sb(struct super_block *sb)
fs/proc/root.c
348
struct proc_fs_info *fs_info = proc_sb_info(sb);
fs/proc/root.c
350
kill_anon_super(sb);
fs/proc_namespace.c
106
struct super_block *sb = mnt_path.dentry->d_sb;
fs/proc_namespace.c
109
if (sb->s_op->show_devname) {
fs/proc_namespace.c
110
err = sb->s_op->show_devname(m, mnt_path.dentry);
fs/proc_namespace.c
122
show_type(m, sb);
fs/proc_namespace.c
124
err = show_sb_opts(m, sb);
fs/proc_namespace.c
128
if (sb->s_op->show_options)
fs/proc_namespace.c
129
err = sb->s_op->show_options(m, mnt_path.dentry);
fs/proc_namespace.c
139
struct super_block *sb = mnt->mnt_sb;
fs/proc_namespace.c
144
MAJOR(sb->s_dev), MINOR(sb->s_dev));
fs/proc_namespace.c
173
show_type(m, sb);
fs/proc_namespace.c
175
if (sb->s_op->show_devname) {
fs/proc_namespace.c
176
err = sb->s_op->show_devname(m, mnt->mnt_root);
fs/proc_namespace.c
182
seq_puts(m, sb_rdonly(sb) ? " ro" : " rw");
fs/proc_namespace.c
183
err = show_sb_opts(m, sb);
fs/proc_namespace.c
186
if (sb->s_op->show_options)
fs/proc_namespace.c
187
err = sb->s_op->show_options(m, mnt->mnt_root);
fs/proc_namespace.c
198
struct super_block *sb = mnt_path.dentry->d_sb;
fs/proc_namespace.c
203
if (sb->s_op->show_devname) {
fs/proc_namespace.c
204
err = sb->s_op->show_devname(m, mnt_path.dentry);
fs/proc_namespace.c
221
show_type(m, sb);
fs/proc_namespace.c
224
if (sb->s_op->show_stats) {
fs/proc_namespace.c
226
err = sb->s_op->show_stats(m, mnt_path.dentry);
fs/proc_namespace.c
45
static int show_sb_opts(struct seq_file *m, struct super_block *sb)
fs/proc_namespace.c
57
if (sb->s_flags & fs_infop->flag)
fs/proc_namespace.c
61
return security_sb_show_options(m, sb);
fs/proc_namespace.c
92
static void show_type(struct seq_file *m, struct super_block *sb)
fs/proc_namespace.c
94
mangle(m, sb->s_type->name);
fs/proc_namespace.c
95
if (sb->s_subtype) {
fs/proc_namespace.c
97
mangle(m, sb->s_subtype);
fs/pstore/inode.c
218
static struct inode *pstore_get_inode(struct super_block *sb)
fs/pstore/inode.c
220
struct inode *inode = new_inode(sb);
fs/pstore/inode.c
411
static int pstore_fill_super(struct super_block *sb, struct fs_context *fc)
fs/pstore/inode.c
416
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/pstore/inode.c
417
sb->s_blocksize = PAGE_SIZE;
fs/pstore/inode.c
418
sb->s_blocksize_bits = PAGE_SHIFT;
fs/pstore/inode.c
419
sb->s_magic = PSTOREFS_MAGIC;
fs/pstore/inode.c
420
sb->s_op = &pstore_ops;
fs/pstore/inode.c
421
sb->s_time_gran = 1;
fs/pstore/inode.c
425
inode = pstore_get_inode(sb);
fs/pstore/inode.c
432
sb->s_root = d_make_root(inode);
fs/pstore/inode.c
433
if (!sb->s_root)
fs/pstore/inode.c
437
pstore_sb = sb;
fs/pstore/inode.c
464
static void pstore_kill_sb(struct super_block *sb)
fs/pstore/inode.c
467
WARN_ON(pstore_sb && pstore_sb != sb);
fs/pstore/inode.c
469
kill_anon_super(sb);
fs/qnx4/bitmap.c
21
unsigned long qnx4_count_free_blocks(struct super_block *sb)
fs/qnx4/bitmap.c
23
int start = le32_to_cpu(qnx4_sb(sb)->BitMap->di_first_xtnt.xtnt_blk) - 1;
fs/qnx4/bitmap.c
27
int size = le32_to_cpu(qnx4_sb(sb)->BitMap->di_size);
fs/qnx4/bitmap.c
33
if ((bh = sb_bread(sb, start + offset)) == NULL) {
fs/qnx4/inode.c
137
struct super_block *sb = dentry->d_sb;
fs/qnx4/inode.c
138
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/qnx4/inode.c
140
buf->f_type = sb->s_magic;
fs/qnx4/inode.c
141
buf->f_bsize = sb->s_blocksize;
fs/qnx4/inode.c
142
buf->f_blocks = le32_to_cpu(qnx4_sb(sb)->BitMap->di_size) * 8;
fs/qnx4/inode.c
143
buf->f_bfree = qnx4_count_free_blocks(sb);
fs/qnx4/inode.c
156
static const char *qnx4_checkroot(struct super_block *sb,
fs/qnx4/inode.c
166
QNX4DEBUG((KERN_NOTICE "QNX4 filesystem found on dev %s.\n", sb->s_id));
fs/qnx4/inode.c
170
bh = sb_bread(sb, rd + j); /* root dir, first block */
fs/qnx4/inode.c
178
qnx4_sb(sb)->BitMap = kmemdup(rootdir,
fs/qnx4/inode.c
182
if (!qnx4_sb(sb)->BitMap)
fs/qnx4/inode.c
257
static void qnx4_kill_sb(struct super_block *sb)
fs/qnx4/inode.c
259
struct qnx4_sb_info *qs = qnx4_sb(sb);
fs/qnx4/inode.c
260
kill_block_super(sb);
fs/qnx4/inode.c
282
struct inode *qnx4_iget(struct super_block *sb, unsigned long ino)
fs/qnx4/inode.c
290
inode = iget_locked(sb, ino);
fs/qnx4/inode.c
303
sb->s_id, ino);
fs/qnx4/inode.c
309
if (!(bh = sb_bread(sb, block))) {
fs/qnx4/inode.c
311
"%s\n", sb->s_id);
fs/qnx4/inode.c
32
static struct inode *qnx4_alloc_inode(struct super_block *sb);
fs/qnx4/inode.c
343
ino, sb->s_id);
fs/qnx4/inode.c
355
static struct inode *qnx4_alloc_inode(struct super_block *sb)
fs/qnx4/inode.c
358
ei = alloc_inode_sb(sb, qnx4_inode_cachep, GFP_KERNEL);
fs/qnx4/inode.c
46
struct super_block *sb = fc->root->d_sb;
fs/qnx4/inode.c
49
sync_filesystem(sb);
fs/qnx4/inode.c
50
qs = qnx4_sb(sb);
fs/qnx4/qnx4.h
26
extern unsigned long qnx4_count_free_blocks(struct super_block *sb);
fs/qnx4/qnx4.h
31
extern int qnx4_is_free(struct super_block *sb, long block);
fs/qnx4/qnx4.h
33
static inline struct qnx4_sb_info *qnx4_sb(struct super_block *sb)
fs/qnx4/qnx4.h
35
return sb->s_fs_info;
fs/qnx6/dir.c
48
static struct qnx6_long_filename *qnx6_longname(struct super_block *sb,
fs/qnx6/dir.c
52
struct qnx6_sb_info *sbi = QNX6_SB(sb);
fs/qnx6/dir.c
54
u32 n = s >> (PAGE_SHIFT - sb->s_blocksize_bits); /* in pages */
fs/qnx6/dir.c
61
offs = offset_in_folio(folio, s << sb->s_blocksize_bits);
fs/qnx6/inode.c
158
struct super_block *sb = dentry->d_sb;
fs/qnx6/inode.c
159
struct qnx6_sb_info *sbi = QNX6_SB(sb);
fs/qnx6/inode.c
160
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/qnx6/inode.c
162
buf->f_type = sb->s_magic;
fs/qnx6/inode.c
163
buf->f_bsize = sb->s_blocksize;
fs/qnx6/inode.c
164
buf->f_blocks = fs32_to_cpu(sbi, sbi->sb->sb_num_blocks);
fs/qnx6/inode.c
165
buf->f_bfree = fs32_to_cpu(sbi, sbi->sb->sb_free_blocks);
fs/qnx6/inode.c
166
buf->f_files = fs32_to_cpu(sbi, sbi->sb->sb_num_inodes);
fs/qnx6/inode.c
167
buf->f_ffree = fs32_to_cpu(sbi, sbi->sb->sb_free_inodes);
fs/qnx6/inode.c
201
void qnx6_superblock_debug(struct qnx6_super_block *sb, struct super_block *s)
fs/qnx6/inode.c
205
pr_debug("magic: %08x\n", fs32_to_cpu(sbi, sb->sb_magic));
fs/qnx6/inode.c
206
pr_debug("checksum: %08x\n", fs32_to_cpu(sbi, sb->sb_checksum));
fs/qnx6/inode.c
207
pr_debug("serial: %llx\n", fs64_to_cpu(sbi, sb->sb_serial));
fs/qnx6/inode.c
208
pr_debug("flags: %08x\n", fs32_to_cpu(sbi, sb->sb_flags));
fs/qnx6/inode.c
209
pr_debug("blocksize: %08x\n", fs32_to_cpu(sbi, sb->sb_blocksize));
fs/qnx6/inode.c
210
pr_debug("num_inodes: %08x\n", fs32_to_cpu(sbi, sb->sb_num_inodes));
fs/qnx6/inode.c
211
pr_debug("free_inodes: %08x\n", fs32_to_cpu(sbi, sb->sb_free_inodes));
fs/qnx6/inode.c
212
pr_debug("num_blocks: %08x\n", fs32_to_cpu(sbi, sb->sb_num_blocks));
fs/qnx6/inode.c
213
pr_debug("free_blocks: %08x\n", fs32_to_cpu(sbi, sb->sb_free_blocks));
fs/qnx6/inode.c
214
pr_debug("inode_levels: %02x\n", sb->Inode.levels);
fs/qnx6/inode.c
256
struct qnx6_super_block *sb;
fs/qnx6/inode.c
265
sb = (struct qnx6_super_block *)bh->b_data;
fs/qnx6/inode.c
266
if (fs32_to_cpu(sbi, sb->sb_magic) != QNX6_SUPER_MAGIC) {
fs/qnx6/inode.c
268
if (fs32_to_cpu(sbi, sb->sb_magic) == QNX6_SUPER_MAGIC) {
fs/qnx6/inode.c
31
static void qnx6_put_super(struct super_block *sb);
fs/qnx6/inode.c
32
static struct inode *qnx6_alloc_inode(struct super_block *sb);
fs/qnx6/inode.c
398
sbi->sb = (struct qnx6_super_block *)bh1->b_data;
fs/qnx6/inode.c
404
sbi->sb = (struct qnx6_super_block *)bh2->b_data;
fs/qnx6/inode.c
474
static void qnx6_put_super(struct super_block *sb)
fs/qnx6/inode.c
476
struct qnx6_sb_info *qs = QNX6_SB(sb);
fs/qnx6/inode.c
48
struct super_block *sb = root->d_sb;
fs/qnx6/inode.c
481
sb->s_fs_info = NULL;
fs/qnx6/inode.c
49
struct qnx6_sb_info *sbi = QNX6_SB(sb);
fs/qnx6/inode.c
511
struct inode *qnx6_iget(struct super_block *sb, unsigned ino)
fs/qnx6/inode.c
513
struct qnx6_sb_info *sbi = QNX6_SB(sb);
fs/qnx6/inode.c
521
inode = iget_locked(sb, ino);
fs/qnx6/inode.c
533
sb->s_id, ino);
fs/qnx6/inode.c
542
sb->s_id);
fs/qnx6/inode.c
58
struct super_block *sb = fc->root->d_sb;
fs/qnx6/inode.c
584
static struct inode *qnx6_alloc_inode(struct super_block *sb)
fs/qnx6/inode.c
587
ei = alloc_inode_sb(sb, qnx6_inode_cachep, GFP_KERNEL);
fs/qnx6/inode.c
60
sync_filesystem(sb);
fs/qnx6/inode.c
65
static unsigned qnx6_get_devblock(struct super_block *sb, __fs32 block)
fs/qnx6/inode.c
67
struct qnx6_sb_info *sbi = QNX6_SB(sb);
fs/qnx6/qnx6.h
31
struct qnx6_super_block *sb; /* our superblock */
fs/qnx6/qnx6.h
47
extern struct inode *qnx6_iget(struct super_block *sb, unsigned ino);
fs/qnx6/qnx6.h
59
static inline struct qnx6_sb_info *QNX6_SB(struct super_block *sb)
fs/qnx6/qnx6.h
61
return sb->s_fs_info;
fs/qnx6/qnx6.h
71
#define test_opt(sb, opt) (QNX6_SB(sb)->s_mount_opt & \
fs/qnx6/super_mmi.c
119
sbi->sb = (struct qnx6_super_block *)bh1->b_data;
fs/qnx6/super_mmi.c
131
sbi->sb = (struct qnx6_super_block *)bh2->b_data;
fs/qnx6/super_mmi.c
141
return sbi->sb;
fs/qnx6/super_mmi.c
19
struct qnx6_mmi_super_block *sb)
fs/qnx6/super_mmi.c
21
qsb->sb_magic = sb->sb_magic;
fs/qnx6/super_mmi.c
22
qsb->sb_checksum = sb->sb_checksum;
fs/qnx6/super_mmi.c
23
qsb->sb_serial = sb->sb_serial;
fs/qnx6/super_mmi.c
24
qsb->sb_blocksize = sb->sb_blocksize;
fs/qnx6/super_mmi.c
25
qsb->sb_num_inodes = sb->sb_num_inodes;
fs/qnx6/super_mmi.c
26
qsb->sb_free_inodes = sb->sb_free_inodes;
fs/qnx6/super_mmi.c
27
qsb->sb_num_blocks = sb->sb_num_blocks;
fs/qnx6/super_mmi.c
28
qsb->sb_free_blocks = sb->sb_free_blocks;
fs/qnx6/super_mmi.c
31
memcpy(&qsb->Inode, &sb->Inode, sizeof(sb->Inode));
fs/qnx6/super_mmi.c
32
memcpy(&qsb->Bitmap, &sb->Bitmap, sizeof(sb->Bitmap));
fs/qnx6/super_mmi.c
33
memcpy(&qsb->Longfile, &sb->Longfile, sizeof(sb->Longfile));
fs/quota/dquot.c
1025
static int add_dquot_ref(struct super_block *sb, int type)
fs/quota/dquot.c
1033
spin_lock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1034
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/quota/dquot.c
1044
spin_unlock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1067
spin_lock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1069
spin_unlock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1074
quota_error(sb, "Writes happened before quota was turned on "
fs/quota/dquot.c
1082
static void remove_dquot_ref(struct super_block *sb, int type)
fs/quota/dquot.c
1089
spin_lock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1090
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/quota/dquot.c
1114
spin_unlock(&sb->s_inode_list_lock);
fs/quota/dquot.c
1119
"inconsistent. Please run quotacheck(8).\n", sb->s_id);
fs/quota/dquot.c
1125
static void drop_dquot_ref(struct super_block *sb, int type)
fs/quota/dquot.c
1127
if (sb->dq_op)
fs/quota/dquot.c
1128
remove_dquot_ref(sb, type);
fs/quota/dquot.c
1344
struct super_block *sb = dquot->dq_sb;
fs/quota/dquot.c
1348
if (!sb_has_quota_limits_enabled(sb, dquot->dq_id.type) ||
fs/quota/dquot.c
136
void __quota_error(struct super_block *sb, const char *func,
fs/quota/dquot.c
1381
sb_dqopt(sb)->info[dquot->dq_id.type].dqi_bgrace;
fs/quota/dquot.c
1445
struct super_block *sb = inode->i_sb;
fs/quota/dquot.c
1449
return sb_any_quota_loaded(sb) & ~sb_any_quota_suspended(sb);
fs/quota/dquot.c
1463
struct super_block *sb = inode->i_sb;
fs/quota/dquot.c
1489
if (!sb_has_quota_active(sb, cnt))
fs/quota/dquot.c
149
sb->s_id, func, &vaf);
fs/quota/dquot.c
1508
dquot = dqget(sb, qid);
fs/quota/dquot.c
1531
if (!sb_has_quota_active(sb, cnt))
fs/quota/dquot.c
2133
struct super_block *sb = inode->i_sb;
fs/quota/dquot.c
2143
dquot = dqget(sb, make_kqid_uid(kuid));
fs/quota/dquot.c
2157
dquot = dqget(sb, make_kqid_gid(kgid));
fs/quota/dquot.c
2177
int dquot_commit_info(struct super_block *sb, int type)
fs/quota/dquot.c
2179
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2181
return dqopt->ops[type]->write_file_info(sb, type);
fs/quota/dquot.c
2185
int dquot_get_next_id(struct super_block *sb, struct kqid *qid)
fs/quota/dquot.c
2187
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2189
if (!sb_has_quota_active(sb, qid->type))
fs/quota/dquot.c
2193
return dqopt->ops[qid->type]->get_next_id(sb, qid);
fs/quota/dquot.c
2226
static void vfs_cleanup_quota_inode(struct super_block *sb, int type)
fs/quota/dquot.c
2228
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2245
int dquot_disable(struct super_block *sb, int type, unsigned int flags)
fs/quota/dquot.c
2248
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2250
rwsem_assert_held_write(&sb->s_umount);
fs/quota/dquot.c
2264
if (!sb_any_quota_loaded(sb))
fs/quota/dquot.c
2270
if (!sb_has_quota_loaded(sb, cnt))
fs/quota/dquot.c
2282
if (!sb_has_quota_loaded(sb, cnt) &&
fs/quota/dquot.c
2283
sb_has_quota_suspended(sb, cnt)) {
fs/quota/dquot.c
2287
vfs_cleanup_quota_inode(sb, cnt);
fs/quota/dquot.c
2294
if (sb_has_quota_loaded(sb, cnt) && !(flags & DQUOT_SUSPENDED))
fs/quota/dquot.c
2298
drop_dquot_ref(sb, cnt);
fs/quota/dquot.c
2299
invalidate_dquots(sb, cnt);
fs/quota/dquot.c
2305
sb->dq_op->write_info(sb, cnt);
fs/quota/dquot.c
2307
dqopt->ops[cnt]->free_file_info(sb, cnt);
fs/quota/dquot.c
2321
if (sb->s_op->sync_fs)
fs/quota/dquot.c
2322
sb->s_op->sync_fs(sb, 1);
fs/quota/dquot.c
2323
sync_blockdev(sb->s_bdev);
fs/quota/dquot.c
2330
if (!sb_has_quota_loaded(sb, cnt) && dqopt->files[cnt]) {
fs/quota/dquot.c
2335
if (sb->s_bdev)
fs/quota/dquot.c
2336
invalidate_bdev(sb->s_bdev);
fs/quota/dquot.c
2343
if (!sb_has_quota_loaded(sb, cnt))
fs/quota/dquot.c
2344
vfs_cleanup_quota_inode(sb, cnt);
fs/quota/dquot.c
2349
int dquot_quota_off(struct super_block *sb, int type)
fs/quota/dquot.c
2351
return dquot_disable(sb, type,
fs/quota/dquot.c
2362
struct super_block *sb = inode->i_sb;
fs/quota/dquot.c
2363
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2371
if (sb_has_quota_loaded(sb, type))
fs/quota/dquot.c
2407
int dquot_load_quota_sb(struct super_block *sb, int type, int format_id,
fs/quota/dquot.c
2411
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2414
lockdep_assert_held_write(&sb->s_umount);
fs/quota/dquot.c
2423
if (!sb->dq_op || !sb->s_qcop ||
fs/quota/dquot.c
2424
(type == PRJQUOTA && sb->dq_op->get_projid == NULL)) {
fs/quota/dquot.c
2429
if (sb->s_user_ns != &init_user_ns) {
fs/quota/dquot.c
2438
if (sb_has_quota_loaded(sb, type)) {
fs/quota/dquot.c
2450
sync_filesystem(sb);
fs/quota/dquot.c
2451
invalidate_bdev(sb->s_bdev);
fs/quota/dquot.c
2455
if (!fmt->qf_ops->check_quota_file(sb, type))
fs/quota/dquot.c
2462
error = dqopt->ops[type]->read_file_info(sb, type);
fs/quota/dquot.c
2474
error = add_dquot_ref(sb, type);
fs/quota/dquot.c
2476
dquot_disable(sb, type,
fs/quota/dquot.c
2507
int dquot_resume(struct super_block *sb, int type)
fs/quota/dquot.c
2509
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2513
rwsem_assert_held_write(&sb->s_umount);
fs/quota/dquot.c
2518
if (!sb_has_quota_suspended(sb, cnt))
fs/quota/dquot.c
2529
ret = dquot_load_quota_sb(sb, cnt, dqopt->info[cnt].dqi_fmt_id,
fs/quota/dquot.c
2532
vfs_cleanup_quota_inode(sb, cnt);
fs/quota/dquot.c
2539
int dquot_quota_on(struct super_block *sb, int type, int format_id,
fs/quota/dquot.c
2546
if (path->dentry->d_sb != sb)
fs/quota/dquot.c
2560
int dquot_quota_on_mount(struct super_block *sb, char *qf_name,
fs/quota/dquot.c
2566
dentry = lookup_noperm_positive_unlocked(&QSTR(qf_name), sb->s_root);
fs/quota/dquot.c
2580
static int dquot_quota_enable(struct super_block *sb, unsigned int flags)
fs/quota/dquot.c
2584
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2596
if (!sb_has_quota_usage_enabled(sb, type)) {
fs/quota/dquot.c
2600
if (sb_has_quota_limits_enabled(sb, type)) {
fs/quota/dquot.c
2614
dquot_disable(sb, type, DQUOT_LIMITS_ENABLED);
fs/quota/dquot.c
2619
static int dquot_quota_disable(struct super_block *sb, unsigned int flags)
fs/quota/dquot.c
2623
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2638
if (!sb_has_quota_limits_enabled(sb, type))
fs/quota/dquot.c
2645
ret = dquot_disable(sb, type, DQUOT_LIMITS_ENABLED);
fs/quota/dquot.c
2682
int dquot_get_dqblk(struct super_block *sb, struct kqid qid,
fs/quota/dquot.c
2687
dquot = dqget(sb, qid);
fs/quota/dquot.c
2697
int dquot_get_next_dqblk(struct super_block *sb, struct kqid *qid,
fs/quota/dquot.c
2703
if (!sb->dq_op->get_next_id)
fs/quota/dquot.c
2705
err = sb->dq_op->get_next_id(sb, qid);
fs/quota/dquot.c
2708
dquot = dqget(sb, *qid);
fs/quota/dquot.c
277
hashfn(const struct super_block *sb, struct kqid qid)
fs/quota/dquot.c
2817
int dquot_set_dqblk(struct super_block *sb, struct kqid qid,
fs/quota/dquot.c
2823
dquot = dqget(sb, qid);
fs/quota/dquot.c
283
tmp = (((unsigned long)sb>>L1_CACHE_SHIFT) ^ id) * (MAXQUOTAS - type);
fs/quota/dquot.c
2836
int dquot_get_state(struct super_block *sb, struct qc_state *state)
fs/quota/dquot.c
2840
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
2845
if (!sb_has_quota_active(sb, type))
fs/quota/dquot.c
2848
mi = sb_dqopt(sb)->info + type;
fs/quota/dquot.c
2855
if (sb_has_quota_limits_enabled(sb, type))
fs/quota/dquot.c
2871
int dquot_set_dqinfo(struct super_block *sb, int type, struct qc_info *ii)
fs/quota/dquot.c
2878
if (!sb_has_quota_active(sb, type))
fs/quota/dquot.c
2880
mi = sb_dqopt(sb)->info + type;
fs/quota/dquot.c
2898
mark_info_dirty(sb, type);
fs/quota/dquot.c
2900
return sb->dq_op->write_info(sb, type);
fs/quota/dquot.c
302
static struct dquot *find_dquot(unsigned int hashent, struct super_block *sb,
fs/quota/dquot.c
308
if (dquot->dq_sb == sb && qid_eq(dquot->dq_id, qid))
fs/quota/dquot.c
443
void mark_info_dirty(struct super_block *sb, int type)
fs/quota/dquot.c
446
sb_dqopt(sb)->info[type].dqi_flags |= DQF_INFO_DIRTY;
fs/quota/dquot.c
574
static void invalidate_dquots(struct super_block *sb, int type)
fs/quota/dquot.c
583
if (dquot->dq_sb != sb)
fs/quota/dquot.c
629
int dquot_scan_active(struct super_block *sb,
fs/quota/dquot.c
636
WARN_ON_ONCE(!rwsem_is_locked(&sb->s_umount));
fs/quota/dquot.c
642
if (dquot->dq_sb != sb)
fs/quota/dquot.c
684
int dquot_writeback_dquots(struct super_block *sb, int type)
fs/quota/dquot.c
688
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
692
WARN_ON_ONCE(!rwsem_is_locked(&sb->s_umount));
fs/quota/dquot.c
699
if (!sb_has_quota_active(sb, cnt))
fs/quota/dquot.c
732
if ((cnt == type || type == -1) && sb_has_quota_active(sb, cnt)
fs/quota/dquot.c
734
sb->dq_op->write_info(sb, cnt);
fs/quota/dquot.c
742
int dquot_quota_sync(struct super_block *sb, int type)
fs/quota/dquot.c
744
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/dquot.c
748
ret = dquot_writeback_dquots(sb, type);
fs/quota/dquot.c
757
if (sb->s_op->sync_fs) {
fs/quota/dquot.c
758
ret = sb->s_op->sync_fs(sb, 1);
fs/quota/dquot.c
762
ret = sync_blockdev(sb->s_bdev);
fs/quota/dquot.c
773
if (!sb_has_quota_active(sb, cnt))
fs/quota/dquot.c
891
struct dquot *dquot_alloc(struct super_block *sb, int type)
fs/quota/dquot.c
897
static struct dquot *get_empty_dquot(struct super_block *sb, int type)
fs/quota/dquot.c
901
dquot = sb->dq_op->alloc_dquot(sb, type);
fs/quota/dquot.c
910
dquot->dq_sb = sb;
fs/quota/dquot.c
926
struct dquot *dqget(struct super_block *sb, struct kqid qid)
fs/quota/dquot.c
928
unsigned int hashent = hashfn(sb, qid);
fs/quota/dquot.c
931
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/dquot.c
934
if (!sb_has_quota_active(sb, qid.type))
fs/quota/dquot.c
939
if (!sb_has_quota_active(sb, qid.type)) {
fs/quota/dquot.c
947
dquot = find_dquot(hashent, sb, qid);
fs/quota/dquot.c
951
empty = get_empty_dquot(sb, qid.type);
fs/quota/dquot.c
980
err = sb->dq_op->acquire_dquot(dquot);
fs/quota/quota.c
100
if (!sb->s_qcop->quota_off && !sb->s_qcop->quota_disable)
fs/quota/quota.c
1000
ret = do_quotactl(sb, type, cmds, id, addr, ERR_PTR(-EINVAL));
fs/quota/quota.c
1003
up_write(&sb->s_umount);
fs/quota/quota.c
1005
up_read(&sb->s_umount);
fs/quota/quota.c
102
if (sb->s_qcop->quota_disable)
fs/quota/quota.c
103
return sb->s_qcop->quota_disable(sb, qtype_enforce_flag(type));
fs/quota/quota.c
104
return sb->s_qcop->quota_off(sb, type);
fs/quota/quota.c
107
static int quota_getfmt(struct super_block *sb, int type, void __user *addr)
fs/quota/quota.c
111
if (!sb_has_quota_active(sb, type))
fs/quota/quota.c
113
fmt = sb_dqopt(sb)->info[type].dqi_format->qf_fmt_id;
fs/quota/quota.c
119
static int quota_getinfo(struct super_block *sb, int type, void __user *addr)
fs/quota/quota.c
126
if (!sb->s_qcop->get_state)
fs/quota/quota.c
128
ret = sb->s_qcop->get_state(sb, &state);
fs/quota/quota.c
147
static int quota_setinfo(struct super_block *sb, int type, void __user *addr)
fs/quota/quota.c
154
if (!sb->s_qcop->set_info)
fs/quota/quota.c
174
return sb->s_qcop->set_info(sb, type, &qinfo);
fs/quota/quota.c
201
static int quota_getquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
209
if (!sb->s_qcop->get_dqblk)
fs/quota/quota.c
212
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
214
ret = sb->s_qcop->get_dqblk(sb, qid, &fdq);
fs/quota/quota.c
237
static int quota_getnextquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
245
if (!sb->s_qcop->get_nextdqblk)
fs/quota/quota.c
248
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
250
ret = sb->s_qcop->get_nextdqblk(sb, &qid, &fdq);
fs/quota/quota.c
27
static int check_quotactl_permission(struct super_block *sb, int type, int cmd,
fs/quota/quota.c
287
static int quota_setquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
304
if (!sb->s_qcop->set_dqblk)
fs/quota/quota.c
307
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
310
return sb->s_qcop->set_dqblk(sb, qid, &fdq);
fs/quota/quota.c
313
static int quota_enable(struct super_block *sb, void __user *addr)
fs/quota/quota.c
319
if (!sb->s_qcop->quota_enable)
fs/quota/quota.c
321
return sb->s_qcop->quota_enable(sb, flags);
fs/quota/quota.c
324
static int quota_disable(struct super_block *sb, void __user *addr)
fs/quota/quota.c
330
if (!sb->s_qcop->quota_disable)
fs/quota/quota.c
332
return sb->s_qcop->quota_disable(sb, flags);
fs/quota/quota.c
354
static int quota_getstate(struct super_block *sb, int type,
fs/quota/quota.c
361
ret = sb->s_qcop->get_state(sb, &state);
fs/quota/quota.c
434
static int quota_getxstate(struct super_block *sb, int type, void __user *addr)
fs/quota/quota.c
439
if (!sb->s_qcop->get_state)
fs/quota/quota.c
441
ret = quota_getstate(sb, type, &fqs);
fs/quota/quota.c
452
static int quota_getstatev(struct super_block *sb, int type,
fs/quota/quota.c
459
ret = sb->s_qcop->get_state(sb, &state);
fs/quota/quota.c
497
static int quota_getxstatev(struct super_block *sb, int type, void __user *addr)
fs/quota/quota.c
502
if (!sb->s_qcop->get_state)
fs/quota/quota.c
51
return security_quotactl(cmd, type, id, sb);
fs/quota/quota.c
516
ret = quota_getstatev(sb, type, &fqs);
fs/quota/quota.c
54
static void quota_sync_one(struct super_block *sb, void *arg)
fs/quota/quota.c
58
if (sb->s_qcop && sb->s_qcop->quota_sync &&
fs/quota/quota.c
59
(sb->s_quota_types & (1 << type)))
fs/quota/quota.c
60
sb->s_qcop->quota_sync(sb, type);
fs/quota/quota.c
624
static int quota_setxquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
633
if (!sb->s_qcop->set_dqblk)
fs/quota/quota.c
636
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
639
if (from_kqid(sb->s_user_ns, qid) == 0 &&
fs/quota/quota.c
644
if (!sb->s_qcop->set_info)
fs/quota/quota.c
647
ret = sb->s_qcop->set_info(sb, type, &qinfo);
fs/quota/quota.c
654
return sb->s_qcop->set_dqblk(sb, qid, &qdq);
fs/quota/quota.c
705
static int quota_getxquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
713
if (!sb->s_qcop->get_dqblk)
fs/quota/quota.c
716
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
718
ret = sb->s_qcop->get_dqblk(sb, qid, &qdq);
fs/quota/quota.c
731
static int quota_getnextxquota(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
740
if (!sb->s_qcop->get_nextdqblk)
fs/quota/quota.c
743
if (!qid_has_mapping(sb->s_user_ns, qid))
fs/quota/quota.c
745
ret = sb->s_qcop->get_nextdqblk(sb, &qid, &qdq);
fs/quota/quota.c
755
static int quota_rmxquota(struct super_block *sb, void __user *addr)
fs/quota/quota.c
761
if (!sb->s_qcop->rm_xquota)
fs/quota/quota.c
763
return sb->s_qcop->rm_xquota(sb, flags);
fs/quota/quota.c
767
static int do_quotactl(struct super_block *sb, int type, int cmd, qid_t id,
fs/quota/quota.c
777
if (!sb->s_qcop)
fs/quota/quota.c
779
if (!(sb->s_quota_types & (1 << type)))
fs/quota/quota.c
782
ret = check_quotactl_permission(sb, type, cmd, id);
fs/quota/quota.c
788
return quota_quotaon(sb, type, id, path);
fs/quota/quota.c
790
return quota_quotaoff(sb, type);
fs/quota/quota.c
792
return quota_getfmt(sb, type, addr);
fs/quota/quota.c
794
return quota_getinfo(sb, type, addr);
fs/quota/quota.c
796
return quota_setinfo(sb, type, addr);
fs/quota/quota.c
798
return quota_getquota(sb, type, id, addr);
fs/quota/quota.c
800
return quota_getnextquota(sb, type, id, addr);
fs/quota/quota.c
802
return quota_setquota(sb, type, id, addr);
fs/quota/quota.c
804
if (!sb->s_qcop->quota_sync)
fs/quota/quota.c
806
return sb->s_qcop->quota_sync(sb, type);
fs/quota/quota.c
808
return quota_enable(sb, addr);
fs/quota/quota.c
810
return quota_disable(sb, addr);
fs/quota/quota.c
812
return quota_rmxquota(sb, addr);
fs/quota/quota.c
814
return quota_getxstate(sb, type, addr);
fs/quota/quota.c
816
return quota_getxstatev(sb, type, addr);
fs/quota/quota.c
818
return quota_setxquota(sb, type, id, addr);
fs/quota/quota.c
820
return quota_getxquota(sb, type, id, addr);
fs/quota/quota.c
822
return quota_getnextxquota(sb, type, id, addr);
fs/quota/quota.c
824
if (sb_rdonly(sb))
fs/quota/quota.c
86
static int quota_quotaon(struct super_block *sb, int type, qid_t id,
fs/quota/quota.c
869
struct super_block *sb;
fs/quota/quota.c
889
sb = user_get_super(dev, excl);
fs/quota/quota.c
89
if (!sb->s_qcop->quota_on && !sb->s_qcop->quota_enable)
fs/quota/quota.c
890
if (!sb)
fs/quota/quota.c
892
if (thawed && sb->s_writers.frozen != SB_UNFROZEN) {
fs/quota/quota.c
894
up_write(&sb->s_umount);
fs/quota/quota.c
896
up_read(&sb->s_umount);
fs/quota/quota.c
898
sb_start_write(sb);
fs/quota/quota.c
899
sb_end_write(sb);
fs/quota/quota.c
900
put_super(sb);
fs/quota/quota.c
904
return sb;
fs/quota/quota.c
91
if (sb->s_qcop->quota_enable)
fs/quota/quota.c
92
return sb->s_qcop->quota_enable(sb, qtype_enforce_flag(type));
fs/quota/quota.c
921
struct super_block *sb = NULL;
fs/quota/quota.c
95
return sb->s_qcop->quota_on(sb, type, id, path);
fs/quota/quota.c
955
sb = quotactl_block(special, cmds);
fs/quota/quota.c
956
if (IS_ERR(sb)) {
fs/quota/quota.c
957
ret = PTR_ERR(sb);
fs/quota/quota.c
961
ret = do_quotactl(sb, type, cmds, id, addr, pathp);
fs/quota/quota.c
964
drop_super(sb);
fs/quota/quota.c
966
drop_super_exclusive(sb);
fs/quota/quota.c
976
struct super_block *sb;
fs/quota/quota.c
98
static int quota_quotaoff(struct super_block *sb, int type)
fs/quota/quota.c
994
sb = fd_file(f)->f_path.mnt->mnt_sb;
fs/quota/quota.c
996
down_write(&sb->s_umount);
fs/quota/quota.c
998
down_read(&sb->s_umount);
fs/quota/quota_tree.c
442
struct super_block *sb = dquot->dq_sb;
fs/quota/quota_tree.c
453
quota_error(sb, "Error %zd occurred while creating "
fs/quota/quota_tree.c
462
ret = sb->s_op->quota_write(sb, type, ddquot, info->dqi_entry_size,
fs/quota/quota_tree.c
465
quota_error(sb, "dquota write failed");
fs/quota/quota_tree.c
58
struct super_block *sb = info->dqi_sb;
fs/quota/quota_tree.c
61
return sb->s_op->quota_read(sb, info->dqi_type, buf,
fs/quota/quota_tree.c
67
struct super_block *sb = info->dqi_sb;
fs/quota/quota_tree.c
70
ret = sb->s_op->quota_write(sb, info->dqi_type, buf,
fs/quota/quota_tree.c
722
struct super_block *sb = dquot->dq_sb;
fs/quota/quota_tree.c
73
quota_error(sb, "dquota write failed");
fs/quota/quota_tree.c
730
quota_error(sb, "Quota invalidated while reading!");
fs/quota/quota_tree.c
739
quota_error(sb,"Can't read quota structure "
fs/quota/quota_tree.c
754
ret = sb->s_op->quota_read(sb, type, ddquot, info->dqi_entry_size,
fs/quota/quota_tree.c
759
quota_error(sb, "Error while reading quota structure for id %u",
fs/quota/quota_tree.c
80
static inline int do_check_range(struct super_block *sb, const char *val_name,
fs/quota/quota_tree.c
84
quota_error(sb, "Getting %s %u out of range %u-%u",
fs/quota/quota_v1.c
127
static int v1_check_quota_file(struct super_block *sb, int type)
fs/quota/quota_v1.c
129
struct inode *inode = sb_dqopt(sb)->files[type];
fs/quota/quota_v1.c
147
size = sb->s_op->quota_read(sb, type, (char *)&dqhead,
fs/quota/quota_v1.c
155
" It probably contains newer quota format.\n", sb->s_id);
fs/quota/quota_v1.c
159
static int v1_read_file_info(struct super_block *sb, int type)
fs/quota/quota_v1.c
161
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/quota_v1.c
168
ret = sb->s_op->quota_read(sb, type, (char *)&dqblk,
fs/quota/quota_v1.c
189
static int v1_write_file_info(struct super_block *sb, int type)
fs/quota/quota_v1.c
191
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/quota_v1.c
198
ret = sb->s_op->quota_read(sb, type, (char *)&dqblk,
fs/quota/quota_v1.c
210
ret = sb->s_op->quota_write(sb, type, (char *)&dqblk,
fs/quota/quota_v2.c
104
ret = v2_read_header(sb, type, &dqhead);
fs/quota/quota_v2.c
114
size = sb->s_op->quota_read(sb, type, (char *)&dinfo,
fs/quota/quota_v2.c
117
quota_error(sb, "Can't read info structure");
fs/quota/quota_v2.c
147
qinfo->dqi_sb = sb;
fs/quota/quota_v2.c
165
i_size_read(sb_dqopt(sb)->files[type])) {
fs/quota/quota_v2.c
166
quota_error(sb, "Number of blocks too big for quota file size (%llu > %llu).",
fs/quota/quota_v2.c
168
i_size_read(sb_dqopt(sb)->files[type]));
fs/quota/quota_v2.c
173
quota_error(sb, "Free block number %u out of range (%u, %u).",
fs/quota/quota_v2.c
179
quota_error(sb, "Block with free entry %u out of range (%u, %u).",
fs/quota/quota_v2.c
197
static int v2_write_file_info(struct super_block *sb, int type)
fs/quota/quota_v2.c
200
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/quota_v2.c
218
size = sb->s_op->quota_write(sb, type, (char *)&dinfo,
fs/quota/quota_v2.c
223
quota_error(sb, "Can't write info structure");
fs/quota/quota_v2.c
398
static int v2_free_file_info(struct super_block *sb, int type)
fs/quota/quota_v2.c
400
kfree(sb_dqinfo(sb, type)->dqi_priv);
fs/quota/quota_v2.c
404
static int v2_get_next_id(struct super_block *sb, struct kqid *qid)
fs/quota/quota_v2.c
406
struct quota_info *dqopt = sb_dqopt(sb);
fs/quota/quota_v2.c
412
ret = qtree_get_next_id(sb_dqinfo(sb, qid->type)->dqi_priv, qid);
fs/quota/quota_v2.c
57
static int v2_read_header(struct super_block *sb, int type,
fs/quota/quota_v2.c
62
size = sb->s_op->quota_read(sb, type, (char *)dqhead,
fs/quota/quota_v2.c
65
quota_error(sb, "Failed header read: expected=%zd got=%zd",
fs/quota/quota_v2.c
75
static int v2_check_quota_file(struct super_block *sb, int type)
fs/quota/quota_v2.c
81
if (v2_read_header(sb, type, &dqhead))
fs/quota/quota_v2.c
90
static int v2_read_file_info(struct super_block *sb, int type)
fs/quota/quota_v2.c
94
struct quota_info *dqopt = sb_dqopt(sb);
fs/ramfs/inode.c
260
static int ramfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ramfs/inode.c
262
struct ramfs_fs_info *fsi = sb->s_fs_info;
fs/ramfs/inode.c
265
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/ramfs/inode.c
266
sb->s_blocksize = PAGE_SIZE;
fs/ramfs/inode.c
267
sb->s_blocksize_bits = PAGE_SHIFT;
fs/ramfs/inode.c
268
sb->s_magic = RAMFS_MAGIC;
fs/ramfs/inode.c
269
sb->s_op = &ramfs_ops;
fs/ramfs/inode.c
270
sb->s_d_flags = DCACHE_DONTCACHE;
fs/ramfs/inode.c
271
sb->s_time_gran = 1;
fs/ramfs/inode.c
273
inode = ramfs_get_inode(sb, NULL, S_IFDIR | fsi->mount_opts.mode, 0);
fs/ramfs/inode.c
274
sb->s_root = d_make_root(inode);
fs/ramfs/inode.c
275
if (!sb->s_root)
fs/ramfs/inode.c
311
void ramfs_kill_sb(struct super_block *sb)
fs/ramfs/inode.c
313
kfree(sb->s_fs_info);
fs/ramfs/inode.c
314
kill_anon_super(sb);
fs/ramfs/inode.c
57
struct inode *ramfs_get_inode(struct super_block *sb,
fs/ramfs/inode.c
60
struct inode * inode = new_inode(sb);
fs/resctrl/rdtgroup.c
3168
static void rdt_kill_sb(struct super_block *sb)
fs/resctrl/rdtgroup.c
3187
kernfs_kill_sb(sb);
fs/romfs/internal.h
16
static inline size_t romfs_maxsize(struct super_block *sb)
fs/romfs/internal.h
18
return (size_t) (unsigned long) sb->s_fs_info;
fs/romfs/internal.h
38
extern int romfs_dev_read(struct super_block *sb, unsigned long pos,
fs/romfs/internal.h
40
extern ssize_t romfs_dev_strnlen(struct super_block *sb,
fs/romfs/internal.h
42
extern int romfs_dev_strcmp(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
103
static int romfs_blk_read(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
114
bh = sb_bread(sb, pos >> ROMBSBITS);
fs/romfs/storage.c
130
static ssize_t romfs_blk_strnlen(struct super_block *sb,
fs/romfs/storage.c
143
bh = sb_bread(sb, pos >> ROMBSBITS);
fs/romfs/storage.c
163
static int romfs_blk_strcmp(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
175
bh = sb_bread(sb, pos >> ROMBSBITS);
fs/romfs/storage.c
18
#define ROMFS_MTD_READ(sb, ...) mtd_read((sb)->s_mtd, ##__VA_ARGS__)
fs/romfs/storage.c
198
bh = sb_bread(sb, pos >> ROMBSBITS);
fs/romfs/storage.c
214
int romfs_dev_read(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
219
limit = romfs_maxsize(sb);
fs/romfs/storage.c
224
if (sb->s_mtd)
fs/romfs/storage.c
225
return romfs_mtd_read(sb, pos, buf, buflen);
fs/romfs/storage.c
228
if (sb->s_bdev)
fs/romfs/storage.c
229
return romfs_blk_read(sb, pos, buf, buflen);
fs/romfs/storage.c
23
static int romfs_mtd_read(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
237
ssize_t romfs_dev_strnlen(struct super_block *sb,
fs/romfs/storage.c
242
limit = romfs_maxsize(sb);
fs/romfs/storage.c
249
if (sb->s_mtd)
fs/romfs/storage.c
250
return romfs_mtd_strnlen(sb, pos, maxlen);
fs/romfs/storage.c
253
if (sb->s_bdev)
fs/romfs/storage.c
254
return romfs_blk_strnlen(sb, pos, maxlen);
fs/romfs/storage.c
265
int romfs_dev_strcmp(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
270
limit = romfs_maxsize(sb);
fs/romfs/storage.c
279
if (sb->s_mtd)
fs/romfs/storage.c
280
return romfs_mtd_strcmp(sb, pos, str, size);
fs/romfs/storage.c
283
if (sb->s_bdev)
fs/romfs/storage.c
284
return romfs_blk_strcmp(sb, pos, str, size);
fs/romfs/storage.c
29
ret = ROMFS_MTD_READ(sb, pos, buflen, &rlen, buf);
fs/romfs/storage.c
36
static ssize_t romfs_mtd_strnlen(struct super_block *sb,
fs/romfs/storage.c
48
ret = ROMFS_MTD_READ(sb, pos, segment, &len, buf);
fs/romfs/storage.c
66
static int romfs_mtd_strcmp(struct super_block *sb, unsigned long pos,
fs/romfs/storage.c
79
ret = ROMFS_MTD_READ(sb, pos, segment, &len, buf);
fs/romfs/super.c
269
static struct inode *romfs_iget(struct super_block *sb, unsigned long pos)
fs/romfs/super.c
282
ret = romfs_dev_read(sb, pos, &ri, sizeof(ri));
fs/romfs/super.c
296
nlen = romfs_dev_strnlen(sb, pos + ROMFH_SIZE, ROMFS_MAXFN);
fs/romfs/super.c
301
i = iget_locked(sb, pos);
fs/romfs/super.c
365
static struct inode *romfs_alloc_inode(struct super_block *sb)
fs/romfs/super.c
369
inode = alloc_inode_sb(sb, romfs_inode_cachep, GFP_KERNEL);
fs/romfs/super.c
386
struct super_block *sb = dentry->d_sb;
fs/romfs/super.c
399
if (sb->s_bdev)
fs/romfs/super.c
400
id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/romfs/super.c
401
else if (sb->s_dev)
fs/romfs/super.c
402
id = huge_encode_dev(sb->s_dev);
fs/romfs/super.c
450
static int romfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/romfs/super.c
460
if (!sb->s_mtd) {
fs/romfs/super.c
461
if (!sb_set_blocksize(sb, ROMBSIZE)) {
fs/romfs/super.c
466
sb->s_blocksize = ROMBSIZE;
fs/romfs/super.c
467
sb->s_blocksize_bits = blksize_bits(ROMBSIZE);
fs/romfs/super.c
471
sb->s_maxbytes = 0xFFFFFFFF;
fs/romfs/super.c
472
sb->s_magic = ROMFS_MAGIC;
fs/romfs/super.c
473
sb->s_flags |= SB_RDONLY | SB_NOATIME;
fs/romfs/super.c
474
sb->s_time_min = 0;
fs/romfs/super.c
475
sb->s_time_max = 0;
fs/romfs/super.c
476
sb->s_op = &romfs_super_ops;
fs/romfs/super.c
480
if (sb->s_mtd)
fs/romfs/super.c
481
sb->s_dev = MKDEV(MTD_BLOCK_MAJOR, sb->s_mtd->index);
fs/romfs/super.c
488
sb->s_fs_info = (void *) 512;
fs/romfs/super.c
489
ret = romfs_dev_read(sb, 0, rsb, 512);
fs/romfs/super.c
495
if (sb->s_mtd && img_size > sb->s_mtd->size)
fs/romfs/super.c
498
sb->s_fs_info = (void *) img_size;
fs/romfs/super.c
504
sb->s_id);
fs/romfs/super.c
509
pr_err("bad initial checksum on dev %s.\n", sb->s_id);
fs/romfs/super.c
513
storage = sb->s_mtd ? "MTD" : "the block layer";
fs/romfs/super.c
526
root = romfs_iget(sb, pos);
fs/romfs/super.c
530
sb->s_root = d_make_root(root);
fs/romfs/super.c
531
if (!sb->s_root)
fs/romfs/super.c
577
static void romfs_kill_sb(struct super_block *sb)
fs/romfs/super.c
579
generic_shutdown_super(sb);
fs/romfs/super.c
582
if (sb->s_mtd) {
fs/romfs/super.c
583
put_mtd_device(sb->s_mtd);
fs/romfs/super.c
584
sb->s_mtd = NULL;
fs/romfs/super.c
588
if (sb->s_bdev) {
fs/romfs/super.c
589
sync_blockdev(sb->s_bdev);
fs/romfs/super.c
590
bdev_fput(sb->s_bdev_file);
fs/romfs/super.c
97
static struct inode *romfs_iget(struct super_block *sb, unsigned long pos);
fs/smb/client/cifsacl.c
1714
struct super_block *sb = dentry->d_sb;
fs/smb/client/cifsacl.c
1715
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsacl.c
1738
if (sb->s_flags & SB_POSIXACL)
fs/smb/client/cifsacl.c
1746
if (sb->s_flags & SB_POSIXACL)
fs/smb/client/cifsacl.c
1777
struct super_block *sb = dentry->d_sb;
fs/smb/client/cifsacl.c
1778
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsacl.c
1815
if (sb->s_flags & SB_POSIXACL)
fs/smb/client/cifsacl.c
1823
if (sb->s_flags & SB_POSIXACL)
fs/smb/client/cifsfs.c
1003
static int cifs_set_super(struct super_block *sb, void *data)
fs/smb/client/cifsfs.c
1006
sb->s_fs_info = mnt_data->cifs_sb;
fs/smb/client/cifsfs.c
1007
return set_anon_super(sb, NULL);
fs/smb/client/cifsfs.c
1016
struct super_block *sb;
fs/smb/client/cifsfs.c
1063
sb = sget(fs_type, cifs_match_super, cifs_set_super, flags, &mnt_data);
fs/smb/client/cifsfs.c
1064
if (IS_ERR(sb)) {
fs/smb/client/cifsfs.c
1066
return ERR_CAST(sb);
fs/smb/client/cifsfs.c
1069
if (sb->s_root) {
fs/smb/client/cifsfs.c
1074
rc = cifs_read_super(sb);
fs/smb/client/cifsfs.c
1080
sb->s_flags |= SB_ACTIVE;
fs/smb/client/cifsfs.c
1083
root = cifs_get_root(cifs_sb ? cifs_sb->ctx : old_ctx, sb);
fs/smb/client/cifsfs.c
1094
deactivate_locked_super(sb);
fs/smb/client/cifsfs.c
209
cifs_sb_active(struct super_block *sb)
fs/smb/client/cifsfs.c
211
struct cifs_sb_info *server = CIFS_SB(sb);
fs/smb/client/cifsfs.c
214
atomic_inc(&sb->s_active);
fs/smb/client/cifsfs.c
218
cifs_sb_deactive(struct super_block *sb)
fs/smb/client/cifsfs.c
220
struct cifs_sb_info *server = CIFS_SB(sb);
fs/smb/client/cifsfs.c
223
deactivate_super(sb);
fs/smb/client/cifsfs.c
227
cifs_read_super(struct super_block *sb)
fs/smb/client/cifsfs.c
236
cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
241
sb->s_flags |= SB_POSIXACL;
fs/smb/client/cifsfs.c
244
sb->s_flags |= SB_RDONLY;
fs/smb/client/cifsfs.c
247
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/smb/client/cifsfs.c
249
sb->s_maxbytes = MAX_NON_LFS;
fs/smb/client/cifsfs.c
262
sb->s_time_gran = 1000000000; /* 1 second is max allowed gran */
fs/smb/client/cifsfs.c
264
sb->s_time_min = ts.tv_sec;
fs/smb/client/cifsfs.c
267
sb->s_time_max = ts.tv_sec;
fs/smb/client/cifsfs.c
273
sb->s_time_gran = 100;
fs/smb/client/cifsfs.c
275
sb->s_time_min = ts.tv_sec;
fs/smb/client/cifsfs.c
277
sb->s_time_max = ts.tv_sec;
fs/smb/client/cifsfs.c
280
sb->s_magic = CIFS_SUPER_MAGIC;
fs/smb/client/cifsfs.c
281
sb->s_op = &cifs_super_ops;
fs/smb/client/cifsfs.c
282
sb->s_xattr = cifs_xattr_handlers;
fs/smb/client/cifsfs.c
283
rc = super_setup_bdi(sb);
fs/smb/client/cifsfs.c
291
sb->s_bdi->ra_pages = cifs_sb->ctx->rasize / PAGE_SIZE;
fs/smb/client/cifsfs.c
293
sb->s_bdi->ra_pages = 2 * (cifs_sb->ctx->rsize / PAGE_SIZE);
fs/smb/client/cifsfs.c
295
sb->s_blocksize = CIFS_MAX_MSGSIZE;
fs/smb/client/cifsfs.c
296
sb->s_blocksize_bits = 14; /* default 2**14 = CIFS_MAX_MSGSIZE */
fs/smb/client/cifsfs.c
297
inode = cifs_root_iget(sb);
fs/smb/client/cifsfs.c
305
set_default_d_op(sb, &cifs_ci_dentry_ops);
fs/smb/client/cifsfs.c
307
set_default_d_op(sb, &cifs_dentry_ops);
fs/smb/client/cifsfs.c
309
sb->s_root = d_make_root(inode);
fs/smb/client/cifsfs.c
310
if (!sb->s_root) {
fs/smb/client/cifsfs.c
318
sb->s_export_op = &cifs_export_ops;
fs/smb/client/cifsfs.c
329
static void cifs_kill_sb(struct super_block *sb)
fs/smb/client/cifsfs.c
331
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
349
kill_anon_super(sb);
fs/smb/client/cifsfs.c
356
struct super_block *sb = dentry->d_sb;
fs/smb/client/cifsfs.c
357
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
453
cifs_alloc_inode(struct super_block *sb)
fs/smb/client/cifsfs.c
456
cifs_inode = alloc_inode_sb(sb, cifs_inode_cachep, GFP_KERNEL);
fs/smb/client/cifsfs.c
846
static void cifs_umount_begin(struct super_block *sb)
fs/smb/client/cifsfs.c
848
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
890
static int cifs_freeze(struct super_block *sb)
fs/smb/client/cifsfs.c
892
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
952
cifs_get_root(struct smb3_fs_context *ctx, struct super_block *sb)
fs/smb/client/cifsfs.c
955
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifsfs.c
961
return dget(sb->s_root);
fs/smb/client/cifsfs.c
971
dentry = dget(sb->s_root);
fs/smb/client/cifsfs.h
46
void cifs_sb_active(struct super_block *sb);
fs/smb/client/cifsfs.h
47
void cifs_sb_deactive(struct super_block *sb);
fs/smb/client/cifsfs.h
51
struct inode *cifs_root_iget(struct super_block *sb);
fs/smb/client/cifsglob.h
1589
static inline void *super_to_fsinfo(struct super_block *sb)
fs/smb/client/cifsglob.h
1591
return sb->s_fs_info;
fs/smb/client/cifsglob.h
443
struct cifs_sb_info *sb);
fs/smb/client/cifsglob.h
632
struct super_block *sb,
fs/smb/client/cifsproto.h
173
struct super_block *sb, int mode, unsigned int f_flags,
fs/smb/client/cifsproto.h
175
void cifs_fill_uniqueid(struct super_block *sb, struct cifs_fattr *fattr);
fs/smb/client/cifsproto.h
184
struct inode *cifs_iget(struct super_block *sb, struct cifs_fattr *fattr);
fs/smb/client/cifsproto.h
188
struct super_block *sb, int xid,
fs/smb/client/cifsproto.h
192
struct super_block *sb,
fs/smb/client/cifsproto.h
196
struct super_block *sb, unsigned int xid);
fs/smb/client/cifsproto.h
238
int cifs_match_super(struct super_block *sb, void *data);
fs/smb/client/cifsproto.h
390
void cifs_put_tcp_super(struct super_block *sb);
fs/smb/client/cifssmb.c
3085
struct super_block *sb,
fs/smb/client/cifssmb.c
3093
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/cifssmb.c
3191
rc = cifs_get_inode_info(&new, full_path, data, sb, xid, NULL);
fs/smb/client/connect.c
2917
compare_mount_options(struct super_block *sb, struct cifs_mnt_data *mnt_data)
fs/smb/client/connect.c
2919
struct cifs_sb_info *old = CIFS_SB(sb);
fs/smb/client/connect.c
2924
if ((sb->s_flags & CIFS_MS_MASK) != (mnt_data->flags & CIFS_MS_MASK))
fs/smb/client/connect.c
2970
static int match_prepath(struct super_block *sb,
fs/smb/client/connect.c
2975
struct cifs_sb_info *old = CIFS_SB(sb);
fs/smb/client/connect.c
2995
cifs_match_super(struct super_block *sb, void *data)
fs/smb/client/connect.c
3007
cifs_sb = CIFS_SB(sb);
fs/smb/client/connect.c
3035
!match_prepath(sb, tcon, mnt_data)) {
fs/smb/client/connect.c
3040
rc = compare_mount_options(sb, mnt_data);
fs/smb/client/dfs.c
393
struct super_block *sb = NULL;
fs/smb/client/dfs.c
433
sb = cifs_get_dfs_tcon_super(tcon);
fs/smb/client/dfs.c
434
if (!IS_ERR(sb))
fs/smb/client/dfs.c
435
cifs_sb = CIFS_SB(sb);
fs/smb/client/dfs.c
451
cifs_put_tcp_super(sb);
fs/smb/client/file.c
3157
struct super_block *sb = inode->i_sb;
fs/smb/client/file.c
3158
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/file.c
479
struct super_block *sb, int mode, unsigned int f_flags,
fs/smb/client/file.c
485
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/file.c
524
cifs_fill_uniqueid(sb, &fattr);
fs/smb/client/file.c
525
*pinode = cifs_iget(sb, &fattr);
fs/smb/client/file.c
859
struct super_block *sb = inode->i_sb;
fs/smb/client/file.c
860
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1063
struct super_block *sb,
fs/smb/client/inode.c
1068
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1100
static void cifs_set_fattr_ino(int xid, struct cifs_tcon *tcon, struct super_block *sb,
fs/smb/client/inode.c
1104
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1112
fattr->cf_uniqueid = iunique(sb, ROOT_I);
fs/smb/client/inode.c
1130
fattr->cf_uniqueid = iunique(sb, ROOT_I);
fs/smb/client/inode.c
1156
struct super_block *sb,
fs/smb/client/inode.c
1163
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1198
cifs_create_junction_fattr(fattr, sb);
fs/smb/client/inode.c
1223
cifs_create_junction_fattr(fattr, sb);
fs/smb/client/inode.c
1246
smb311_posix_info_to_fattr(fattr, data, sb);
fs/smb/client/inode.c
1248
cifs_open_info_to_fattr(fattr, data, sb);
fs/smb/client/inode.c
1256
struct super_block *sb, int xid,
fs/smb/client/inode.c
1262
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1300
rc = reparse_info_to_fattr(data, sb, xid, tcon,
fs/smb/client/inode.c
1303
cifs_open_info_to_fattr(fattr, data, sb);
fs/smb/client/inode.c
1311
cifs_create_junction_fattr(fattr, sb);
fs/smb/client/inode.c
1330
rc = cifs_backup_query_path_info(xid, tcon, sb,
fs/smb/client/inode.c
1362
cifs_set_fattr_ino(xid, tcon, sb, inode, full_path, data, fattr);
fs/smb/client/inode.c
1425
struct super_block *sb, int xid,
fs/smb/client/inode.c
1436
rc = cifs_get_fattr(data, sb, xid, fid, &fattr, inode, full_path);
fs/smb/client/inode.c
1440
rc = update_inode_info(sb, &fattr, inode);
fs/smb/client/inode.c
1449
struct super_block *sb,
fs/smb/client/inode.c
1454
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1482
rc = reparse_info_to_fattr(data, sb, xid, tcon,
fs/smb/client/inode.c
1485
smb311_posix_info_to_fattr(fattr, data, sb);
fs/smb/client/inode.c
1490
cifs_create_junction_fattr(fattr, sb);
fs/smb/client/inode.c
1524
struct super_block *sb,
fs/smb/client/inode.c
1535
rc = smb311_posix_get_fattr(data, &fattr, full_path, sb, xid);
fs/smb/client/inode.c
1539
rc = update_inode_info(sb, &fattr, inode);
fs/smb/client/inode.c
1610
cifs_iget(struct super_block *sb, struct cifs_fattr *fattr)
fs/smb/client/inode.c
1621
inode = iget5_locked(sb, hash, cifs_find_inode, cifs_init_inode, fattr);
fs/smb/client/inode.c
1628
cifs_autodisable_serverino(CIFS_SB(sb));
fs/smb/client/inode.c
1630
fattr->cf_uniqueid = iunique(sb, ROOT_I);
fs/smb/client/inode.c
1637
if (sb->s_flags & SB_NOATIME)
fs/smb/client/inode.c
1650
struct inode *cifs_root_iget(struct super_block *sb)
fs/smb/client/inode.c
1652
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
1677
rc = cifs_get_unix_fattr(path, sb, &fattr, &inode, xid);
fs/smb/client/inode.c
1687
rc = smb311_posix_get_fattr(NULL, &fattr, path, sb, xid);
fs/smb/client/inode.c
1689
rc = cifs_get_fattr(NULL, sb, xid, NULL, &fattr, &inode, path);
fs/smb/client/inode.c
1697
inode = cifs_iget(sb, &fattr);
fs/smb/client/inode.c
1940
struct super_block *sb = dir->i_sb;
fs/smb/client/inode.c
1941
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
242
cifs_fill_uniqueid(struct super_block *sb, struct cifs_fattr *fattr)
fs/smb/client/inode.c
244
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
247
fattr->cf_uniqueid = iunique(sb, ROOT_I);
fs/smb/client/inode.c
2824
struct super_block *sb = dentry->d_sb;
fs/smb/client/inode.c
2849
if (cifs_sb_master_tcon(CIFS_SB(sb))->posix_extensions) {
fs/smb/client/inode.c
2851
NULL, sb, xid);
fs/smb/client/inode.c
2852
} else if (cifs_sb_master_tcon(CIFS_SB(sb))->unix_ext) {
fs/smb/client/inode.c
2853
rc = cifs_get_inode_info_unix(&inode, full_path, sb, xid);
fs/smb/client/inode.c
2855
rc = cifs_get_inode_info(&inode, full_path, NULL, sb,
fs/smb/client/inode.c
347
struct super_block *sb)
fs/smb/client/inode.c
349
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
364
static int update_inode_info(struct super_block *sb,
fs/smb/client/inode.c
368
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
372
*inode = cifs_iget(sb, fattr);
fs/smb/client/inode.c
428
struct super_block *sb,
fs/smb/client/inode.c
434
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
457
cifs_create_junction_fattr(fattr, sb);
fs/smb/client/inode.c
464
cifs_fill_uniqueid(sb, fattr);
fs/smb/client/inode.c
485
struct super_block *sb, unsigned int xid)
fs/smb/client/inode.c
490
rc = cifs_get_unix_fattr(full_path, sb, &fattr, pinode, xid);
fs/smb/client/inode.c
494
rc = update_inode_info(sb, &fattr, pinode);
fs/smb/client/inode.c
501
struct super_block *sb,
fs/smb/client/inode.c
511
struct super_block *sb, unsigned int xid)
fs/smb/client/inode.c
830
struct super_block *sb)
fs/smb/client/inode.c
833
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/inode.c
888
struct super_block *sb)
fs/smb/client/inode.c
891
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/ioctl.c
166
static int cifs_shutdown(struct super_block *sb, unsigned long arg)
fs/smb/client/ioctl.c
168
struct cifs_sb_info *sbi = CIFS_SB(sb);
fs/smb/client/misc.c
888
struct super_block *sb;
fs/smb/client/misc.c
891
static void tcon_super_cb(struct super_block *sb, void *arg)
fs/smb/client/misc.c
897
if (sd->sb)
fs/smb/client/misc.c
900
cifs_sb = CIFS_SB(sb);
fs/smb/client/misc.c
909
sd->sb = sb;
fs/smb/client/misc.c
918
.sb = NULL,
fs/smb/client/misc.c
926
if (sd.sb) {
fs/smb/client/misc.c
932
cifs_sb_active(sd.sb);
fs/smb/client/misc.c
933
return sd.sb;
fs/smb/client/misc.c
940
static void __cifs_put_super(struct super_block *sb)
fs/smb/client/misc.c
942
if (!IS_ERR_OR_NULL(sb))
fs/smb/client/misc.c
943
cifs_sb_deactive(sb);
fs/smb/client/misc.c
957
void cifs_put_tcp_super(struct super_block *sb)
fs/smb/client/misc.c
959
__cifs_put_super(sb);
fs/smb/client/readdir.c
1029
fattr.cf_uniqueid = iunique(sb, ROOT_I);
fs/smb/client/readdir.c
163
inode = cifs_iget(sb, fattr);
fs/smb/client/readdir.c
72
struct super_block *sb = parent->d_sb;
fs/smb/client/readdir.c
73
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/readdir.c
956
struct super_block *sb = file_inode(file)->i_sb;
fs/smb/client/readdir.c
957
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/smb1proto.h
113
struct super_block *sb,
fs/smb/client/smb2inode.c
1434
struct super_block *sb,
fs/smb/client/smb2inode.c
1443
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/smb2inode.c
1483
data, sb, xid);
fs/smb/client/smb2inode.c
1492
data, sb, xid, NULL);
fs/smb/client/smb2proto.h
50
struct super_block *sb,
fs/smb/client/xattr.c
101
struct super_block *sb = dentry->d_sb;
fs/smb/client/xattr.c
102
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/client/xattr.c
277
struct super_block *sb = dentry->d_sb;
fs/smb/client/xattr.c
278
struct cifs_sb_info *cifs_sb = CIFS_SB(sb);
fs/smb/server/vfs_cache.c
159
static unsigned long inode_hash(struct super_block *sb, unsigned long hashval)
fs/smb/server/vfs_cache.c
163
tmp = (hashval * (unsigned long)sb) ^ (GOLDEN_RATIO_PRIME + hashval) /
fs/squashfs/block.c
216
static int squashfs_bio_read(struct super_block *sb, u64 index, int length,
fs/squashfs/block.c
219
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/block.c
234
bio_init_inline(bio, sb->s_bdev, page_count, REQ_OP_READ);
fs/squashfs/block.c
290
int squashfs_read_data(struct super_block *sb, u64 index, int length,
fs/squashfs/block.c
293
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/block.c
319
res = squashfs_bio_read(sb, index, 2, &bio, &offset);
fs/squashfs/block.c
360
res = squashfs_bio_read(sb, index, length, &bio, &offset);
fs/squashfs/cache.c
110
entry->length = squashfs_read_data(sb, block, length,
fs/squashfs/cache.c
335
int squashfs_read_metadata(struct super_block *sb, void *buffer,
fs/squashfs/cache.c
338
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/cache.c
351
entry = squashfs_cache_get(sb, msblk->block_cache, *block, 0);
fs/squashfs/cache.c
386
struct squashfs_cache_entry *squashfs_get_fragment(struct super_block *sb,
fs/squashfs/cache.c
389
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/cache.c
391
return squashfs_cache_get(sb, msblk->fragment_cache, start_block,
fs/squashfs/cache.c
401
struct squashfs_cache_entry *squashfs_get_datablock(struct super_block *sb,
fs/squashfs/cache.c
404
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/cache.c
406
return squashfs_cache_get(sb, msblk->read_page, start_block, length);
fs/squashfs/cache.c
413
void *squashfs_read_table(struct super_block *sb, u64 block, int length)
fs/squashfs/cache.c
439
res = squashfs_read_data(sb, block, length |
fs/squashfs/cache.c
52
struct squashfs_cache_entry *squashfs_cache_get(struct super_block *sb,
fs/squashfs/decompressor.c
110
length = squashfs_read_data(sb,
fs/squashfs/decompressor.c
128
void *squashfs_decompressor_setup(struct super_block *sb, unsigned short flags)
fs/squashfs/decompressor.c
130
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/decompressor.c
131
void *stream, *comp_opts = get_comp_opts(sb, flags);
fs/squashfs/decompressor.c
87
static void *get_comp_opts(struct super_block *sb, unsigned short flags)
fs/squashfs/decompressor.c
89
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/dir.c
39
static int get_dir_index_using_offset(struct super_block *sb,
fs/squashfs/dir.c
43
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/dir.c
61
err = squashfs_read_metadata(sb, &dir_index, &index_start,
fs/squashfs/dir.c
79
err = squashfs_read_metadata(sb, NULL, &index_start,
fs/squashfs/export.c
115
__le64 *squashfs_read_inode_lookup_table(struct super_block *sb,
fs/squashfs/export.c
139
table = squashfs_read_table(sb, lookup_table_start, length);
fs/squashfs/export.c
39
static long long squashfs_inode_lookup(struct super_block *sb, int ino_num)
fs/squashfs/export.c
41
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/export.c
55
err = squashfs_read_metadata(sb, &ino, &start, &offset, sizeof(ino));
fs/squashfs/export.c
66
static struct dentry *squashfs_export_iget(struct super_block *sb,
fs/squashfs/export.c
74
ino = squashfs_inode_lookup(sb, ino_num);
fs/squashfs/export.c
76
dentry = d_obtain_alias(squashfs_iget(sb, ino, ino_num));
fs/squashfs/export.c
82
static struct dentry *squashfs_fh_to_dentry(struct super_block *sb,
fs/squashfs/export.c
89
return squashfs_export_iget(sb, fid->i32.ino);
fs/squashfs/export.c
93
static struct dentry *squashfs_fh_to_parent(struct super_block *sb,
fs/squashfs/export.c
99
return squashfs_export_iget(sb, fid->i32.parent_ino);
fs/squashfs/file.c
162
static long long read_indexes(struct super_block *sb, int n,
fs/squashfs/file.c
177
err = squashfs_read_metadata(sb, blist, start_block,
fs/squashfs/file.c
673
struct super_block *sb = inode->i_sb;
fs/squashfs/file.c
674
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/file.c
716
offset = squashfs_read_metadata(sb, blist, &start, &s_offset, indexes << 2);
fs/squashfs/fragment.c
35
int squashfs_frag_lookup(struct super_block *sb, unsigned int fragment,
fs/squashfs/fragment.c
38
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/fragment.c
50
size = squashfs_read_metadata(sb, &fragment_entry, &start_block,
fs/squashfs/fragment.c
63
__le64 *squashfs_read_fragment_index_table(struct super_block *sb,
fs/squashfs/fragment.c
77
table = squashfs_read_table(sb, fragment_table_start, length);
fs/squashfs/id.c
32
int squashfs_get_id(struct super_block *sb, unsigned int index,
fs/squashfs/id.c
35
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/id.c
47
err = squashfs_read_metadata(sb, &disk_id, &start_block, &offset,
fs/squashfs/id.c
60
__le64 *squashfs_read_id_index_table(struct super_block *sb,
fs/squashfs/id.c
84
table = squashfs_read_table(sb, id_table_start, length);
fs/squashfs/inode.c
109
struct super_block *sb = inode->i_sb;
fs/squashfs/inode.c
110
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/inode.c
122
err = squashfs_read_metadata(sb, sqshb_ino, &block,
fs/squashfs/inode.c
127
err = squashfs_new_inode(sb, inode, sqshb_ino);
fs/squashfs/inode.c
142
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
159
frag_size = squashfs_frag_lookup(sb, frag, &frag_blk);
fs/squashfs/inode.c
194
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
215
frag_size = squashfs_frag_lookup(sb, frag, &frag_blk);
fs/squashfs/inode.c
251
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
275
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
303
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
326
err = squashfs_read_metadata(sb, NULL, &block,
fs/squashfs/inode.c
330
err = squashfs_read_metadata(sb, &xattr, &block,
fs/squashfs/inode.c
347
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
370
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
394
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
412
err = squashfs_read_metadata(sb, sqsh_ino, &block, &offset,
fs/squashfs/inode.c
434
err = squashfs_xattr_lookup(sb, xattr_id,
fs/squashfs/inode.c
44
static int squashfs_new_inode(struct super_block *sb, struct inode *inode,
fs/squashfs/inode.c
55
err = squashfs_get_id(sb, le16_to_cpu(sqsh_ino->uid), &i_uid);
fs/squashfs/inode.c
59
err = squashfs_get_id(sb, le16_to_cpu(sqsh_ino->guid), &i_gid);
fs/squashfs/inode.c
79
struct inode *squashfs_iget(struct super_block *sb, long long ino,
fs/squashfs/inode.c
82
struct inode *inode = iget_locked(sb, ino_number);
fs/squashfs/namei.c
63
static int get_dir_index_using_name(struct super_block *sb,
fs/squashfs/namei.c
67
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/namei.c
81
err = squashfs_read_metadata(sb, index, &index_start,
fs/squashfs/namei.c
91
err = squashfs_read_metadata(sb, index->name, &index_start,
fs/squashfs/super.c
180
static int squashfs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/squashfs/super.c
190
int err, devblksize = sb_min_blocksize(sb, SQUASHFS_DEVBLK_SIZE);
fs/squashfs/super.c
199
sb->s_fs_info = kzalloc_obj(*msblk);
fs/squashfs/super.c
200
if (sb->s_fs_info == NULL) {
fs/squashfs/super.c
204
msblk = sb->s_fs_info;
fs/squashfs/super.c
221
sblk = squashfs_read_table(sb, SQUASHFS_START, sizeof(*sblk));
fs/squashfs/super.c
233
sb->s_magic = le32_to_cpu(sblk->s_magic);
fs/squashfs/super.c
234
if (sb->s_magic != SQUASHFS_MAGIC) {
fs/squashfs/super.c
237
sb->s_bdev);
fs/squashfs/super.c
260
msblk->bytes_used > bdev_nr_bytes(sb->s_bdev))
fs/squashfs/super.c
299
TRACE("Found valid superblock on %pg\n", sb->s_bdev);
fs/squashfs/super.c
316
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/squashfs/super.c
317
sb->s_time_min = 0;
fs/squashfs/super.c
318
sb->s_time_max = U32_MAX;
fs/squashfs/super.c
319
sb->s_flags |= SB_RDONLY;
fs/squashfs/super.c
320
sb->s_op = &squashfs_super_ops;
fs/squashfs/super.c
339
struct inode *cache = new_inode(sb);
fs/squashfs/super.c
353
msblk->stream = squashfs_decompressor_setup(sb, flags);
fs/squashfs/super.c
361
sb->s_xattr = squashfs_xattr_handlers;
fs/squashfs/super.c
369
msblk->xattr_id_table = squashfs_read_xattr_id_table(sb,
fs/squashfs/super.c
382
msblk->id_table = squashfs_read_id_index_table(sb,
fs/squashfs/super.c
398
msblk->inode_lookup_table = squashfs_read_inode_lookup_table(sb,
fs/squashfs/super.c
408
sb->s_export_op = &squashfs_export_ops;
fs/squashfs/super.c
423
msblk->fragment_index = squashfs_read_fragment_index_table(sb,
fs/squashfs/super.c
447
root = new_inode(sb);
fs/squashfs/super.c
461
sb->s_root = d_make_root(root);
fs/squashfs/super.c
462
if (sb->s_root == NULL) {
fs/squashfs/super.c
485
kfree(sb->s_fs_info);
fs/squashfs/super.c
486
sb->s_fs_info = NULL;
fs/squashfs/super.c
498
struct super_block *sb = fc->root->d_sb;
fs/squashfs/super.c
499
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/super.c
524
struct super_block *sb = root->d_sb;
fs/squashfs/super.c
525
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/super.c
591
static void squashfs_put_super(struct super_block *sb)
fs/squashfs/super.c
593
if (sb->s_fs_info) {
fs/squashfs/super.c
594
struct squashfs_sb_info *sbi = sb->s_fs_info;
fs/squashfs/super.c
606
kfree(sb->s_fs_info);
fs/squashfs/super.c
607
sb->s_fs_info = NULL;
fs/squashfs/super.c
670
static struct inode *squashfs_alloc_inode(struct super_block *sb)
fs/squashfs/super.c
673
alloc_inode_sb(sb, squashfs_inode_cachep, GFP_KERNEL);
fs/squashfs/symlink.c
36
struct super_block *sb = inode->i_sb;
fs/squashfs/symlink.c
37
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/symlink.c
53
bytes = squashfs_read_metadata(sb, NULL, &block, &offset,
fs/squashfs/symlink.c
72
entry = squashfs_cache_get(sb, msblk->block_cache, block, 0);
fs/squashfs/xattr.c
108
struct super_block *sb = inode->i_sb;
fs/squashfs/xattr.c
109
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/xattr.c
127
err = squashfs_read_metadata(sb, &entry, &start, &offset,
fs/squashfs/xattr.c
137
err = squashfs_read_metadata(sb, target, &start,
fs/squashfs/xattr.c
140
err = squashfs_read_metadata(sb, NULL, &start,
fs/squashfs/xattr.c
152
err = squashfs_read_metadata(sb, &val, &start,
fs/squashfs/xattr.c
156
err = squashfs_read_metadata(sb, &xattr_val,
fs/squashfs/xattr.c
166
err = squashfs_read_metadata(sb, &val, &start, &offset,
fs/squashfs/xattr.c
177
err = squashfs_read_metadata(sb, buffer, &start,
fs/squashfs/xattr.c
186
err = squashfs_read_metadata(sb, &val, &start, &offset,
fs/squashfs/xattr.c
190
err = squashfs_read_metadata(sb, NULL, &start, &offset,
fs/squashfs/xattr.c
30
struct super_block *sb = inode->i_sb;
fs/squashfs/xattr.c
31
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/xattr.c
50
err = squashfs_read_metadata(sb, &entry, &start, &offset,
fs/squashfs/xattr.c
69
err = squashfs_read_metadata(sb, buffer, &start,
fs/squashfs/xattr.c
80
err = squashfs_read_metadata(sb, NULL, &start,
fs/squashfs/xattr.c
88
err = squashfs_read_metadata(sb, &val, &start, &offset,
fs/squashfs/xattr.c
93
err = squashfs_read_metadata(sb, NULL, &start, &offset,
fs/squashfs/xattr.h
17
static inline __le64 *squashfs_read_xattr_id_table(struct super_block *sb,
fs/squashfs/xattr.h
22
id_table = squashfs_read_table(sb, start, sizeof(*id_table));
fs/squashfs/xattr.h
33
static inline int squashfs_xattr_lookup(struct super_block *sb,
fs/squashfs/xattr_id.c
28
int squashfs_xattr_lookup(struct super_block *sb, unsigned int index,
fs/squashfs/xattr_id.c
31
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/xattr_id.c
43
err = squashfs_read_metadata(sb, &id, &start_block, &offset,
fs/squashfs/xattr_id.c
58
__le64 *squashfs_read_xattr_id_table(struct super_block *sb, u64 table_start,
fs/squashfs/xattr_id.c
61
struct squashfs_sb_info *msblk = sb->s_fs_info;
fs/squashfs/xattr_id.c
68
id_table = squashfs_read_table(sb, table_start, sizeof(*id_table));
fs/squashfs/xattr_id.c
95
table = squashfs_read_table(sb, start, len);
fs/super.c
1001
locked = super_lock(sb, excl);
fs/super.c
1003
return sb;
fs/super.c
1006
__put_super(sb);
fs/super.c
1021
struct super_block *sb = fc->root->d_sb;
fs/super.c
1029
if (sb->s_writers.frozen != SB_UNFROZEN)
fs/super.c
1032
retval = security_sb_remount(sb, fc->security);
fs/super.c
1038
if (!(fc->sb_flags & SB_RDONLY) && sb->s_bdev &&
fs/super.c
1039
bdev_read_only(sb->s_bdev))
fs/super.c
1042
remount_rw = !(fc->sb_flags & SB_RDONLY) && sb_rdonly(sb);
fs/super.c
1043
remount_ro = (fc->sb_flags & SB_RDONLY) && !sb_rdonly(sb);
fs/super.c
1047
if (!hlist_empty(&sb->s_pins)) {
fs/super.c
1048
super_unlock_excl(sb);
fs/super.c
1049
group_pin_kill(&sb->s_pins);
fs/super.c
1050
__super_lock_excl(sb);
fs/super.c
1051
if (!sb->s_root)
fs/super.c
1053
if (sb->s_writers.frozen != SB_UNFROZEN)
fs/super.c
1055
remount_ro = !sb_rdonly(sb);
fs/super.c
1058
shrink_dcache_sb(sb);
fs/super.c
1065
sb_start_ro_state_change(sb);
fs/super.c
1067
retval = sb_prepare_remount_readonly(sb);
fs/super.c
1076
sb_start_ro_state_change(sb);
fs/super.c
1086
sb->s_type->name, retval);
fs/super.c
1090
WRITE_ONCE(sb->s_flags, ((sb->s_flags & ~fc->sb_flags_mask) |
fs/super.c
1092
sb_end_ro_state_change(sb);
fs/super.c
1102
if (remount_ro && sb->s_bdev)
fs/super.c
1103
invalidate_bdev(sb->s_bdev);
fs/super.c
1107
sb_end_ro_state_change(sb);
fs/super.c
111
static __must_check bool super_lock(struct super_block *sb, bool excl)
fs/super.c
1111
static void do_emergency_remount_callback(struct super_block *sb, void *unused)
fs/super.c
1113
if (sb->s_bdev && !sb_rdonly(sb)) {
fs/super.c
1116
fc = fs_context_for_reconfigure(sb->s_root,
fs/super.c
113
lockdep_assert_not_held(&sb->s_umount);
fs/super.c
1145
static void do_thaw_all_callback(struct super_block *sb, void *unused)
fs/super.c
1148
while (sb->s_bdev && !bdev_thaw(sb->s_bdev))
fs/super.c
1149
pr_warn("Emergency Thaw on %pg\n", sb->s_bdev);
fs/super.c
1150
thaw_super_locked(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/super.c
116
wait_var_event(&sb->s_flags, super_flags(sb, SB_BORN | SB_DYING));
fs/super.c
1177
static inline bool get_active_super(struct super_block *sb)
fs/super.c
1181
if (super_lock_excl(sb)) {
fs/super.c
1182
active = atomic_inc_not_zero(&sb->s_active);
fs/super.c
1183
super_unlock_excl(sb);
fs/super.c
119
if (super_flags(sb, SB_DYING))
fs/super.c
1190
static void filesystems_freeze_callback(struct super_block *sb, void *freeze_all_ptr)
fs/super.c
1192
if (!sb->s_op->freeze_fs && !sb->s_op->freeze_super)
fs/super.c
1195
if (!freeze_all_ptr && !(sb->s_type->fs_flags & FS_POWER_FREEZE))
fs/super.c
1198
if (!get_active_super(sb))
fs/super.c
1201
if (sb->s_op->freeze_super)
fs/super.c
1202
sb->s_op->freeze_super(sb, FREEZE_EXCL | FREEZE_HOLDER_KERNEL,
fs/super.c
1205
freeze_super(sb, FREEZE_EXCL | FREEZE_HOLDER_KERNEL,
fs/super.c
1208
deactivate_super(sb);
fs/super.c
122
__super_lock(sb, excl);
fs/super.c
1221
static void filesystems_thaw_callback(struct super_block *sb, void *unused)
fs/super.c
1223
if (!sb->s_op->freeze_fs && !sb->s_op->freeze_super)
fs/super.c
1226
if (!get_active_super(sb))
fs/super.c
1229
if (sb->s_op->thaw_super)
fs/super.c
1230
sb->s_op->thaw_super(sb, FREEZE_EXCL | FREEZE_HOLDER_KERNEL,
fs/super.c
1233
thaw_super(sb, FREEZE_EXCL | FREEZE_HOLDER_KERNEL,
fs/super.c
1236
deactivate_super(sb);
fs/super.c
1289
void kill_anon_super(struct super_block *sb)
fs/super.c
129
if (sb->s_flags & SB_DYING) {
fs/super.c
1291
dev_t dev = sb->s_dev;
fs/super.c
1292
generic_shutdown_super(sb);
fs/super.c
1293
kill_super_notify(sb);
fs/super.c
1298
int set_anon_super_fc(struct super_block *sb, struct fs_context *fc)
fs/super.c
130
super_unlock(sb, excl);
fs/super.c
1300
return set_anon_super(sb, NULL);
fs/super.c
1304
static int test_keyed_super(struct super_block *sb, struct fs_context *fc)
fs/super.c
1306
return sb->s_fs_info == fc->s_fs_info;
fs/super.c
1316
int (*fill_super)(struct super_block *sb,
fs/super.c
1319
struct super_block *sb;
fs/super.c
1322
sb = sget_fc(fc, test, set_anon_super_fc);
fs/super.c
1323
if (IS_ERR(sb))
fs/super.c
1324
return PTR_ERR(sb);
fs/super.c
1326
if (!sb->s_root) {
fs/super.c
1327
err = fill_super(sb, fc);
fs/super.c
1331
sb->s_flags |= SB_ACTIVE;
fs/super.c
1334
fc->root = dget(sb->s_root);
fs/super.c
1338
deactivate_locked_super(sb);
fs/super.c
134
WARN_ON_ONCE(!(sb->s_flags & SB_BORN));
fs/super.c
1343
int (*fill_super)(struct super_block *sb,
fs/super.c
1351
int (*fill_super)(struct super_block *sb,
fs/super.c
1359
int (*fill_super)(struct super_block *sb,
fs/super.c
139
static inline bool super_lock_shared(struct super_block *sb)
fs/super.c
141
return super_lock(sb, false);
fs/super.c
1423
struct super_block *sb = bdev->bd_holder;
fs/super.c
1427
lockdep_assert_not_held(&sb->s_umount);
fs/super.c
1432
sb->s_count++;
fs/super.c
1437
locked = super_lock(sb, excl);
fs/super.c
1443
put_super(sb);
fs/super.c
1448
if (!sb->s_root || !(sb->s_flags & SB_ACTIVE)) {
fs/super.c
1449
super_unlock(sb, excl);
fs/super.c
145
static inline bool super_lock_excl(struct super_block *sb)
fs/super.c
1453
return sb;
fs/super.c
1458
struct super_block *sb;
fs/super.c
1460
sb = bdev_super_lock(bdev, false);
fs/super.c
1461
if (!sb)
fs/super.c
1464
if (sb->s_op->remove_bdev) {
fs/super.c
1467
ret = sb->s_op->remove_bdev(sb, bdev);
fs/super.c
1469
super_unlock_shared(sb);
fs/super.c
147
return super_lock(sb, true);
fs/super.c
1476
sync_filesystem(sb);
fs/super.c
1477
shrink_dcache_sb(sb);
fs/super.c
1478
evict_inodes(sb);
fs/super.c
1479
if (sb->s_op->shutdown)
fs/super.c
1480
sb->s_op->shutdown(sb);
fs/super.c
1482
super_unlock_shared(sb);
fs/super.c
1487
struct super_block *sb;
fs/super.c
1489
sb = bdev_super_lock(bdev, false);
fs/super.c
1490
if (!sb)
fs/super.c
1493
sync_filesystem(sb);
fs/super.c
1494
super_unlock_shared(sb);
fs/super.c
1500
struct super_block *sb;
fs/super.c
1502
sb = bdev_super_lock(bdev, true);
fs/super.c
1503
if (sb) {
fs/super.c
1504
active = atomic_inc_not_zero(&sb->s_active);
fs/super.c
1505
super_unlock_excl(sb);
fs/super.c
1509
return sb;
fs/super.c
152
static void super_wake(struct super_block *sb, unsigned int flag)
fs/super.c
1529
struct super_block *sb;
fs/super.c
1534
sb = get_bdev_super(bdev);
fs/super.c
1535
if (!sb)
fs/super.c
1538
if (sb->s_op->freeze_super)
fs/super.c
1539
error = sb->s_op->freeze_super(sb,
fs/super.c
1542
error = freeze_super(sb,
fs/super.c
1546
deactivate_super(sb);
fs/super.c
1569
struct super_block *sb;
fs/super.c
1583
sb = get_bdev_super(bdev);
fs/super.c
1584
if (!sb)
fs/super.c
1587
if (sb->s_op->thaw_super)
fs/super.c
1588
error = sb->s_op->thaw_super(sb,
fs/super.c
1591
error = thaw_super(sb,
fs/super.c
1593
deactivate_super(sb);
fs/super.c
1605
int setup_bdev_super(struct super_block *sb, int sb_flags,
fs/super.c
1612
bdev_file = bdev_file_open_by_dev(sb->s_dev, mode, sb, &fs_holder_ops);
fs/super.c
162
smp_store_release(&sb->s_flags, sb->s_flags | flag);
fs/super.c
1641
sb->s_bdev_file = bdev_file;
fs/super.c
1642
sb->s_bdev = bdev;
fs/super.c
1643
sb->s_bdi = bdi_get(bdev->bd_disk->bdi);
fs/super.c
1645
sb->s_iflags |= SB_I_STABLE_WRITES;
fs/super.c
1648
snprintf(sb->s_id, sizeof(sb->s_id), "%pg", bdev);
fs/super.c
1649
shrinker_debugfs_rename(sb->s_shrink, "sb-%s:%s", sb->s_type->name,
fs/super.c
1650
sb->s_id);
fs/super.c
1651
sb_set_blocksize(sb, block_size(bdev));
fs/super.c
1663
int (*fill_super)(struct super_block *sb,
fs/super.c
169
wake_up_var(&sb->s_flags);
fs/super.c
1721
void kill_block_super(struct super_block *sb)
fs/super.c
1723
struct block_device *bdev = sb->s_bdev;
fs/super.c
1725
generic_shutdown_super(sb);
fs/super.c
1728
bdev_fput(sb->s_bdev_file);
fs/super.c
1745
struct super_block *sb;
fs/super.c
1767
sb = fc->root->d_sb;
fs/super.c
1768
WARN_ON(!sb->s_bdi);
fs/super.c
1777
super_wake(sb, SB_BORN);
fs/super.c
1779
error = security_sb_set_mnt_opts(sb, fc->security, 0, NULL);
fs/super.c
1791
WARN((sb->s_maxbytes < 0), "%s set sb->s_maxbytes to "
fs/super.c
1792
"negative value (%lld)\n", fc->fs_type->name, sb->s_maxbytes);
fs/super.c
1802
int super_setup_bdi_name(struct super_block *sb, char *fmt, ...)
fs/super.c
1819
WARN_ON(sb->s_bdi != &noop_backing_dev_info);
fs/super.c
182
struct super_block *sb;
fs/super.c
1820
sb->s_bdi = bdi;
fs/super.c
1821
sb->s_iflags |= SB_I_PERSB_BDI;
fs/super.c
1831
int super_setup_bdi(struct super_block *sb)
fs/super.c
1835
return super_setup_bdi_name(sb, "%.28s-%ld", sb->s_type->name,
fs/super.c
1848
static void sb_wait_write(struct super_block *sb, int level)
fs/super.c
1850
percpu_down_write(sb->s_writers.rw_sem + level-1);
fs/super.c
1857
static void lockdep_sb_freeze_release(struct super_block *sb)
fs/super.c
1862
percpu_rwsem_release(sb->s_writers.rw_sem + level, _THIS_IP_);
fs/super.c
1868
static void lockdep_sb_freeze_acquire(struct super_block *sb)
fs/super.c
1873
percpu_rwsem_acquire(sb->s_writers.rw_sem + level, 0, _THIS_IP_);
fs/super.c
1876
static void sb_freeze_unlock(struct super_block *sb, int level)
fs/super.c
1879
percpu_up_write(sb->s_writers.rw_sem + level);
fs/super.c
1882
static int wait_for_partially_frozen(struct super_block *sb)
fs/super.c
1887
unsigned short old = sb->s_writers.frozen;
fs/super.c
1889
up_write(&sb->s_umount);
fs/super.c
189
sb = shrink->private_data;
fs/super.c
1890
ret = wait_var_event_killable(&sb->s_writers.frozen,
fs/super.c
1891
sb->s_writers.frozen != old);
fs/super.c
1892
down_write(&sb->s_umount);
fs/super.c
1894
sb->s_writers.frozen != SB_UNFROZEN &&
fs/super.c
1895
sb->s_writers.frozen != SB_FREEZE_COMPLETE);
fs/super.c
1903
static inline int freeze_inc(struct super_block *sb, enum freeze_holder who)
fs/super.c
1909
++sb->s_writers.freeze_kcount;
fs/super.c
1911
++sb->s_writers.freeze_ucount;
fs/super.c
1912
return sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount;
fs/super.c
1915
static inline int freeze_dec(struct super_block *sb, enum freeze_holder who)
fs/super.c
1920
if ((who & FREEZE_HOLDER_KERNEL) && sb->s_writers.freeze_kcount)
fs/super.c
1921
--sb->s_writers.freeze_kcount;
fs/super.c
1922
if ((who & FREEZE_HOLDER_USERSPACE) && sb->s_writers.freeze_ucount)
fs/super.c
1923
--sb->s_writers.freeze_ucount;
fs/super.c
1924
return sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount;
fs/super.c
1927
static inline bool may_freeze(struct super_block *sb, enum freeze_holder who,
fs/super.c
1930
lockdep_assert_held(&sb->s_umount);
fs/super.c
1943
if (sb->s_writers.freeze_owner)
fs/super.c
1950
if (sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount)
fs/super.c
1951
sb->s_writers.freeze_owner = freeze_owner;
fs/super.c
1957
sb->s_writers.freeze_kcount == 0;
fs/super.c
1960
sb->s_writers.freeze_ucount == 0;
fs/super.c
1964
static inline bool may_unfreeze(struct super_block *sb, enum freeze_holder who,
fs/super.c
1967
lockdep_assert_held(&sb->s_umount);
fs/super.c
1979
if (WARN_ON_ONCE(sb->s_writers.freeze_kcount == 0))
fs/super.c
198
if (!super_trylock_shared(sb))
fs/super.c
1982
if (!sb->s_writers.freeze_owner)
fs/super.c
1985
if (sb->s_writers.freeze_owner != freeze_owner)
fs/super.c
1992
if ((sb->s_writers.freeze_kcount + sb->s_writers.freeze_ucount) > 1)
fs/super.c
1993
sb->s_writers.freeze_owner = NULL;
fs/super.c
2002
if (sb->s_writers.freeze_kcount == 1 &&
fs/super.c
2003
sb->s_writers.freeze_owner)
fs/super.c
2005
return sb->s_writers.freeze_kcount > 0;
fs/super.c
2009
return sb->s_writers.freeze_ucount > 0;
fs/super.c
201
if (sb->s_op->nr_cached_objects)
fs/super.c
202
fs_objects = sb->s_op->nr_cached_objects(sb, sc);
fs/super.c
204
inodes = list_lru_shrink_count(&sb->s_inode_lru, sc);
fs/super.c
205
dentries = list_lru_shrink_count(&sb->s_dentry_lru, sc);
fs/super.c
2070
int freeze_super(struct super_block *sb, enum freeze_holder who, const void *freeze_owner)
fs/super.c
2074
if (!super_lock_excl(sb)) {
fs/super.c
2078
atomic_inc(&sb->s_active);
fs/super.c
2081
if (sb->s_writers.frozen == SB_FREEZE_COMPLETE) {
fs/super.c
2082
if (may_freeze(sb, who, freeze_owner))
fs/super.c
2083
ret = !!WARN_ON_ONCE(freeze_inc(sb, who) == 1);
fs/super.c
2087
deactivate_locked_super(sb);
fs/super.c
2091
if (sb->s_writers.frozen != SB_UNFROZEN) {
fs/super.c
2092
ret = wait_for_partially_frozen(sb);
fs/super.c
2094
deactivate_locked_super(sb);
fs/super.c
2101
if (sb_rdonly(sb)) {
fs/super.c
2103
WARN_ON_ONCE(freeze_inc(sb, who) > 1);
fs/super.c
2104
sb->s_writers.freeze_owner = freeze_owner;
fs/super.c
2105
sb->s_writers.frozen = SB_FREEZE_COMPLETE;
fs/super.c
2106
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2107
super_unlock_excl(sb);
fs/super.c
2111
sb->s_writers.frozen = SB_FREEZE_WRITE;
fs/super.c
2113
super_unlock_excl(sb);
fs/super.c
2114
sb_wait_write(sb, SB_FREEZE_WRITE);
fs/super.c
2115
__super_lock_excl(sb);
fs/super.c
2118
sb->s_writers.frozen = SB_FREEZE_PAGEFAULT;
fs/super.c
2119
sb_wait_write(sb, SB_FREEZE_PAGEFAULT);
fs/super.c
2122
ret = sync_filesystem(sb);
fs/super.c
2124
sb->s_writers.frozen = SB_UNFROZEN;
fs/super.c
2125
sb_freeze_unlock(sb, SB_FREEZE_PAGEFAULT);
fs/super.c
2126
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2127
deactivate_locked_super(sb);
fs/super.c
2132
sb->s_writers.frozen = SB_FREEZE_FS;
fs/super.c
2133
sb_wait_write(sb, SB_FREEZE_FS);
fs/super.c
2135
if (sb->s_op->freeze_fs) {
fs/super.c
2136
ret = sb->s_op->freeze_fs(sb);
fs/super.c
2140
sb->s_writers.frozen = SB_UNFROZEN;
fs/super.c
2141
sb_freeze_unlock(sb, SB_FREEZE_FS);
fs/super.c
2142
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2143
deactivate_locked_super(sb);
fs/super.c
2151
WARN_ON_ONCE(freeze_inc(sb, who) > 1);
fs/super.c
2152
sb->s_writers.freeze_owner = freeze_owner;
fs/super.c
2153
sb->s_writers.frozen = SB_FREEZE_COMPLETE;
fs/super.c
2154
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2155
lockdep_sb_freeze_release(sb);
fs/super.c
2156
super_unlock_excl(sb);
fs/super.c
2167
static int thaw_super_locked(struct super_block *sb, enum freeze_holder who,
fs/super.c
2172
if (sb->s_writers.frozen != SB_FREEZE_COMPLETE)
fs/super.c
2175
if (!may_unfreeze(sb, who, freeze_owner))
fs/super.c
2182
if (freeze_dec(sb, who))
fs/super.c
2185
if (sb_rdonly(sb)) {
fs/super.c
2186
sb->s_writers.frozen = SB_UNFROZEN;
fs/super.c
2187
sb->s_writers.freeze_owner = NULL;
fs/super.c
2188
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2192
lockdep_sb_freeze_acquire(sb);
fs/super.c
2194
if (sb->s_op->unfreeze_fs) {
fs/super.c
2195
error = sb->s_op->unfreeze_fs(sb);
fs/super.c
2198
freeze_inc(sb, who);
fs/super.c
2199
lockdep_sb_freeze_release(sb);
fs/super.c
2204
sb->s_writers.frozen = SB_UNFROZEN;
fs/super.c
2205
sb->s_writers.freeze_owner = NULL;
fs/super.c
2206
wake_up_var(&sb->s_writers.frozen);
fs/super.c
2207
sb_freeze_unlock(sb, SB_FREEZE_FS);
fs/super.c
2209
deactivate_locked_super(sb);
fs/super.c
2213
super_unlock_excl(sb);
fs/super.c
223
freed = prune_dcache_sb(sb, sc);
fs/super.c
2235
int thaw_super(struct super_block *sb, enum freeze_holder who,
fs/super.c
2238
if (!super_lock_excl(sb)) {
fs/super.c
2242
return thaw_super_locked(sb, who, freeze_owner);
fs/super.c
225
freed += prune_icache_sb(sb, sc);
fs/super.c
2252
int sb_init_dio_done_wq(struct super_block *sb)
fs/super.c
2258
sb->s_id);
fs/super.c
2266
if (!try_cmpxchg(&sb->s_dio_done_wq, &old, wq)) {
fs/super.c
229
freed += sb->s_op->free_cached_objects(sb, sc);
fs/super.c
232
super_unlock_shared(sb);
fs/super.c
239
struct super_block *sb;
fs/super.c
242
sb = shrink->private_data;
fs/super.c
258
if (!(sb->s_flags & SB_BORN))
fs/super.c
262
if (sb->s_op && sb->s_op->nr_cached_objects)
fs/super.c
263
total_objects = sb->s_op->nr_cached_objects(sb, sc);
fs/super.c
265
total_objects += list_lru_shrink_count(&sb->s_dentry_lru, sc);
fs/super.c
266
total_objects += list_lru_shrink_count(&sb->s_inode_lru, sc);
fs/super.c
425
void put_super(struct super_block *sb)
fs/super.c
428
__put_super(sb);
fs/super.c
43
static int thaw_super_locked(struct super_block *sb, enum freeze_holder who,
fs/super.c
432
static void kill_super_notify(struct super_block *sb)
fs/super.c
434
lockdep_assert_not_held(&sb->s_umount);
fs/super.c
437
if (sb->s_flags & SB_DEAD)
fs/super.c
448
hlist_del_init(&sb->s_instances);
fs/super.c
457
super_wake(sb, SB_DEAD);
fs/super.c
527
static bool grab_super(struct super_block *sb)
fs/super.c
531
sb->s_count++;
fs/super.c
533
locked = super_lock_excl(sb);
fs/super.c
535
if (atomic_inc_not_zero(&sb->s_active)) {
fs/super.c
536
put_super(sb);
fs/super.c
539
super_unlock_excl(sb);
fs/super.c
541
wait_var_event(&sb->s_flags, super_flags(sb, SB_DEAD));
fs/super.c
542
put_super(sb);
fs/super.c
55
static inline void __super_lock(struct super_block *sb, bool excl)
fs/super.c
563
bool super_trylock_shared(struct super_block *sb)
fs/super.c
565
if (down_read_trylock(&sb->s_umount)) {
fs/super.c
566
if (!(sb->s_flags & SB_DYING) && sb->s_root &&
fs/super.c
567
(sb->s_flags & SB_BORN))
fs/super.c
569
super_unlock_shared(sb);
fs/super.c
58
down_write(&sb->s_umount);
fs/super.c
591
void retire_super(struct super_block *sb)
fs/super.c
593
WARN_ON(!sb->s_bdev);
fs/super.c
594
__super_lock_excl(sb);
fs/super.c
595
if (sb->s_iflags & SB_I_PERSB_BDI) {
fs/super.c
596
bdi_unregister(sb->s_bdi);
fs/super.c
597
sb->s_iflags &= ~SB_I_PERSB_BDI;
fs/super.c
599
sb->s_iflags |= SB_I_RETIRED;
fs/super.c
60
down_read(&sb->s_umount);
fs/super.c
600
super_unlock_excl(sb);
fs/super.c
618
void generic_shutdown_super(struct super_block *sb)
fs/super.c
620
const struct super_operations *sop = sb->s_op;
fs/super.c
622
if (sb->s_root) {
fs/super.c
623
fsnotify_sb_delete(sb);
fs/super.c
624
shrink_dcache_for_umount(sb);
fs/super.c
625
sync_filesystem(sb);
fs/super.c
626
sb->s_flags &= ~SB_ACTIVE;
fs/super.c
628
fserror_unmount(sb);
fs/super.c
629
cgroup_writeback_umount(sb);
fs/super.c
63
static inline void super_unlock(struct super_block *sb, bool excl)
fs/super.c
632
evict_inodes(sb);
fs/super.c
638
security_sb_delete(sb);
fs/super.c
640
if (sb->s_dio_done_wq) {
fs/super.c
641
destroy_workqueue(sb->s_dio_done_wq);
fs/super.c
642
sb->s_dio_done_wq = NULL;
fs/super.c
646
sop->put_super(sb);
fs/super.c
652
fscrypt_destroy_keyring(sb);
fs/super.c
654
if (CHECK_DATA_CORRUPTION(!list_empty(&sb->s_inodes), NULL,
fs/super.c
656
sb->s_id, sb->s_type->name)) {
fs/super.c
66
up_write(&sb->s_umount);
fs/super.c
664
spin_lock(&sb->s_inode_list_lock);
fs/super.c
665
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
fs/super.c
670
spin_unlock(&sb->s_inode_list_lock);
fs/super.c
68
up_read(&sb->s_umount);
fs/super.c
682
super_wake(sb, SB_DYING);
fs/super.c
683
super_unlock_excl(sb);
fs/super.c
684
if (sb->s_bdi != &noop_backing_dev_info) {
fs/super.c
685
if (sb->s_iflags & SB_I_PERSB_BDI)
fs/super.c
686
bdi_unregister(sb->s_bdi);
fs/super.c
687
bdi_put(sb->s_bdi);
fs/super.c
688
sb->s_bdi = &noop_backing_dev_info;
fs/super.c
71
static inline void __super_lock_excl(struct super_block *sb)
fs/super.c
73
__super_lock(sb, true);
fs/super.c
76
static inline void super_unlock_excl(struct super_block *sb)
fs/super.c
78
super_unlock(sb, true);
fs/super.c
81
static inline void super_unlock_shared(struct super_block *sb)
fs/super.c
83
super_unlock(sb, false);
fs/super.c
86
static bool super_flags(const struct super_block *sb, unsigned int flags)
fs/super.c
872
void drop_super(struct super_block *sb)
fs/super.c
874
super_unlock_shared(sb);
fs/super.c
875
put_super(sb);
fs/super.c
880
void drop_super_exclusive(struct super_block *sb)
fs/super.c
882
super_unlock_excl(sb);
fs/super.c
883
put_super(sb);
fs/super.c
900
static inline struct super_block *next_super(struct super_block *sb,
fs/super.c
904
return list_prev_entry(sb, s_list);
fs/super.c
905
return list_next_entry(sb, s_list);
fs/super.c
911
struct super_block *sb, *p = NULL;
fs/super.c
916
for (sb = first_super(flags);
fs/super.c
917
!list_entry_is_head(sb, &super_blocks, s_list);
fs/super.c
918
sb = next_super(sb, flags)) {
fs/super.c
919
if (super_flags(sb, SB_DYING))
fs/super.c
92
return smp_load_acquire(&sb->s_flags) & flags;
fs/super.c
921
sb->s_count++;
fs/super.c
925
f(sb, arg);
fs/super.c
926
} else if (super_lock(sb, excl)) {
fs/super.c
927
f(sb, arg);
fs/super.c
928
super_unlock(sb, excl);
fs/super.c
934
p = sb;
fs/super.c
957
struct super_block *sb, *p = NULL;
fs/super.c
960
hlist_for_each_entry(sb, &type->fs_supers, s_instances) {
fs/super.c
963
if (super_flags(sb, SB_DYING))
fs/super.c
966
sb->s_count++;
fs/super.c
969
locked = super_lock_shared(sb);
fs/super.c
971
f(sb, arg);
fs/super.c
972
super_unlock_shared(sb);
fs/super.c
978
p = sb;
fs/super.c
989
struct super_block *sb;
fs/super.c
992
list_for_each_entry(sb, &super_blocks, s_list) {
fs/super.c
995
if (sb->s_dev != dev)
fs/super.c
998
sb->s_count++;
fs/sync.c
151
struct super_block *sb;
fs/sync.c
156
sb = fd_file(f)->f_path.dentry->d_sb;
fs/sync.c
158
down_read(&sb->s_umount);
fs/sync.c
159
ret = sync_filesystem(sb);
fs/sync.c
160
up_read(&sb->s_umount);
fs/sync.c
162
ret2 = errseq_check_and_advance(&sb->s_wb_err, &fd_file(f)->f_sb_err);
fs/sync.c
30
int sync_filesystem(struct super_block *sb)
fs/sync.c
38
WARN_ON(!rwsem_is_locked(&sb->s_umount));
fs/sync.c
43
if (sb_rdonly(sb))
fs/sync.c
54
writeback_inodes_sb(sb, WB_REASON_SYNC);
fs/sync.c
55
if (sb->s_op->sync_fs) {
fs/sync.c
56
ret = sb->s_op->sync_fs(sb, 0);
fs/sync.c
60
ret = sync_blockdev_nowait(sb->s_bdev);
fs/sync.c
64
sync_inodes_sb(sb);
fs/sync.c
65
if (sb->s_op->sync_fs) {
fs/sync.c
66
ret = sb->s_op->sync_fs(sb, 1);
fs/sync.c
70
return sync_blockdev(sb->s_bdev);
fs/sync.c
74
static void sync_inodes_one_sb(struct super_block *sb, void *arg)
fs/sync.c
76
if (!sb_rdonly(sb))
fs/sync.c
77
sync_inodes_sb(sb);
fs/sync.c
80
static void sync_fs_one_sb(struct super_block *sb, void *arg)
fs/sync.c
82
if (!sb_rdonly(sb) && !(sb->s_iflags & SB_I_SKIP_SYNC) &&
fs/sync.c
83
sb->s_op->sync_fs)
fs/sync.c
84
sb->s_op->sync_fs(sb, *(int *)arg);
fs/sysfs/mount.c
82
static void sysfs_kill_sb(struct super_block *sb)
fs/sysfs/mount.c
84
void *ns = (void *)kernfs_super_ns(sb);
fs/sysfs/mount.c
86
kernfs_kill_sb(sb);
fs/tracefs/inode.c
271
struct inode *tracefs_get_inode(struct super_block *sb)
fs/tracefs/inode.c
273
struct inode *inode = new_inode(sb);
fs/tracefs/inode.c
333
static int tracefs_apply_options(struct super_block *sb, bool remount)
fs/tracefs/inode.c
335
struct tracefs_fs_info *fsi = sb->s_fs_info;
fs/tracefs/inode.c
336
struct inode *inode = d_inode(sb->s_root);
fs/tracefs/inode.c
391
struct super_block *sb = fc->root->d_sb;
fs/tracefs/inode.c
392
struct tracefs_fs_info *sb_opts = sb->s_fs_info;
fs/tracefs/inode.c
398
sync_filesystem(sb);
fs/tracefs/inode.c
40
static struct inode *tracefs_alloc_inode(struct super_block *sb)
fs/tracefs/inode.c
402
return tracefs_apply_options(sb, true);
fs/tracefs/inode.c
45
ti = alloc_inode_sb(sb, tracefs_inode_cachep, GFP_KERNEL);
fs/tracefs/inode.c
484
static int tracefs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/tracefs/inode.c
489
err = simple_fill_super(sb, TRACEFS_MAGIC, trace_files);
fs/tracefs/inode.c
493
sb->s_op = &tracefs_super_operations;
fs/tracefs/inode.c
494
set_default_d_op(sb, &tracefs_dentry_operations);
fs/tracefs/internal.h
74
struct inode *tracefs_get_inode(struct super_block *sb);
fs/ubifs/debug.c
2331
struct ubifs_scan_node *sa, *sb;
fs/ubifs/debug.c
2342
sb = container_of(cur->next, struct ubifs_scan_node, list);
fs/ubifs/debug.c
2349
if (sb->type != UBIFS_DATA_NODE) {
fs/ubifs/debug.c
2350
ubifs_err(c, "bad node type %d", sb->type);
fs/ubifs/debug.c
2351
ubifs_dump_node(c, sb->node, c->leb_size - sb->offs);
fs/ubifs/debug.c
2356
inumb = key_inum(c, &sb->key);
fs/ubifs/debug.c
2367
blkb = key_block(c, &sb->key);
fs/ubifs/debug.c
2383
ubifs_dump_node(c, sb->node, c->leb_size - sb->offs);
fs/ubifs/debug.c
2398
struct ubifs_scan_node *sa, *sb;
fs/ubifs/debug.c
2409
sb = container_of(cur->next, struct ubifs_scan_node, list);
fs/ubifs/debug.c
2417
if (sb->type != UBIFS_INO_NODE && sb->type != UBIFS_DENT_NODE &&
fs/ubifs/debug.c
2418
sb->type != UBIFS_XENT_NODE) {
fs/ubifs/debug.c
2419
ubifs_err(c, "bad node type %d", sb->type);
fs/ubifs/debug.c
2420
ubifs_dump_node(c, sb->node, c->leb_size - sb->offs);
fs/ubifs/debug.c
2424
if (sa->type != UBIFS_INO_NODE && sb->type == UBIFS_INO_NODE) {
fs/ubifs/debug.c
2429
if (sa->type == UBIFS_INO_NODE && sb->type != UBIFS_INO_NODE)
fs/ubifs/debug.c
2432
if (sa->type == UBIFS_INO_NODE && sb->type == UBIFS_INO_NODE) {
fs/ubifs/debug.c
2434
if (sa->len < sb->len) {
fs/ubifs/debug.c
2446
inumb = key_inum(c, &sb->key);
fs/ubifs/debug.c
2457
hashb = key_block(c, &sb->key);
fs/ubifs/debug.c
2472
ubifs_dump_node(c, sb->node, c->leb_size - sb->offs);
fs/ubifs/gc.c
110
struct ubifs_scan_node *sa, *sb;
fs/ubifs/gc.c
117
sb = list_entry(b, struct ubifs_scan_node, list);
fs/ubifs/gc.c
120
ubifs_assert(c, key_type(c, &sb->key) == UBIFS_DATA_KEY);
fs/ubifs/gc.c
122
ubifs_assert(c, sb->type == UBIFS_DATA_NODE);
fs/ubifs/gc.c
125
inumb = key_inum(c, &sb->key);
fs/ubifs/gc.c
129
unsigned int blkb = key_block(c, &sb->key);
fs/ubifs/gc.c
154
struct ubifs_scan_node *sa, *sb;
fs/ubifs/gc.c
161
sb = list_entry(b, struct ubifs_scan_node, list);
fs/ubifs/gc.c
164
key_type(c, &sb->key) != UBIFS_DATA_KEY);
fs/ubifs/gc.c
166
sb->type != UBIFS_DATA_NODE);
fs/ubifs/gc.c
170
if (sb->type == UBIFS_INO_NODE)
fs/ubifs/gc.c
171
return sb->len - sa->len;
fs/ubifs/gc.c
174
if (sb->type == UBIFS_INO_NODE)
fs/ubifs/gc.c
179
ubifs_assert(c, key_type(c, &sb->key) == UBIFS_DENT_KEY ||
fs/ubifs/gc.c
180
key_type(c, &sb->key) == UBIFS_XENT_KEY);
fs/ubifs/gc.c
183
ubifs_assert(c, sb->type == UBIFS_DENT_NODE ||
fs/ubifs/gc.c
184
sb->type == UBIFS_XENT_NODE);
fs/ubifs/gc.c
187
inumb = key_inum(c, &sb->key);
fs/ubifs/gc.c
191
uint32_t hashb = key_hash(c, &sb->key);
fs/ubifs/super.c
103
struct inode *ubifs_iget(struct super_block *sb, unsigned long inum)
fs/ubifs/super.c
108
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/super.c
114
inode = iget_locked(sb, inum);
fs/ubifs/super.c
1869
static void ubifs_put_super(struct super_block *sb)
fs/ubifs/super.c
1872
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/super.c
1950
struct super_block *sb = fc->root->d_sb;
fs/ubifs/super.c
1952
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/super.c
1954
sync_filesystem(sb);
fs/ubifs/super.c
1955
dbg_gen("old flags %#lx, new flags %#x", sb->s_flags, fc->sb_flags);
fs/ubifs/super.c
2129
static int ubifs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ubifs/super.c
2131
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/super.c
2136
c->vfs_sb = sb;
fs/ubifs/super.c
2167
err = super_setup_bdi_name(sb, "ubifs_%d_%d", c->vi.ubi_num,
fs/ubifs/super.c
2171
sb->s_bdi->ra_pages = 0;
fs/ubifs/super.c
2172
sb->s_bdi->io_pages = 0;
fs/ubifs/super.c
2174
sb->s_fs_info = c;
fs/ubifs/super.c
2175
sb->s_magic = UBIFS_SUPER_MAGIC;
fs/ubifs/super.c
2176
sb->s_blocksize = UBIFS_BLOCK_SIZE;
fs/ubifs/super.c
2177
sb->s_blocksize_bits = UBIFS_BLOCK_SHIFT;
fs/ubifs/super.c
2178
sb->s_maxbytes = c->max_inode_sz = key_max_inode_size(c);
fs/ubifs/super.c
2180
sb->s_maxbytes = c->max_inode_sz = MAX_LFS_FILESIZE;
fs/ubifs/super.c
2181
sb->s_op = &ubifs_super_operations;
fs/ubifs/super.c
2182
sb->s_xattr = ubifs_xattr_handlers;
fs/ubifs/super.c
2183
fscrypt_set_ops(sb, &ubifs_crypt_operations);
fs/ubifs/super.c
2193
root = ubifs_iget(sb, UBIFS_ROOT_INO);
fs/ubifs/super.c
2199
generic_set_sb_d_ops(sb);
fs/ubifs/super.c
2200
sb->s_root = d_make_root(root);
fs/ubifs/super.c
2201
if (!sb->s_root) {
fs/ubifs/super.c
2206
super_set_uuid(sb, c->uuid, sizeof(c->uuid));
fs/ubifs/super.c
2207
super_set_sysfs_name_generic(sb, UBIFS_DFS_DIR_NAME,
fs/ubifs/super.c
2224
static int sb_test(struct super_block *sb, struct fs_context *fc)
fs/ubifs/super.c
2227
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/super.c
2236
struct super_block *sb;
fs/ubifs/super.c
2267
sb = sget_fc(fc, sb_test, set_anon_super_fc);
fs/ubifs/super.c
2268
if (IS_ERR(sb)) {
fs/ubifs/super.c
2269
err = PTR_ERR(sb);
fs/ubifs/super.c
2274
if (sb->s_root) {
fs/ubifs/super.c
2275
struct ubifs_info *c1 = sb->s_fs_info;
fs/ubifs/super.c
2284
err = ubifs_fill_super(sb, fc);
fs/ubifs/super.c
2288
sb->s_flags |= SB_ACTIVE;
fs/ubifs/super.c
2292
sb->s_flags |= SB_NOATIME;
fs/ubifs/super.c
2298
fc->root = dget(sb->s_root);
fs/ubifs/super.c
2302
deactivate_locked_super(sb);
fs/ubifs/super.c
263
static struct inode *ubifs_alloc_inode(struct super_block *sb)
fs/ubifs/super.c
267
ui = alloc_inode_sb(sb, ubifs_inode_slab, GFP_NOFS);
fs/ubifs/super.c
463
static int ubifs_sync_fs(struct super_block *sb, int wait)
fs/ubifs/super.c
466
struct ubifs_info *c = sb->s_fs_info;
fs/ubifs/ubifs.h
2063
struct inode *ubifs_iget(struct super_block *sb, unsigned long inum);
fs/udf/balloc.c
106
static void udf_add_free_space(struct super_block *sb, u16 partition, u32 cnt)
fs/udf/balloc.c
108
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
116
udf_updated_lvid(sb);
fs/udf/balloc.c
119
static void udf_bitmap_free_blocks(struct super_block *sb,
fs/udf/balloc.c
125
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
140
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
141
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
146
if (bit + count > (sb->s_blocksize << 3)) {
fs/udf/balloc.c
147
overflow = bit + count - (sb->s_blocksize << 3);
fs/udf/balloc.c
150
bitmap_nr = load_block_bitmap(sb, bitmap, block_group);
fs/udf/balloc.c
162
udf_add_free_space(sb, sbi->s_partition, count);
fs/udf/balloc.c
174
static int udf_bitmap_prealloc_blocks(struct super_block *sb,
fs/udf/balloc.c
179
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
196
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
198
bitmap_nr = load_block_bitmap(sb, bitmap, block_group);
fs/udf/balloc.c
203
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
205
while (bit < (sb->s_blocksize << 3) && block_count > 0) {
fs/udf/balloc.c
217
udf_add_free_space(sb, partition, -alloc_count);
fs/udf/balloc.c
222
static udf_pblk_t udf_bitmap_new_block(struct super_block *sb,
fs/udf/balloc.c
226
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
244
block_group = block >> (sb->s_blocksize_bits + 3);
fs/udf/balloc.c
247
bitmap_nr = load_block_bitmap(sb, bitmap, block_group);
fs/udf/balloc.c
252
sb->s_blocksize - group_start);
fs/udf/balloc.c
254
if ((ptr - ((char *)bh->b_data)) < sb->s_blocksize) {
fs/udf/balloc.c
255
bit = block % (sb->s_blocksize << 3);
fs/udf/balloc.c
265
sb->s_blocksize - ((bit + 7) >> 3));
fs/udf/balloc.c
267
if (newbit < sb->s_blocksize << 3) {
fs/udf/balloc.c
273
sb->s_blocksize << 3, bit);
fs/udf/balloc.c
274
if (newbit < sb->s_blocksize << 3) {
fs/udf/balloc.c
286
bitmap_nr = load_block_bitmap(sb, bitmap, block_group);
fs/udf/balloc.c
292
sb->s_blocksize - group_start);
fs/udf/balloc.c
293
if ((ptr - ((char *)bh->b_data)) < sb->s_blocksize) {
fs/udf/balloc.c
299
sb->s_blocksize << 3,
fs/udf/balloc.c
301
if (bit < sb->s_blocksize << 3)
fs/udf/balloc.c
309
if (bit < sb->s_blocksize << 3)
fs/udf/balloc.c
31
static int read_block_bitmap(struct super_block *sb,
fs/udf/balloc.c
312
bit = udf_find_next_one_bit(bh->b_data, sb->s_blocksize << 3,
fs/udf/balloc.c
314
if (bit >= sb->s_blocksize << 3) {
fs/udf/balloc.c
328
newblock = bit + (block_group << (sb->s_blocksize_bits + 3)) -
fs/udf/balloc.c
336
udf_err(sb, "bitmap for partition %d corrupted (block %u marked"
fs/udf/balloc.c
349
udf_add_free_space(sb, partition, -1);
fs/udf/balloc.c
360
static void udf_table_free_blocks(struct super_block *sb,
fs/udf/balloc.c
366
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
377
udf_add_free_space(sb, sbi->s_partition, count);
fs/udf/balloc.c
394
(elen >> sb->s_blocksize_bits)) == start)) {
fs/udf/balloc.c
396
(count << sb->s_blocksize_bits)) {
fs/udf/balloc.c
398
sb->s_blocksize_bits);
fs/udf/balloc.c
402
(0x40000000 - sb->s_blocksize);
fs/udf/balloc.c
406
(count << sb->s_blocksize_bits));
fs/udf/balloc.c
41
loc.partitionReferenceNum = UDF_SB(sb)->s_partition;
fs/udf/balloc.c
413
(count << sb->s_blocksize_bits)) {
fs/udf/balloc.c
415
sb->s_blocksize_bits);
fs/udf/balloc.c
420
(0x40000000 - sb->s_blocksize);
fs/udf/balloc.c
425
(count << sb->s_blocksize_bits));
fs/udf/balloc.c
43
bh = sb_bread(sb, udf_get_lb_pblock(sb, &loc, block));
fs/udf/balloc.c
461
(count << sb->s_blocksize_bits);
fs/udf/balloc.c
470
if (epos.offset + (2 * adsize) > sb->s_blocksize) {
fs/udf/balloc.c
476
elen -= sb->s_blocksize;
fs/udf/balloc.c
49
max_bits = sb->s_blocksize * 8;
fs/udf/balloc.c
492
static int udf_table_prealloc_blocks(struct super_block *sb,
fs/udf/balloc.c
496
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
535
alloc_count = (elen >> sb->s_blocksize_bits);
fs/udf/balloc.c
539
elen -= (alloc_count << sb->s_blocksize_bits);
fs/udf/balloc.c
552
udf_add_free_space(sb, partition, -alloc_count);
fs/udf/balloc.c
557
static udf_pblk_t udf_table_new_block(struct super_block *sb,
fs/udf/balloc.c
561
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/balloc.c
59
(bitmap->s_nr_groups >> (sb->s_blocksize_bits + 3)) + 2)
fs/udf/balloc.c
600
(elen >> sb->s_blocksize_bits))
fs/udf/balloc.c
604
(elen >> sb->s_blocksize_bits);
fs/udf/balloc.c
640
goal_elen -= sb->s_blocksize;
fs/udf/balloc.c
648
udf_add_free_space(sb, partition, -1);
fs/udf/balloc.c
655
void udf_free_blocks(struct super_block *sb, struct inode *inode,
fs/udf/balloc.c
660
struct udf_part_map *map = &UDF_SB(sb)->s_partmaps[partition];
fs/udf/balloc.c
674
udf_bitmap_free_blocks(sb, map->s_uspace.s_bitmap,
fs/udf/balloc.c
677
udf_table_free_blocks(sb, map->s_uspace.s_table,
fs/udf/balloc.c
683
((sector_t)count) << sb->s_blocksize_bits);
fs/udf/balloc.c
687
inline int udf_prealloc_blocks(struct super_block *sb,
fs/udf/balloc.c
692
struct udf_part_map *map = &UDF_SB(sb)->s_partmaps[partition];
fs/udf/balloc.c
696
allocated = udf_bitmap_prealloc_blocks(sb,
fs/udf/balloc.c
701
allocated = udf_table_prealloc_blocks(sb,
fs/udf/balloc.c
709
inode_add_bytes(inode, allocated << sb->s_blocksize_bits);
fs/udf/balloc.c
713
inline udf_pblk_t udf_new_block(struct super_block *sb,
fs/udf/balloc.c
717
struct udf_part_map *map = &UDF_SB(sb)->s_partmaps[partition];
fs/udf/balloc.c
721
block = udf_bitmap_new_block(sb,
fs/udf/balloc.c
725
block = udf_table_new_block(sb,
fs/udf/balloc.c
733
inode_add_bytes(inode, sb->s_blocksize);
fs/udf/balloc.c
77
static int load_block_bitmap(struct super_block *sb,
fs/udf/balloc.c
99
retval = read_block_bitmap(sb, bitmap, block_group, block_group);
fs/udf/dir.c
107
flen = udf_get_filename(sb, iter.name,
fs/udf/dir.c
113
iblock = udf_get_lb_pblock(sb, &tloc, 0);
fs/udf/dir.c
43
struct super_block *sb = dir->i_sb;
fs/udf/dir.c
92
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_UNDELETE))
fs/udf/dir.c
97
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_UNHIDE))
fs/udf/file.c
219
struct super_block *sb = inode->i_sb;
fs/udf/file.c
227
UDF_QUERY_FLAG(sb, UDF_FLAG_UID_SET) &&
fs/udf/file.c
228
!uid_eq(attr->ia_uid, UDF_SB(sb)->s_uid))
fs/udf/file.c
231
UDF_QUERY_FLAG(sb, UDF_FLAG_GID_SET) &&
fs/udf/file.c
232
!gid_eq(attr->ia_gid, UDF_SB(sb)->s_gid))
fs/udf/ialloc.c
32
struct super_block *sb = dir->i_sb;
fs/udf/ialloc.c
33
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/ialloc.c
41
inode = new_inode(sb);
fs/udf/ialloc.c
76
iinfo->i_unique = lvid_get_unique_id(sb);
fs/udf/ialloc.c
80
if (UDF_QUERY_FLAG(sb, UDF_FLAG_UID_SET))
fs/udf/ialloc.c
82
if (UDF_QUERY_FLAG(sb, UDF_FLAG_GID_SET))
fs/udf/ialloc.c
88
inode->i_ino = udf_get_lb_pblock(sb, &iinfo->i_location, 0);
fs/udf/inode.c
1952
struct inode *__udf_iget(struct super_block *sb, struct kernel_lb_addr *ino,
fs/udf/inode.c
1955
unsigned long block = udf_get_lb_pblock(sb, ino, 0);
fs/udf/inode.c
1956
struct inode *inode = iget_locked(sb, block);
fs/udf/inode.c
1984
struct super_block *sb = inode->i_sb;
fs/udf/inode.c
2002
bh = sb_getblk(sb, udf_get_lb_pblock(sb, &neloc, 0));
fs/udf/inode.c
2006
memset(bh->b_data, 0x00, sb->s_blocksize);
fs/udf/inode.c
2012
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_STRICT)) {
fs/udf/inode.c
2017
if (UDF_SB(sb)->s_udfrev >= 0x0200)
fs/udf/inode.c
2032
if (epos->offset + adsize > sb->s_blocksize) {
fs/udf/inode.c
2045
sb->s_blocksize | EXT_NEXT_EXTENT_ALLOCDESCS, 0);
fs/udf/inode.c
2048
sb->s_blocksize | EXT_NEXT_EXTENT_ALLOCDESCS, 0);
fs/udf/inode.c
2118
struct super_block *sb = inode->i_sb;
fs/udf/inode.c
2127
if (epos->offset + (2 * adsize) > sb->s_blocksize) {
fs/udf/inode.c
2131
new_block = udf_new_block(sb, NULL,
fs/udf/inode.c
514
struct super_block *sb = inode->i_sb;
fs/udf/inode.c
525
if (last_ext->extLength & (sb->s_blocksize - 1)) {
fs/udf/inode.c
529
sb->s_blocksize - 1) & ~(sb->s_blocksize - 1));
fs/udf/inode.c
531
(iinfo->i_lenExtents + sb->s_blocksize - 1) &
fs/udf/inode.c
532
~(sb->s_blocksize - 1);
fs/udf/inode.c
539
add = (1 << 30) - sb->s_blocksize -
fs/udf/inode.c
582
add = (1 << 30) - sb->s_blocksize;
fs/udf/inode.c
652
struct super_block *sb = inode->i_sb;
fs/udf/inode.c
653
sector_t first_block = newsize >> sb->s_blocksize_bits, offset;
fs/udf/inode.c
700
(newsize & (sb->s_blocksize - 1));
fs/udf/lowlevel.c
24
unsigned int udf_get_last_session(struct super_block *sb)
fs/udf/lowlevel.c
26
struct cdrom_device_info *cdi = disk_to_cdi(sb->s_bdev->bd_disk);
fs/udf/lowlevel.c
44
udf_pblk_t udf_get_last_block(struct super_block *sb)
fs/udf/lowlevel.c
46
struct cdrom_device_info *cdi = disk_to_cdi(sb->s_bdev->bd_disk);
fs/udf/lowlevel.c
54
if (sb_bdev_nr_blocks(sb) > ~(udf_pblk_t)0)
fs/udf/lowlevel.c
56
lblock = sb_bdev_nr_blocks(sb);
fs/udf/misc.c
188
struct buffer_head *udf_read_tagged(struct super_block *sb, uint32_t block,
fs/udf/misc.c
199
bh = sb_bread(sb, block);
fs/udf/misc.c
201
udf_err(sb, "read failed, block=%u, location=%u\n",
fs/udf/misc.c
219
udf_err(sb, "tag checksum failed, block %u: 0x%02x != 0x%02x\n",
fs/udf/misc.c
227
udf_err(sb, "tag version 0x%04x != 0x0002 || 0x0003, block %u\n",
fs/udf/misc.c
233
if (le16_to_cpu(tag_p->descCRCLength) + sizeof(struct tag) > sb->s_blocksize ||
fs/udf/misc.c
247
struct buffer_head *udf_read_ptagged(struct super_block *sb,
fs/udf/misc.c
251
return udf_read_tagged(sb, udf_get_lb_pblock(sb, loc, offset),
fs/udf/namei.c
331
static void udf_add_fid_counter(struct super_block *sb, bool dir, int val)
fs/udf/namei.c
333
struct logicalVolIntegrityDescImpUse *lvidiu = udf_sb_lvidiu(sb);
fs/udf/namei.c
337
mutex_lock(&UDF_SB(sb)->s_alloc_mutex);
fs/udf/namei.c
342
udf_updated_lvid(sb);
fs/udf/namei.c
343
mutex_unlock(&UDF_SB(sb)->s_alloc_mutex);
fs/udf/namei.c
57
struct super_block *sb = dir->i_sb;
fs/udf/namei.c
584
struct super_block *sb = dir->i_sb;
fs/udf/namei.c
608
block = udf_new_block(sb, inode,
fs/udf/namei.c
619
bsize = sb->s_blocksize;
fs/udf/namei.c
624
udf_free_blocks(sb, inode, &eloc, 0, 1);
fs/udf/namei.c
628
block = udf_get_pblock(sb, block,
fs/udf/namei.c
631
epos.bh = sb_getblk(sb, block);
fs/udf/namei.c
634
udf_free_blocks(sb, inode, &eloc, 0, 1);
fs/udf/namei.c
646
eoffset = sb->s_blocksize - udf_ext0_offset(inode);
fs/udf/namei.c
686
namelen = udf_put_filename(sb, compstart,
fs/udf/namei.c
70
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_UNDELETE))
fs/udf/namei.c
75
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_UNHIDE))
fs/udf/namei.c
86
flen = udf_get_filename(sb, iter->name,
fs/udf/namei.c
931
static struct dentry *udf_nfs_get_inode(struct super_block *sb, u32 block,
fs/udf/namei.c
942
inode = udf_iget(sb, &loc);
fs/udf/namei.c
954
static struct dentry *udf_fh_to_dentry(struct super_block *sb,
fs/udf/namei.c
962
return udf_nfs_get_inode(sb, fid->udf.block, fid->udf.partref,
fs/udf/namei.c
966
static struct dentry *udf_fh_to_parent(struct super_block *sb,
fs/udf/namei.c
972
return udf_nfs_get_inode(sb, fid->udf.parent_block,
fs/udf/partition.c
101
inline uint32_t udf_get_pblock_virt20(struct super_block *sb, uint32_t block,
fs/udf/partition.c
104
return udf_get_pblock_virt15(sb, block, partition, offset);
fs/udf/partition.c
107
uint32_t udf_get_pblock_spar15(struct super_block *sb, uint32_t block,
fs/udf/partition.c
112
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/partition.c
147
int udf_relocate_blocks(struct super_block *sb, long old_block, long *new_block)
fs/udf/partition.c
154
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/partition.c
25
uint32_t udf_get_pblock(struct super_block *sb, uint32_t block,
fs/udf/partition.c
278
struct super_block *sb = inode->i_sb;
fs/udf/partition.c
28
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/partition.c
292
map = &UDF_SB(sb)->s_partmaps[partition];
fs/udf/partition.c
294
phyblock = udf_get_pblock(sb, eloc.logicalBlockNum,
fs/udf/partition.c
303
uint32_t udf_get_pblock_meta25(struct super_block *sb, uint32_t block,
fs/udf/partition.c
306
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/partition.c
323
udf_warn(sb, "error reading from METADATA, trying to read from MIRROR\n");
fs/udf/partition.c
325
mdata->s_mirror_fe = udf_find_metadata_inode_efe(sb,
fs/udf/partition.c
37
return map->s_partition_func(sb, block, partition, offset);
fs/udf/partition.c
42
uint32_t udf_get_pblock_virt15(struct super_block *sb, uint32_t block,
fs/udf/partition.c
49
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/partition.c
69
index = (sb->s_blocksize - vdata->s_start_offset) / sizeof(uint32_t);
fs/udf/partition.c
72
newblock = 1 + (block / (sb->s_blocksize / sizeof(uint32_t)));
fs/udf/partition.c
73
index = block % (sb->s_blocksize / sizeof(uint32_t));
fs/udf/partition.c
82
sb, block, partition);
fs/udf/partition.c
96
return udf_get_pblock(sb, loc,
fs/udf/super.c
1001
udf_err(sb, "Both metadata and mirror metadata inode efe can not found\n");
fs/udf/super.c
1021
fe = udf_iget_special(sb, &addr);
fs/udf/super.c
1023
if (sb_rdonly(sb))
fs/udf/super.c
1024
udf_warn(sb, "bitmap inode efe not found but it's ok since the disc is mounted read-only\n");
fs/udf/super.c
1026
udf_err(sb, "bitmap inode efe not found and attempted read-write mount\n");
fs/udf/super.c
1037
int udf_compute_nr_groups(struct super_block *sb, u32 partition)
fs/udf/super.c
1039
struct udf_part_map *map = &UDF_SB(sb)->s_partmaps[partition];
fs/udf/super.c
1042
sb->s_blocksize * 8);
fs/udf/super.c
1045
static struct udf_bitmap *udf_sb_alloc_bitmap(struct super_block *sb, u32 index)
fs/udf/super.c
1048
int nr_groups = udf_compute_nr_groups(sb, index);
fs/udf/super.c
1058
static int check_partition_desc(struct super_block *sb,
fs/udf/super.c
109
struct logicalVolIntegrityDescImpUse *udf_sb_lvidiu(struct super_block *sb)
fs/udf/super.c
1101
if (!sb_rdonly(sb))
fs/udf/super.c
1103
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
1107
static int udf_fill_partdesc_info(struct super_block *sb,
fs/udf/super.c
1111
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1122
udf_err(sb, "Partition %d has invalid location %u + %u\n",
fs/udf/super.c
1140
err = check_partition_desc(sb, p, map);
fs/udf/super.c
1149
if (UDF_QUERY_FLAG(sb, UDF_FLAG_RW_INCOMPAT))
fs/udf/super.c
115
if (!UDF_SB(sb)->s_lvid_bh)
fs/udf/super.c
1161
inode = udf_iget_special(sb, &loc);
fs/udf/super.c
117
lvid = (struct logicalVolIntegrityDesc *)UDF_SB(sb)->s_lvid_bh->b_data;
fs/udf/super.c
1174
struct udf_bitmap *bitmap = udf_sb_alloc_bitmap(sb, p_index);
fs/udf/super.c
1185
udf_err(sb, "Partition %d is too long (%u)\n", p_index,
fs/udf/super.c
1196
static void udf_find_vat_block(struct super_block *sb, int p_index,
fs/udf/super.c
1199
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1214
inode = udf_iget_special(sb, &ino);
fs/udf/super.c
1222
static int udf_load_vat(struct super_block *sb, int p_index, int type1_index)
fs/udf/super.c
1224
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1229
sector_t blocks = sb_bdev_nr_blocks(sb);
fs/udf/super.c
1231
udf_find_vat_block(sb, p_index, type1_index, sbi->s_last_block);
fs/udf/super.c
1237
udf_find_vat_block(sb, p_index, type1_index, blocks - 1);
fs/udf/super.c
1280
static int udf_load_partdesc(struct super_block *sb, sector_t block)
fs/udf/super.c
1285
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1291
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1320
ret = udf_fill_partdesc_info(sb, p, i);
fs/udf/super.c
1345
ret = udf_fill_partdesc_info(sb, p, i);
fs/udf/super.c
1350
ret = udf_load_metadata_files(sb, i, type1_idx);
fs/udf/super.c
1352
udf_err(sb, "error loading MetaData partition map %d\n",
fs/udf/super.c
1362
if (!sb_rdonly(sb)) {
fs/udf/super.c
1366
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
1367
ret = udf_load_vat(sb, i, type1_idx);
fs/udf/super.c
1378
static int udf_load_sparable_map(struct super_block *sb,
fs/udf/super.c
1392
udf_err(sb, "error loading logical volume descriptor: "
fs/udf/super.c
1398
udf_err(sb, "error loading logical volume descriptor: "
fs/udf/super.c
1403
if (le32_to_cpu(spm->sizeSparingTable) > sb->s_blocksize) {
fs/udf/super.c
1404
udf_err(sb, "error loading logical volume descriptor: "
fs/udf/super.c
1412
bh = udf_read_tagged(sb, loc, loc, &ident);
fs/udf/super.c
1421
sb->s_blocksize) {
fs/udf/super.c
1432
static int udf_load_logicalvol(struct super_block *sb, sector_t block,
fs/udf/super.c
1438
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1445
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1451
if (table_len > sb->s_blocksize - sizeof(*lvd)) {
fs/udf/super.c
1452
udf_err(sb, "error loading logical volume descriptor: "
fs/udf/super.c
1454
sb->s_blocksize - sizeof(*lvd));
fs/udf/super.c
1459
ret = udf_verify_domain_identifier(sb, &lvd->domainIdent,
fs/udf/super.c
1466
udf_err(sb, "error loading logical volume descriptor: "
fs/udf/super.c
1472
ret = udf_sb_alloc_partition_maps(sb, part_map_count);
fs/udf/super.c
150
static struct inode *udf_alloc_inode(struct super_block *sb)
fs/udf/super.c
1512
ret = udf_load_sparable_map(sb, map,
fs/udf/super.c
153
ei = alloc_inode_sb(sb, udf_inode_cachep, GFP_KERNEL);
fs/udf/super.c
1579
udf_load_logicalvolint(sb, leea_to_cpu(lvd->integritySeqExt));
fs/udf/super.c
1584
if (sb_rdonly(sb)) {
fs/udf/super.c
1585
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
1587
udf_warn(sb, "Damaged or missing LVID, forcing "
fs/udf/super.c
1597
static bool udf_lvid_valid(struct super_block *sb,
fs/udf/super.c
1604
if (parts >= sb->s_blocksize || impuselen >= sb->s_blocksize ||
fs/udf/super.c
1606
2 * parts * sizeof(u32) > sb->s_blocksize)
fs/udf/super.c
1614
static void udf_load_logicalvolint(struct super_block *sb, struct kernel_extent_ad loc)
fs/udf/super.c
1618
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1625
(bh = udf_read_tagged(sb, loc.extLocation,
fs/udf/super.c
1635
loc.extLength -= sb->s_blocksize;
fs/udf/super.c
1643
if (udf_lvid_valid(sb, lvid)) {
fs/udf/super.c
1647
udf_warn(sb, "Corrupted LVID (parts=%u, impuselen=%u), "
fs/udf/super.c
1659
udf_warn(sb, "Too many LVID indirections (max %u), ignoring.\n",
fs/udf/super.c
1740
struct super_block *sb,
fs/udf/super.c
1769
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1779
udf_err(sb, "too many Volume Descriptor "
fs/udf/super.c
1791
sb->s_blocksize_bits;
fs/udf/super.c
1826
udf_err(sb, "Primary Volume Descriptor not found!\n");
fs/udf/super.c
1830
ret = udf_load_pvoldesc(sb, data.vds[VDS_POS_PRIMARY_VOL_DESC].block);
fs/udf/super.c
1835
ret = udf_load_logicalvol(sb,
fs/udf/super.c
1844
ret = udf_load_partdesc(sb, data.part_descs_loc[i].rec.block);
fs/udf/super.c
1859
static int udf_load_sequence(struct super_block *sb, struct buffer_head *bh,
fs/udf/super.c
1871
main_e = main_e >> sb->s_blocksize_bits;
fs/udf/super.c
1877
reserve_e = reserve_e >> sb->s_blocksize_bits;
fs/udf/super.c
1882
ret = udf_process_sequence(sb, main_s, main_e, fileset);
fs/udf/super.c
1885
udf_sb_free_partitions(sb);
fs/udf/super.c
1886
ret = udf_process_sequence(sb, reserve_s, reserve_e, fileset);
fs/udf/super.c
1888
udf_sb_free_partitions(sb);
fs/udf/super.c
1903
static int udf_check_anchor_block(struct super_block *sb, sector_t block,
fs/udf/super.c
1910
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
1917
ret = udf_load_sequence(sb, bh, fileset);
fs/udf/super.c
1928
static int udf_scan_anchors(struct super_block *sb, udf_pblk_t *lastblock,
fs/udf/super.c
1933
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
1939
ret = udf_check_anchor_block(sb, sbi->s_anchor, fileset);
fs/udf/super.c
1950
ret = udf_check_anchor_block(sb, sbi->s_session + 256, fileset);
fs/udf/super.c
1969
if (last[i] >= sb_bdev_nr_blocks(sb))
fs/udf/super.c
1971
ret = udf_check_anchor_block(sb, last[i], fileset);
fs/udf/super.c
1979
ret = udf_check_anchor_block(sb, last[i] - 256, fileset);
fs/udf/super.c
1988
return udf_check_anchor_block(sb, sbi->s_session + 512, fileset);
fs/udf/super.c
1998
static int udf_load_vrs(struct super_block *sb, struct udf_options *uopt,
fs/udf/super.c
2001
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2005
if (!sb_set_blocksize(sb, uopt->blocksize)) {
fs/udf/super.c
2007
udf_warn(sb, "Bad block size\n");
fs/udf/super.c
2011
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_NOVRS)) {
fs/udf/super.c
2013
nsr = udf_check_vsd(sb);
fs/udf/super.c
2016
udf_warn(sb, "No VRS found\n");
fs/udf/super.c
2024
sbi->s_last_block = udf_get_last_block(sb);
fs/udf/super.c
2031
ret = udf_scan_anchors(sb, &sbi->s_last_block, fileset);
fs/udf/super.c
2034
udf_warn(sb, "No anchor found\n");
fs/udf/super.c
2052
static void udf_open_lvid(struct super_block *sb)
fs/udf/super.c
2054
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2062
lvidiu = udf_sb_lvidiu(sb);
fs/udf/super.c
2072
UDF_SET_FLAG(sb, UDF_FLAG_INCONSISTENT);
fs/udf/super.c
2082
static void udf_close_lvid(struct super_block *sb)
fs/udf/super.c
2084
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2092
lvidiu = udf_sb_lvidiu(sb);
fs/udf/super.c
2105
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_INCONSISTENT))
fs/udf/super.c
2122
u64 lvid_get_unique_id(struct super_block *sb)
fs/udf/super.c
2125
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2143
udf_updated_lvid(sb);
fs/udf/super.c
2149
static int udf_fill_super(struct super_block *sb, struct fs_context *fc)
fs/udf/super.c
2163
sb->s_fs_info = sbi;
fs/udf/super.c
2181
sbi->s_session = udf_get_last_session(sb);
fs/udf/super.c
2188
sb->s_op = &udf_sb_ops;
fs/udf/super.c
2189
sb->s_export_op = &udf_export_ops;
fs/udf/super.c
2191
sb->s_magic = UDF_SUPER_MAGIC;
fs/udf/super.c
2192
sb->s_time_gran = 1000;
fs/udf/super.c
2195
ret = udf_load_vrs(sb, uopt, silent, &fileset);
fs/udf/super.c
2197
uopt->blocksize = bdev_logical_block_size(sb->s_bdev);
fs/udf/super.c
2199
ret = udf_load_vrs(sb, uopt, silent, &fileset);
fs/udf/super.c
2221
udf_warn(sb, "No partition found (1)\n");
fs/udf/super.c
2231
udf_sb_lvidiu(sb);
fs/udf/super.c
2242
udf_err(sb, "minUDFReadRev=%x (max is %x)\n",
fs/udf/super.c
2248
if (!sb_rdonly(sb)) {
fs/udf/super.c
2252
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
2258
UDF_SET_FLAG(sb, UDF_FLAG_USE_EXTENDED_FE);
fs/udf/super.c
2260
UDF_SET_FLAG(sb, UDF_FLAG_USE_STREAMS);
fs/udf/super.c
2264
udf_warn(sb, "No partition found (2)\n");
fs/udf/super.c
2271
if (!sb_rdonly(sb)) {
fs/udf/super.c
2275
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
2278
ret = udf_find_fileset(sb, &fileset, &rootdir);
fs/udf/super.c
2280
udf_warn(sb, "No fileset found\n");
fs/udf/super.c
2292
if (!sb_rdonly(sb)) {
fs/udf/super.c
2293
udf_open_lvid(sb);
fs/udf/super.c
2300
inode = udf_iget(sb, &rootdir);
fs/udf/super.c
2302
udf_err(sb, "Error in udf_iget, block=%u, partition=%u\n",
fs/udf/super.c
2309
sb->s_root = d_make_root(inode);
fs/udf/super.c
2310
if (!sb->s_root) {
fs/udf/super.c
2311
udf_err(sb, "Couldn't allocate root dentry\n");
fs/udf/super.c
2315
sb->s_maxbytes = UDF_MAX_FILESIZE;
fs/udf/super.c
2316
sb->s_max_links = UDF_MAX_LINKS;
fs/udf/super.c
2323
udf_close_lvid(sb);
fs/udf/super.c
2325
udf_sb_free_partitions(sb);
fs/udf/super.c
2327
sb->s_fs_info = NULL;
fs/udf/super.c
2332
void _udf_err(struct super_block *sb, const char *function,
fs/udf/super.c
2343
pr_err("error (device %s): %s: %pV", sb->s_id, function, &vaf);
fs/udf/super.c
2348
void _udf_warn(struct super_block *sb, const char *function,
fs/udf/super.c
2359
pr_warn("warning (device %s): %s: %pV", sb->s_id, function, &vaf);
fs/udf/super.c
2364
static void udf_put_super(struct super_block *sb)
fs/udf/super.c
2368
sbi = UDF_SB(sb);
fs/udf/super.c
2372
if (!sb_rdonly(sb))
fs/udf/super.c
2373
udf_close_lvid(sb);
fs/udf/super.c
2375
udf_sb_free_partitions(sb);
fs/udf/super.c
2377
kfree(sb->s_fs_info);
fs/udf/super.c
2378
sb->s_fs_info = NULL;
fs/udf/super.c
2381
static int udf_sync_fs(struct super_block *sb, int wait)
fs/udf/super.c
2383
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2407
struct super_block *sb = dentry->d_sb;
fs/udf/super.c
2408
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2410
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/udf/super.c
2412
lvidiu = udf_sb_lvidiu(sb);
fs/udf/super.c
2414
buf->f_bsize = sb->s_blocksize;
fs/udf/super.c
2416
buf->f_bfree = udf_count_free(sb);
fs/udf/super.c
242
struct super_block *sb = fc->root->d_sb;
fs/udf/super.c
243
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2432
static unsigned int udf_count_free_bitmap(struct super_block *sb,
fs/udf/super.c
2446
loc.partitionReferenceNum = UDF_SB(sb)->s_partition;
fs/udf/super.c
2447
bh = udf_read_ptagged(sb, &loc, 0, &ident);
fs/udf/super.c
2450
udf_err(sb, "udf_count_free failed\n");
fs/udf/super.c
2454
udf_err(sb, "udf_count_free failed\n");
fs/udf/super.c
2464
u32 cur_bytes = min_t(u32, bytes, sb->s_blocksize - index);
fs/udf/super.c
2470
newblock = udf_get_lb_pblock(sb, &loc, ++block);
fs/udf/super.c
2471
bh = sb_bread(sb, newblock);
fs/udf/super.c
2485
static unsigned int udf_count_free_table(struct super_block *sb,
fs/udf/super.c
2494
mutex_lock(&UDF_SB(sb)->s_alloc_mutex);
fs/udf/super.c
2503
mutex_unlock(&UDF_SB(sb)->s_alloc_mutex);
fs/udf/super.c
2508
static unsigned int udf_count_free(struct super_block *sb)
fs/udf/super.c
2511
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
2544
accum += udf_count_free_bitmap(sb,
fs/udf/super.c
2551
accum += udf_count_free_table(sb,
fs/udf/super.c
319
static int udf_sb_alloc_partition_maps(struct super_block *sb, u32 count)
fs/udf/super.c
321
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
370
static void udf_sb_free_partitions(struct super_block *sb)
fs/udf/super.c
372
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
385
struct super_block *sb = root->d_sb;
fs/udf/super.c
386
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
388
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_STRICT))
fs/udf/super.c
390
if (UDF_QUERY_FLAG(sb, UDF_FLAG_BLOCKSIZE_SET))
fs/udf/super.c
391
seq_printf(seq, ",bs=%lu", sb->s_blocksize);
fs/udf/super.c
392
if (UDF_QUERY_FLAG(sb, UDF_FLAG_UNHIDE))
fs/udf/super.c
394
if (UDF_QUERY_FLAG(sb, UDF_FLAG_UNDELETE))
fs/udf/super.c
396
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_USE_AD_IN_ICB))
fs/udf/super.c
398
if (UDF_QUERY_FLAG(sb, UDF_FLAG_USE_SHORT_AD))
fs/udf/super.c
400
if (UDF_QUERY_FLAG(sb, UDF_FLAG_UID_FORGET))
fs/udf/super.c
402
if (UDF_QUERY_FLAG(sb, UDF_FLAG_GID_FORGET))
fs/udf/super.c
404
if (UDF_QUERY_FLAG(sb, UDF_FLAG_UID_SET))
fs/udf/super.c
406
if (UDF_QUERY_FLAG(sb, UDF_FLAG_GID_SET))
fs/udf/super.c
414
if (UDF_QUERY_FLAG(sb, UDF_FLAG_SESSION_SET))
fs/udf/super.c
416
if (UDF_QUERY_FLAG(sb, UDF_FLAG_LASTBLOCK_SET))
fs/udf/super.c
653
struct super_block *sb = fc->root->d_sb;
fs/udf/super.c
654
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
658
if (!readonly && UDF_QUERY_FLAG(sb, UDF_FLAG_RW_INCOMPAT))
fs/udf/super.c
661
sync_filesystem(sb);
fs/udf/super.c
672
if (readonly == sb_rdonly(sb))
fs/udf/super.c
676
udf_close_lvid(sb);
fs/udf/super.c
678
udf_open_lvid(sb);
fs/udf/super.c
738
static int udf_check_vsd(struct super_block *sb)
fs/udf/super.c
748
sbi = UDF_SB(sb);
fs/udf/super.c
749
if (sb->s_blocksize < sizeof(struct volStructDesc))
fs/udf/super.c
752
sectorsize = sb->s_blocksize;
fs/udf/super.c
754
session_offset = (loff_t)sbi->s_session << sb->s_blocksize_bits;
fs/udf/super.c
758
(unsigned int)(sector >> sb->s_blocksize_bits),
fs/udf/super.c
759
sb->s_blocksize);
fs/udf/super.c
772
bh = sb_bread(sb, sector >> sb->s_blocksize_bits);
fs/udf/super.c
777
(sector & (sb->s_blocksize - 1)));
fs/udf/super.c
789
if (sb->s_blocksize == 4096) {
fs/udf/super.c
806
static int udf_verify_domain_identifier(struct super_block *sb,
fs/udf/super.c
812
udf_warn(sb, "Not OSTA UDF compliant %s descriptor.\n", dname);
fs/udf/super.c
816
udf_warn(sb, "Possibly not OSTA UDF compliant %s descriptor.\n",
fs/udf/super.c
823
if (!sb_rdonly(sb)) {
fs/udf/super.c
824
udf_warn(sb, "Descriptor for %s marked write protected."
fs/udf/super.c
832
if (!sb_rdonly(sb))
fs/udf/super.c
834
UDF_SET_FLAG(sb, UDF_FLAG_RW_INCOMPAT);
fs/udf/super.c
838
static int udf_load_fileset(struct super_block *sb, struct fileSetDesc *fset,
fs/udf/super.c
843
ret = udf_verify_domain_identifier(sb, &fset->domainIdent, "file set");
fs/udf/super.c
848
UDF_SB(sb)->s_serial_number = le16_to_cpu(fset->descTag.tagSerialNum);
fs/udf/super.c
855
static int udf_find_fileset(struct super_block *sb,
fs/udf/super.c
867
bh = udf_read_ptagged(sb, fileset, 0, &ident);
fs/udf/super.c
878
UDF_SB(sb)->s_partition = fileset->partitionReferenceNum;
fs/udf/super.c
879
ret = udf_load_fileset(sb, (struct fileSetDesc *)bh->b_data, root);
fs/udf/super.c
890
static int udf_load_pvoldesc(struct super_block *sb, sector_t block)
fs/udf/super.c
903
bh = udf_read_tagged(sb, block, block, &ident);
fs/udf/super.c
916
udf_disk_stamp_to_time(&UDF_SB(sb)->s_record_time,
fs/udf/super.c
923
ret = udf_dstrCS0toChar(sb, outstr, 31, pvoldesc->volIdent, 32);
fs/udf/super.c
925
strscpy_pad(UDF_SB(sb)->s_volume_ident, "InvalidName");
fs/udf/super.c
929
strscpy_pad(UDF_SB(sb)->s_volume_ident, outstr);
fs/udf/super.c
931
udf_debug("volIdent[] = '%s'\n", UDF_SB(sb)->s_volume_ident);
fs/udf/super.c
933
ret = udf_dstrCS0toChar(sb, outstr, 127, pvoldesc->volSetIdent, 128);
fs/udf/super.c
94
static int udf_fill_super(struct super_block *sb, struct fs_context *fc);
fs/udf/super.c
949
struct inode *udf_find_metadata_inode_efe(struct super_block *sb,
fs/udf/super.c
958
metadata_fe = udf_iget_special(sb, &addr);
fs/udf/super.c
961
udf_warn(sb, "metadata inode efe not found\n");
fs/udf/super.c
965
udf_warn(sb, "metadata inode efe does not have short allocation descriptors!\n");
fs/udf/super.c
973
static int udf_load_metadata_files(struct super_block *sb, int partition,
fs/udf/super.c
976
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/super.c
990
fe = udf_find_metadata_inode_efe(sb, mdata->s_meta_file_loc,
fs/udf/super.c
997
fe = udf_find_metadata_inode_efe(sb, mdata->s_mirror_file_loc,
fs/udf/symlink.c
28
static int udf_pc_to_char(struct super_block *sb, unsigned char *from,
fs/udf/symlink.c
78
comp_len = udf_get_filename(sb, pc->componentIdent,
fs/udf/truncate.c
178
struct super_block *sb = inode->i_sb;
fs/udf/truncate.c
179
struct udf_sb_info *sbi = UDF_SB(sb);
fs/udf/truncate.c
185
if (!UDF_QUERY_FLAG(sb, UDF_FLAG_STRICT) || sbi->s_udfrev >= 0x0201)
fs/udf/truncate.c
203
struct super_block *sb = inode->i_sb;
fs/udf/truncate.c
204
sector_t first_block = inode->i_size >> sb->s_blocksize_bits, offset;
fs/udf/truncate.c
220
byte_offset = (offset << sb->s_blocksize_bits) +
fs/udf/truncate.c
221
(inode->i_size & (sb->s_blocksize - 1));
fs/udf/truncate.c
248
udf_free_blocks(sb, NULL, &epos.block,
fs/udf/truncate.c
259
epos.bh = sb_bread(sb,
fs/udf/truncate.c
260
udf_get_lb_pblock(sb, &eloc, 0));
fs/udf/truncate.c
266
(elen + sb->s_blocksize - 1) >>
fs/udf/truncate.c
267
sb->s_blocksize_bits;
fs/udf/truncate.c
283
udf_free_blocks(sb, NULL, &epos.block, 0, indirect_ext_len);
fs/udf/udf_sb.h
159
static inline struct udf_sb_info *UDF_SB(struct super_block *sb)
fs/udf/udf_sb.h
161
return sb->s_fs_info;
fs/udf/udf_sb.h
164
struct logicalVolIntegrityDescImpUse *udf_sb_lvidiu(struct super_block *sb);
fs/udf/udf_sb.h
166
int udf_compute_nr_groups(struct super_block *sb, u32 partition);
fs/udf/udf_sb.h
168
static inline int UDF_QUERY_FLAG(struct super_block *sb, int flag)
fs/udf/udf_sb.h
170
return test_bit(flag, &UDF_SB(sb)->s_flags);
fs/udf/udf_sb.h
173
static inline void UDF_SET_FLAG(struct super_block *sb, int flag)
fs/udf/udf_sb.h
175
set_bit(flag, &UDF_SB(sb)->s_flags);
fs/udf/udf_sb.h
178
static inline void UDF_CLEAR_FLAG(struct super_block *sb, int flag)
fs/udf/udf_sb.h
180
clear_bit(flag, &UDF_SB(sb)->s_flags);
fs/udf/udfdecl.h
116
static inline void udf_updated_lvid(struct super_block *sb)
fs/udf/udfdecl.h
118
struct buffer_head *bh = UDF_SB(sb)->s_lvid_bh;
fs/udf/udfdecl.h
124
UDF_SB(sb)->s_lvid_dirty = 1;
fs/udf/udfdecl.h
126
extern u64 lvid_get_unique_id(struct super_block *sb);
fs/udf/udfdecl.h
127
struct inode *udf_find_metadata_inode_efe(struct super_block *sb,
fs/udf/udfdecl.h
144
static inline struct inode *udf_iget_special(struct super_block *sb,
fs/udf/udfdecl.h
147
return __udf_iget(sb, ino, true);
fs/udf/udfdecl.h
149
static inline struct inode *udf_iget(struct super_block *sb,
fs/udf/udfdecl.h
152
return __udf_iget(sb, ino, false);
fs/udf/udfdecl.h
21
extern __printf(3, 4) void _udf_err(struct super_block *sb,
fs/udf/udfdecl.h
212
udf_get_lb_pblock(struct super_block *sb, struct kernel_lb_addr *loc,
fs/udf/udfdecl.h
215
return udf_get_pblock(sb, loc->logicalBlockNum,
fs/udf/udfdecl.h
23
#define udf_err(sb, fmt, ...) \
fs/udf/udfdecl.h
24
_udf_err(sb, __func__, fmt, ##__VA_ARGS__)
fs/udf/udfdecl.h
241
extern udf_pblk_t udf_new_block(struct super_block *sb, struct inode *inode,
fs/udf/udfdecl.h
26
extern __printf(3, 4) void _udf_warn(struct super_block *sb,
fs/udf/udfdecl.h
28
#define udf_warn(sb, fmt, ...) \
fs/udf/udfdecl.h
29
_udf_warn(sb, __func__, fmt, ##__VA_ARGS__)
fs/udf/unicode.c
148
static int udf_name_from_CS0(struct super_block *sb,
fs/udf/unicode.c
177
if (UDF_SB(sb)->s_nls_map)
fs/udf/unicode.c
178
conv_f = UDF_SB(sb)->s_nls_map->uni2char;
fs/udf/unicode.c
272
static int udf_name_to_CS0(struct super_block *sb,
fs/udf/unicode.c
285
if (UDF_SB(sb)->s_nls_map)
fs/udf/unicode.c
286
conv_f = UDF_SB(sb)->s_nls_map->char2uni;
fs/udf/unicode.c
356
int udf_dstrCS0toChar(struct super_block *sb, uint8_t *utf_o, int o_len,
fs/udf/unicode.c
373
return udf_name_from_CS0(sb, utf_o, o_len, ocu_i, s_len, 0);
fs/udf/unicode.c
376
int udf_get_filename(struct super_block *sb, const uint8_t *sname, int slen,
fs/udf/unicode.c
387
ret = udf_name_from_CS0(sb, dname, dlen, sname, slen, 1);
fs/udf/unicode.c
394
int udf_put_filename(struct super_block *sb, const uint8_t *sname, int slen,
fs/udf/unicode.c
397
return udf_name_to_CS0(sb, dname, dlen, sname, slen);
fs/ufs/balloc.c
100
ufs_fragacct (sb, blkmap, ucg->cg_frsum, -1);
fs/ufs/balloc.c
105
ufs_error (sb, "ufs_free_fragments",
fs/ufs/balloc.c
110
fs32_add(sb, &ucg->cg_cs.cs_nffree, count);
fs/ufs/balloc.c
112
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, count);
fs/ufs/balloc.c
114
ufs_fragacct(sb, blkmap, ucg->cg_frsum, 1);
fs/ufs/balloc.c
120
fs32_sub(sb, &ucg->cg_cs.cs_nffree, uspi->s_fpb);
fs/ufs/balloc.c
122
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, uspi->s_fpb);
fs/ufs/balloc.c
123
adjust_free_blocks(sb, ucg, ucpi, bbase, 1);
fs/ufs/balloc.c
128
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/balloc.c
130
ufs_mark_sb_dirty(sb);
fs/ufs/balloc.c
132
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
137
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
147
struct super_block * sb;
fs/ufs/balloc.c
153
sb = inode->i_sb;
fs/ufs/balloc.c
154
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
160
ufs_error (sb, "ufs_free_blocks", "internal error, "
fs/ufs/balloc.c
166
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
173
ufs_panic (sb, "ufs_free_blocks", "freeing blocks are outside device");
fs/ufs/balloc.c
183
ucpi = ufs_load_cylinder (sb, cgno);
fs/ufs/balloc.c
187
if (!ufs_cg_chkmagic(sb, ucg)) {
fs/ufs/balloc.c
188
ufs_panic (sb, "ufs_free_blocks", "internal error, bad magic number on cg %u", cgno);
fs/ufs/balloc.c
194
ufs_error(sb, "ufs_free_blocks", "freeing free fragment");
fs/ufs/balloc.c
198
adjust_free_blocks(sb, ucg, ucpi, i, 1);
fs/ufs/balloc.c
203
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/balloc.c
212
ufs_mark_sb_dirty(sb);
fs/ufs/balloc.c
213
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
218
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
337
struct super_block * sb;
fs/ufs/balloc.c
347
sb = inode->i_sb;
fs/ufs/balloc.c
348
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
352
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
353
tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/balloc.c
356
ufs_warning(sb, "ufs_new_fragments", "internal warning"
fs/ufs/balloc.c
36
static void adjust_free_blocks(struct super_block *sb,
fs/ufs/balloc.c
369
ufs_error(sb, "ufs_new_fragments", "internal error, "
fs/ufs/balloc.c
373
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
378
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
385
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
395
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
41
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
418
ufs_cpu_to_data_ptr(sb, p, result);
fs/ufs/balloc.c
423
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
43
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
fs/ufs/balloc.c
44
ufs_clusteracct(sb, ucpi, fragment, delta);
fs/ufs/balloc.c
440
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
448
if (fs32_to_cpu(sb, usb1->fs_optim) == UFS_OPTSPACE) {
fs/ufs/balloc.c
451
usb1->fs_optim = cpu_to_fs32(sb, UFS_OPTTIME);
fs/ufs/balloc.c
455
usb1->fs_optim = cpu_to_fs32(sb, UFS_OPTSPACE);
fs/ufs/balloc.c
46
fs32_add(sb, &ucg->cg_cs.cs_nbfree, delta);
fs/ufs/balloc.c
461
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
467
ufs_cpu_to_data_ptr(sb, p, result);
fs/ufs/balloc.c
478
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
48
fs32_add(sb, &UFS_SB(sb)->fs_cs(ucpi->c_cgx).cs_nbfree, delta);
fs/ufs/balloc.c
500
struct super_block * sb;
fs/ufs/balloc.c
509
sb = inode->i_sb;
fs/ufs/balloc.c
510
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
514
if (fs32_to_cpu(sb, UFS_SB(sb)->fs_cs(cgno).cs_nffree) < count)
fs/ufs/balloc.c
518
ucpi = ufs_load_cylinder (sb, cgno);
fs/ufs/balloc.c
522
if (!ufs_cg_chkmagic(sb, ucg)) {
fs/ufs/balloc.c
523
ufs_panic (sb, "ufs_add_fragments",
fs/ufs/balloc.c
53
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
fs/ufs/balloc.c
539
ucg->cg_time = ufs_get_seconds(sb);
fs/ufs/balloc.c
544
if (!fs32_to_cpu(sb, ucg->cg_frsum[fragsize]))
fs/ufs/balloc.c
545
ufs_panic (sb, "ufs_add_fragments",
fs/ufs/balloc.c
547
fs32_sub(sb, &ucg->cg_frsum[fragsize], 1);
fs/ufs/balloc.c
549
fs32_add(sb, &ucg->cg_frsum[fragsize - count], 1);
fs/ufs/balloc.c
55
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), delta);
fs/ufs/balloc.c
553
fs32_sub(sb, &ucg->cg_cs.cs_nffree, count);
fs/ufs/balloc.c
554
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, count);
fs/ufs/balloc.c
559
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/balloc.c
561
ufs_mark_sb_dirty(sb);
fs/ufs/balloc.c
569
ucg = (struct ufs_cylinder_group *) UFS_SB(sb)->s_ucg[cgno]->b_data; \
fs/ufs/balloc.c
570
if (fs32_to_cpu(sb, ucg->cg_cs.cs_nbfree)) \
fs/ufs/balloc.c
573
if (fs32_to_cpu(sb, ucg->cg_frsum[k])) \
fs/ufs/balloc.c
579
struct super_block * sb;
fs/ufs/balloc.c
589
sb = inode->i_sb;
fs/ufs/balloc.c
590
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
624
ucpi = ufs_load_cylinder (sb, cgno);
fs/ufs/balloc.c
628
if (!ufs_cg_chkmagic(sb, ucg))
fs/ufs/balloc.c
629
ufs_panic (sb, "ufs_alloc_fragments",
fs/ufs/balloc.c
631
ucg->cg_time = ufs_get_seconds(sb);
fs/ufs/balloc.c
64
struct super_block * sb;
fs/ufs/balloc.c
641
if (fs32_to_cpu(sb, ucg->cg_frsum[allocsize]) != 0)
fs/ufs/balloc.c
654
fs32_add(sb, &ucg->cg_cs.cs_nffree, i);
fs/ufs/balloc.c
656
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, i);
fs/ufs/balloc.c
657
fs32_add(sb, &ucg->cg_frsum[i], 1);
fs/ufs/balloc.c
661
result = ufs_bitmap_search (sb, ucpi, goal, allocsize);
fs/ufs/balloc.c
669
fs32_sub(sb, &ucg->cg_cs.cs_nffree, count);
fs/ufs/balloc.c
671
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, count);
fs/ufs/balloc.c
672
fs32_sub(sb, &ucg->cg_frsum[allocsize], 1);
fs/ufs/balloc.c
675
fs32_add(sb, &ucg->cg_frsum[allocsize - count], 1);
fs/ufs/balloc.c
680
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/balloc.c
682
ufs_mark_sb_dirty(sb);
fs/ufs/balloc.c
693
struct super_block * sb;
fs/ufs/balloc.c
70
sb = inode->i_sb;
fs/ufs/balloc.c
700
sb = inode->i_sb;
fs/ufs/balloc.c
701
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
71
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
720
result = ufs_bitmap_search (sb, ucpi, goal, uspi->s_fpb);
fs/ufs/balloc.c
728
adjust_free_blocks(sb, ucg, ucpi, result, -1);
fs/ufs/balloc.c
77
ufs_error (sb, "ufs_free_fragments", "internal error");
fs/ufs/balloc.c
770
static u64 ufs_bitmap_search(struct super_block *sb,
fs/ufs/balloc.c
784
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
79
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/balloc.c
808
ufs_error(sb, "ufs_bitmap_search",
fs/ufs/balloc.c
84
ufs_panic (sb, "ufs_free_fragments", "freeing blocks are outside device");
fs/ufs/balloc.c
840
ufs_error(sb, "ufs_bitmap_search", "block not in map on cg %u\n",
fs/ufs/balloc.c
846
static void ufs_clusteracct(struct super_block * sb,
fs/ufs/balloc.c
849
struct ufs_sb_private_info * uspi = UFS_SB(sb)->s_uspi;
fs/ufs/balloc.c
88
ucpi = ufs_load_cylinder (sb, cgno);
fs/ufs/balloc.c
892
fs32_add(sb, (__fs32*)ubh_get_addr(UCPI_UBH(ucpi), ucpi->c_clustersumoff + (i << 2)), cnt);
fs/ufs/balloc.c
894
fs32_sub(sb, (__fs32*)ubh_get_addr(UCPI_UBH(ucpi), ucpi->c_clustersumoff + (back << 2)), cnt);
fs/ufs/balloc.c
896
fs32_sub(sb, (__fs32*)ubh_get_addr(UCPI_UBH(ucpi), ucpi->c_clustersumoff + (forw << 2)), cnt);
fs/ufs/balloc.c
92
if (!ufs_cg_chkmagic(sb, ucg)) {
fs/ufs/balloc.c
93
ufs_panic (sb, "ufs_free_fragments", "internal error, bad magic number on cg %u", cgno);
fs/ufs/cylinder.c
105
ufs_panic (sb, "ufs_put_cylinder", "internal error");
fs/ufs/cylinder.c
112
ucg->cg_rotor = cpu_to_fs32(sb, ucpi->c_rotor);
fs/ufs/cylinder.c
113
ucg->cg_frotor = cpu_to_fs32(sb, ucpi->c_frotor);
fs/ufs/cylinder.c
114
ucg->cg_irotor = cpu_to_fs32(sb, ucpi->c_irotor);
fs/ufs/cylinder.c
131
struct super_block * sb, unsigned cgno)
fs/ufs/cylinder.c
133
struct ufs_sb_info * sbi = UFS_SB(sb);
fs/ufs/cylinder.c
142
ufs_panic (sb, "ufs_load_cylinder", "internal error, high number of cg");
fs/ufs/cylinder.c
158
ufs_panic (sb, "ufs_load_cylinder", "internal error, wrong number of cg in cache");
fs/ufs/cylinder.c
163
if (unlikely(!ufs_read_cylinder (sb, cgno, cgno))) {
fs/ufs/cylinder.c
193
ufs_put_cylinder (sb, UFS_MAX_GROUP_LOADED-1);
fs/ufs/cylinder.c
200
if (unlikely(!ufs_read_cylinder (sb, cgno, 0))) {
fs/ufs/cylinder.c
29
static bool ufs_read_cylinder(struct super_block *sb,
fs/ufs/cylinder.c
32
struct ufs_sb_info * sbi = UFS_SB(sb);
fs/ufs/cylinder.c
44
UCPI_UBH(ucpi)->count = uspi->s_cgsize >> sb->s_blocksize_bits;
fs/ufs/cylinder.c
50
UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i);
fs/ufs/cylinder.c
56
ucpi->c_cgx = fs32_to_cpu(sb, ucg->cg_cgx);
fs/ufs/cylinder.c
57
ucpi->c_ncyl = fs16_to_cpu(sb, ucg->cg_ncyl);
fs/ufs/cylinder.c
58
ucpi->c_niblk = fs16_to_cpu(sb, ucg->cg_niblk);
fs/ufs/cylinder.c
59
ucpi->c_ndblk = fs32_to_cpu(sb, ucg->cg_ndblk);
fs/ufs/cylinder.c
60
ucpi->c_rotor = fs32_to_cpu(sb, ucg->cg_rotor);
fs/ufs/cylinder.c
61
ucpi->c_frotor = fs32_to_cpu(sb, ucg->cg_frotor);
fs/ufs/cylinder.c
62
ucpi->c_irotor = fs32_to_cpu(sb, ucg->cg_irotor);
fs/ufs/cylinder.c
63
ucpi->c_btotoff = fs32_to_cpu(sb, ucg->cg_btotoff);
fs/ufs/cylinder.c
64
ucpi->c_boff = fs32_to_cpu(sb, ucg->cg_boff);
fs/ufs/cylinder.c
65
ucpi->c_iusedoff = fs32_to_cpu(sb, ucg->cg_iusedoff);
fs/ufs/cylinder.c
66
ucpi->c_freeoff = fs32_to_cpu(sb, ucg->cg_freeoff);
fs/ufs/cylinder.c
67
ucpi->c_nextfreeoff = fs32_to_cpu(sb, ucg->cg_nextfreeoff);
fs/ufs/cylinder.c
68
ucpi->c_clustersumoff = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_clustersumoff);
fs/ufs/cylinder.c
69
ucpi->c_clusteroff = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_clusteroff);
fs/ufs/cylinder.c
70
ucpi->c_nclusterblks = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_nclusterblks);
fs/ufs/cylinder.c
78
ufs_error (sb, "ufs_read_cylinder", "can't read cylinder group block %u", cgno);
fs/ufs/cylinder.c
86
void ufs_put_cylinder (struct super_block * sb, unsigned bitmap_nr)
fs/ufs/cylinder.c
88
struct ufs_sb_info * sbi = UFS_SB(sb);
fs/ufs/dir.c
113
struct super_block *sb = dir->i_sb;
fs/ufs/dir.c
116
const unsigned chunk_mask = UFS_SB(sb)->s_uspi->s_dirblksize - 1;
fs/ufs/dir.c
129
rec_len = fs16_to_cpu(sb, p->d_reclen);
fs/ufs/dir.c
135
if (rec_len < UFS_DIR_REC_LEN(ufs_get_de_namlen(sb, p)))
fs/ufs/dir.c
139
if (fs32_to_cpu(sb, p->d_ino) > (UFS_SB(sb)->s_uspi->s_ipg *
fs/ufs/dir.c
140
UFS_SB(sb)->s_uspi->s_ncg))
fs/ufs/dir.c
152
ufs_error(sb, __func__,
fs/ufs/dir.c
172
ufs_error(sb, __func__, "bad entry in directory #%lu: %s - "
fs/ufs/dir.c
175
rec_len, ufs_get_de_namlen(sb, p));
fs/ufs/dir.c
179
ufs_error(sb, __func__,
fs/ufs/dir.c
225
ufs_next_entry(struct super_block *sb, struct ufs_dir_entry *p)
fs/ufs/dir.c
228
fs16_to_cpu(sb, p->d_reclen));
fs/ufs/dir.c
252
struct super_block *sb = dir->i_sb;
fs/ufs/dir.c
278
if (ufs_match(sb, namelen, name, de))
fs/ufs/dir.c
280
de = ufs_next_entry(sb, de);
fs/ufs/dir.c
303
struct super_block *sb = dir->i_sb;
fs/ufs/dir.c
305
const unsigned int chunk_size = UFS_SB(sb)->s_uspi->s_dirblksize;
fs/ufs/dir.c
336
de->d_reclen = cpu_to_fs16(sb, chunk_size);
fs/ufs/dir.c
347
if (ufs_match(sb, namelen, name, de))
fs/ufs/dir.c
349
name_len = UFS_DIR_REC_LEN(ufs_get_de_namlen(sb, de));
fs/ufs/dir.c
350
rec_len = fs16_to_cpu(sb, de->d_reclen);
fs/ufs/dir.c
36
static inline int ufs_match(struct super_block *sb, int len,
fs/ufs/dir.c
371
de1->d_reclen = cpu_to_fs16(sb, rec_len - name_len);
fs/ufs/dir.c
372
de->d_reclen = cpu_to_fs16(sb, name_len);
fs/ufs/dir.c
377
ufs_set_de_namlen(sb, de, namelen);
fs/ufs/dir.c
379
de->d_ino = cpu_to_fs32(sb, inode->i_ino);
fs/ufs/dir.c
380
ufs_set_de_type(sb, de, inode->i_mode);
fs/ufs/dir.c
39
if (len != ufs_get_de_namlen(sb, de))
fs/ufs/dir.c
397
ufs_validate_entry(struct super_block *sb, char *base,
fs/ufs/dir.c
403
p = ufs_next_entry(sb, p);
fs/ufs/dir.c
416
struct super_block *sb = inode->i_sb;
fs/ufs/dir.c
420
unsigned chunk_mask = ~(UFS_SB(sb)->s_uspi->s_dirblksize - 1);
fs/ufs/dir.c
422
unsigned flags = UFS_SB(sb)->s_flags;
fs/ufs/dir.c
436
ufs_error(sb, __func__,
fs/ufs/dir.c
444
offset = ufs_validate_entry(sb, kaddr, offset, chunk_mask);
fs/ufs/dir.c
452
for ( ;(char*)de <= limit; de = ufs_next_entry(sb, de)) {
fs/ufs/dir.c
457
fs32_to_cpu(sb, de->d_ino));
fs/ufs/dir.c
458
UFSD("namlen %u\n", ufs_get_de_namlen(sb, de));
fs/ufs/dir.c
464
ufs_get_de_namlen(sb, de),
fs/ufs/dir.c
465
fs32_to_cpu(sb, de->d_ino),
fs/ufs/dir.c
471
ctx->pos += fs16_to_cpu(sb, de->d_reclen);
fs/ufs/dir.c
486
struct super_block *sb = inode->i_sb;
fs/ufs/dir.c
496
to = from + fs16_to_cpu(sb, dir->d_reclen);
fs/ufs/dir.c
498
from &= ~(UFS_SB(sb)->s_uspi->s_dirblksize - 1);
fs/ufs/dir.c
502
fs32_to_cpu(sb, de->d_ino),
fs/ufs/dir.c
503
fs16_to_cpu(sb, de->d_reclen),
fs/ufs/dir.c
504
ufs_get_de_namlen(sb, de), de->d_name);
fs/ufs/dir.c
513
de = ufs_next_entry(sb, de);
fs/ufs/dir.c
525
pde->d_reclen = cpu_to_fs16(sb, to - from);
fs/ufs/dir.c
535
struct super_block * sb = dir->i_sb;
fs/ufs/dir.c
538
const unsigned int chunk_size = UFS_SB(sb)->s_uspi->s_dirblksize;
fs/ufs/dir.c
557
de->d_ino = cpu_to_fs32(sb, inode->i_ino);
fs/ufs/dir.c
558
ufs_set_de_type(sb, de, inode->i_mode);
fs/ufs/dir.c
559
ufs_set_de_namlen(sb, de, 1);
fs/ufs/dir.c
560
de->d_reclen = cpu_to_fs16(sb, UFS_DIR_REC_LEN(1));
fs/ufs/dir.c
563
((char *)de + fs16_to_cpu(sb, de->d_reclen));
fs/ufs/dir.c
564
de->d_ino = cpu_to_fs32(sb, dir->i_ino);
fs/ufs/dir.c
565
ufs_set_de_type(sb, de, dir->i_mode);
fs/ufs/dir.c
566
de->d_reclen = cpu_to_fs16(sb, chunk_size - UFS_DIR_REC_LEN(1));
fs/ufs/dir.c
567
ufs_set_de_namlen(sb, de, 2);
fs/ufs/dir.c
583
struct super_block *sb = inode->i_sb;
fs/ufs/dir.c
606
u16 namelen=ufs_get_de_namlen(sb, de);
fs/ufs/dir.c
614
fs32_to_cpu(sb, de->d_ino))
fs/ufs/dir.c
619
de = ufs_next_entry(sb, de);
fs/ufs/ialloc.c
102
fs32_add(sb, &ucg->cg_cs.cs_nifree, 1);
fs/ufs/ialloc.c
104
fs32_add(sb, &UFS_SB(sb)->fs_cs(cg).cs_nifree, 1);
fs/ufs/ialloc.c
107
fs32_sub(sb, &ucg->cg_cs.cs_ndir, 1);
fs/ufs/ialloc.c
109
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cg).cs_ndir, 1);
fs/ufs/ialloc.c
115
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/ialloc.c
118
ufs_mark_sb_dirty(sb);
fs/ufs/ialloc.c
119
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/ialloc.c
129
static void ufs2_init_inodes_chunk(struct super_block *sb,
fs/ufs/ialloc.c
134
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/ialloc.c
137
fs32_to_cpu(sb, ucg->cg_u.cg_u2.cg_initediblk));
fs/ufs/ialloc.c
143
bh = sb_getblk(sb, beg);
fs/ufs/ialloc.c
145
memset(bh->b_data, 0, sb->s_blocksize);
fs/ufs/ialloc.c
149
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/ialloc.c
154
fs32_add(sb, &ucg->cg_u.cg_u2.cg_initediblk, uspi->s_inopb);
fs/ufs/ialloc.c
156
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/ialloc.c
174
struct super_block * sb;
fs/ufs/ialloc.c
190
sb = dir->i_sb;
fs/ufs/ialloc.c
191
inode = new_inode(sb);
fs/ufs/ialloc.c
195
sbi = UFS_SB(sb);
fs/ufs/ialloc.c
239
ucpi = ufs_load_cylinder (sb, cg);
fs/ufs/ialloc.c
245
if (!ufs_cg_chkmagic(sb, ucg))
fs/ufs/ialloc.c
246
ufs_panic (sb, "ufs_new_inode", "internal error, bad cg magic number");
fs/ufs/ialloc.c
253
ufs_error (sb, "ufs_new_inode",
fs/ufs/ialloc.c
263
ufs_panic (sb, "ufs_new_inode", "internal error");
fs/ufs/ialloc.c
269
u32 initediblk = fs32_to_cpu(sb, ucg->cg_u.cg_u2.cg_initediblk);
fs/ufs/ialloc.c
272
initediblk < fs32_to_cpu(sb, ucg->cg_u.cg_u2.cg_niblk))
fs/ufs/ialloc.c
273
ufs2_init_inodes_chunk(sb, ucpi, ucg);
fs/ufs/ialloc.c
276
fs32_sub(sb, &ucg->cg_cs.cs_nifree, 1);
fs/ufs/ialloc.c
278
fs32_sub(sb, &sbi->fs_cs(cg).cs_nifree, 1);
fs/ufs/ialloc.c
281
fs32_add(sb, &ucg->cg_cs.cs_ndir, 1);
fs/ufs/ialloc.c
283
fs32_add(sb, &sbi->fs_cs(cg).cs_ndir, 1);
fs/ufs/ialloc.c
287
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/ialloc.c
289
ufs_mark_sb_dirty(sb);
fs/ufs/ialloc.c
317
bh = sb_bread(sb, uspi->s_sbbase + ufs_inotofsba(inode->i_ino));
fs/ufs/ialloc.c
319
ufs_warning(sb, "ufs_read_inode",
fs/ufs/ialloc.c
329
ufs2_inode->ui_birthtime = cpu_to_fs64(sb, ts.tv_sec);
fs/ufs/ialloc.c
330
ufs2_inode->ui_birthnsec = cpu_to_fs32(sb, ts.tv_nsec);
fs/ufs/ialloc.c
333
if (sb->s_flags & SB_SYNCHRONOUS)
fs/ufs/ialloc.c
59
struct super_block * sb;
fs/ufs/ialloc.c
68
sb = inode->i_sb;
fs/ufs/ialloc.c
69
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/ialloc.c
73
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/ialloc.c
76
ufs_warning(sb, "ufs_free_inode", "reserved inode or nonexistent inode %u\n", ino);
fs/ufs/ialloc.c
77
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/ialloc.c
83
ucpi = ufs_load_cylinder (sb, cg);
fs/ufs/ialloc.c
85
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/ialloc.c
89
if (!ufs_cg_chkmagic(sb, ucg))
fs/ufs/ialloc.c
90
ufs_panic (sb, "ufs_free_fragments", "internal error, bad cg magic number");
fs/ufs/ialloc.c
92
ucg->cg_time = ufs_get_seconds(sb);
fs/ufs/ialloc.c
97
ufs_error(sb, "ufs_free_inode", "bit already cleared for inode %u", ino);
fs/ufs/inode.c
1004
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1018
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1037
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
1039
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
1086
bh = sb_getblk(sb, i + phys64);
fs/ufs/inode.c
1088
memset(bh->b_data, 0, sb->s_blocksize);
fs/ufs/inode.c
1105
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
1106
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
1136
block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
1139
ubh[i] = ubh_bread(sb, block, uspi->s_bsize);
fs/ufs/inode.c
1153
block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
128
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
129
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
134
unsigned flags = UFS_SB(sb)->s_flags;
fs/ufs/inode.c
159
bh = sb_bread(sb, uspi->s_sbbase +
fs/ufs/inode.c
160
fs32_to_cpu(sb, q->key32) + (n>>shift));
fs/ufs/inode.c
169
res = fs32_to_cpu(sb, q->key32);
fs/ufs/inode.c
183
bh = sb_bread(sb, uspi->s_sbbase +
fs/ufs/inode.c
184
fs64_to_cpu(sb, q->key64) + (n>>shift));
fs/ufs/inode.c
193
res = fs64_to_cpu(sb, q->key64);
fs/ufs/inode.c
226
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
227
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
240
tmp = ufs_new_fragments(inode, p, lastfrag, ufs_data_ptr_to_cpu(sb, p),
fs/ufs/inode.c
260
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
261
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
267
tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
279
goal = ufs_data_ptr_to_cpu(sb,
fs/ufs/inode.c
317
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
318
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
327
bh = sb_bread(sb, ind_block + (index >> shift));
fs/ufs/inode.c
339
tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
344
(tmp = fs64_to_cpu(sb, ((__fs64 *)bh->b_data)[index-1])) :
fs/ufs/inode.c
345
(tmp = fs32_to_cpu(sb, ((__fs32 *)bh->b_data)[index-1]))))
fs/ufs/inode.c
377
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
378
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
405
ufs_warning(sb, "ufs_get_block", "block > big");
fs/ufs/inode.c
436
map_bh(bh_result, sb, phys64);
fs/ufs/inode.c
445
map_bh(bh_result, sb, phys64 + frag);
fs/ufs/inode.c
547
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
553
inode->i_mode = mode = fs16_to_cpu(sb, ufs_inode->ui_mode);
fs/ufs/inode.c
554
set_nlink(inode, fs16_to_cpu(sb, ufs_inode->ui_nlink));
fs/ufs/inode.c
561
i_uid_write(inode, ufs_get_inode_uid(sb, ufs_inode));
fs/ufs/inode.c
562
i_gid_write(inode, ufs_get_inode_gid(sb, ufs_inode));
fs/ufs/inode.c
564
inode->i_size = fs64_to_cpu(sb, ufs_inode->ui_size);
fs/ufs/inode.c
566
(signed)fs32_to_cpu(sb, ufs_inode->ui_atime.tv_sec),
fs/ufs/inode.c
569
(signed)fs32_to_cpu(sb, ufs_inode->ui_ctime.tv_sec),
fs/ufs/inode.c
572
(signed)fs32_to_cpu(sb, ufs_inode->ui_mtime.tv_sec),
fs/ufs/inode.c
574
inode->i_blocks = fs32_to_cpu(sb, ufs_inode->ui_blocks);
fs/ufs/inode.c
575
inode->i_generation = fs32_to_cpu(sb, ufs_inode->ui_gen);
fs/ufs/inode.c
576
ufsi->i_flags = fs32_to_cpu(sb, ufs_inode->ui_flags);
fs/ufs/inode.c
577
ufsi->i_shadow = fs32_to_cpu(sb, ufs_inode->ui_u3.ui_sun.ui_shadow);
fs/ufs/inode.c
578
ufsi->i_oeftflag = fs32_to_cpu(sb, ufs_inode->ui_u3.ui_sun.ui_oeftflag);
fs/ufs/inode.c
595
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
602
inode->i_mode = mode = fs16_to_cpu(sb, ufs2_inode->ui_mode);
fs/ufs/inode.c
603
set_nlink(inode, fs16_to_cpu(sb, ufs2_inode->ui_nlink));
fs/ufs/inode.c
610
i_uid_write(inode, fs32_to_cpu(sb, ufs2_inode->ui_uid));
fs/ufs/inode.c
611
i_gid_write(inode, fs32_to_cpu(sb, ufs2_inode->ui_gid));
fs/ufs/inode.c
613
inode->i_size = fs64_to_cpu(sb, ufs2_inode->ui_size);
fs/ufs/inode.c
614
inode_set_atime(inode, fs64_to_cpu(sb, ufs2_inode->ui_atime),
fs/ufs/inode.c
615
fs32_to_cpu(sb, ufs2_inode->ui_atimensec));
fs/ufs/inode.c
616
inode_set_ctime(inode, fs64_to_cpu(sb, ufs2_inode->ui_ctime),
fs/ufs/inode.c
617
fs32_to_cpu(sb, ufs2_inode->ui_ctimensec));
fs/ufs/inode.c
618
inode_set_mtime(inode, fs64_to_cpu(sb, ufs2_inode->ui_mtime),
fs/ufs/inode.c
619
fs32_to_cpu(sb, ufs2_inode->ui_mtimensec));
fs/ufs/inode.c
620
inode->i_blocks = fs64_to_cpu(sb, ufs2_inode->ui_blocks);
fs/ufs/inode.c
621
inode->i_generation = fs32_to_cpu(sb, ufs2_inode->ui_gen);
fs/ufs/inode.c
622
ufsi->i_flags = fs32_to_cpu(sb, ufs2_inode->ui_flags);
fs/ufs/inode.c
639
struct inode *ufs_iget(struct super_block *sb, unsigned long ino)
fs/ufs/inode.c
642
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
650
ufs_warning(sb, "ufs_read_inode", "bad inode number (%lu)\n",
fs/ufs/inode.c
655
inode = iget_locked(sb, ino);
fs/ufs/inode.c
663
bh = sb_bread(sb, uspi->s_sbbase + ufs_inotofsba(inode->i_ino));
fs/ufs/inode.c
665
ufs_warning(sb, "ufs_read_inode", "unable to read inode %lu\n",
fs/ufs/inode.c
669
if ((UFS_SB(sb)->s_flags & UFS_TYPE_MASK) == UFS_TYPE_UFS2) {
fs/ufs/inode.c
703
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
706
ufs_inode->ui_mode = cpu_to_fs16(sb, inode->i_mode);
fs/ufs/inode.c
707
ufs_inode->ui_nlink = cpu_to_fs16(sb, inode->i_nlink);
fs/ufs/inode.c
709
ufs_set_inode_uid(sb, ufs_inode, i_uid_read(inode));
fs/ufs/inode.c
710
ufs_set_inode_gid(sb, ufs_inode, i_gid_read(inode));
fs/ufs/inode.c
712
ufs_inode->ui_size = cpu_to_fs64(sb, inode->i_size);
fs/ufs/inode.c
713
ufs_inode->ui_atime.tv_sec = cpu_to_fs32(sb,
fs/ufs/inode.c
716
ufs_inode->ui_ctime.tv_sec = cpu_to_fs32(sb,
fs/ufs/inode.c
719
ufs_inode->ui_mtime.tv_sec = cpu_to_fs32(sb,
fs/ufs/inode.c
722
ufs_inode->ui_blocks = cpu_to_fs32(sb, inode->i_blocks);
fs/ufs/inode.c
723
ufs_inode->ui_flags = cpu_to_fs32(sb, ufsi->i_flags);
fs/ufs/inode.c
724
ufs_inode->ui_gen = cpu_to_fs32(sb, inode->i_generation);
fs/ufs/inode.c
726
if ((UFS_SB(sb)->s_flags & UFS_UID_MASK) == UFS_UID_EFT) {
fs/ufs/inode.c
727
ufs_inode->ui_u3.ui_sun.ui_shadow = cpu_to_fs32(sb, ufsi->i_shadow);
fs/ufs/inode.c
728
ufs_inode->ui_u3.ui_sun.ui_oeftflag = cpu_to_fs32(sb, ufsi->i_oeftflag);
fs/ufs/inode.c
749
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
753
ufs_inode->ui_mode = cpu_to_fs16(sb, inode->i_mode);
fs/ufs/inode.c
754
ufs_inode->ui_nlink = cpu_to_fs16(sb, inode->i_nlink);
fs/ufs/inode.c
756
ufs_inode->ui_uid = cpu_to_fs32(sb, i_uid_read(inode));
fs/ufs/inode.c
757
ufs_inode->ui_gid = cpu_to_fs32(sb, i_gid_read(inode));
fs/ufs/inode.c
759
ufs_inode->ui_size = cpu_to_fs64(sb, inode->i_size);
fs/ufs/inode.c
760
ufs_inode->ui_atime = cpu_to_fs64(sb, inode_get_atime_sec(inode));
fs/ufs/inode.c
761
ufs_inode->ui_atimensec = cpu_to_fs32(sb,
fs/ufs/inode.c
763
ufs_inode->ui_ctime = cpu_to_fs64(sb, inode_get_ctime_sec(inode));
fs/ufs/inode.c
764
ufs_inode->ui_ctimensec = cpu_to_fs32(sb,
fs/ufs/inode.c
766
ufs_inode->ui_mtime = cpu_to_fs64(sb, inode_get_mtime_sec(inode));
fs/ufs/inode.c
767
ufs_inode->ui_mtimensec = cpu_to_fs32(sb,
fs/ufs/inode.c
770
ufs_inode->ui_blocks = cpu_to_fs64(sb, inode->i_blocks);
fs/ufs/inode.c
771
ufs_inode->ui_flags = cpu_to_fs32(sb, ufsi->i_flags);
fs/ufs/inode.c
772
ufs_inode->ui_gen = cpu_to_fs32(sb, inode->i_generation);
fs/ufs/inode.c
792
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
793
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
800
ufs_warning (sb, "ufs_read_inode", "bad inode number (%lu)\n", inode->i_ino);
fs/ufs/inode.c
804
bh = sb_bread(sb, ufs_inotofsba(inode->i_ino));
fs/ufs/inode.c
806
ufs_warning (sb, "ufs_read_inode", "unable to read inode %lu\n", inode->i_ino);
fs/ufs/inode.c
887
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
888
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
911
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
913
ufs_panic(sb, __func__, "internal error");
fs/ufs/inode.c
925
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
927
ufs_panic(sb, __func__, "internal error");
fs/ufs/inode.c
934
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
948
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
950
ufs_panic(sb, __func__, "internal error");
fs/ufs/inode.c
964
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
965
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/inode.c
966
struct ufs_buffer_head *ubh = ubh_bread(sb, ind_block, uspi->s_bsize);
fs/ufs/inode.c
975
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
984
u64 block = ufs_data_ptr_to_cpu(sb, p);
fs/ufs/inode.c
997
struct super_block *sb = inode->i_sb;
fs/ufs/inode.c
998
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/namei.c
112
struct super_block * sb = dir->i_sb;
fs/ufs/namei.c
117
if (l > sb->s_blocksize)
fs/ufs/namei.c
125
if (l > UFS_SB(sb)->s_uspi->s_maxsymlinklen) {
fs/ufs/super.c
1000
uspi->s_bsize = fs32_to_cpu(sb, usb1->fs_bsize);
fs/ufs/super.c
1001
uspi->s_fsize = fs32_to_cpu(sb, usb1->fs_fsize);
fs/ufs/super.c
1002
uspi->s_sbsize = fs32_to_cpu(sb, usb1->fs_sbsize);
fs/ufs/super.c
1003
uspi->s_fmask = fs32_to_cpu(sb, usb1->fs_fmask);
fs/ufs/super.c
1004
uspi->s_fshift = fs32_to_cpu(sb, usb1->fs_fshift);
fs/ufs/super.c
101
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
1046
ufs_print_super_stuff(sb, usb1, usb2, usb3);
fs/ufs/super.c
1057
(ufs_get_fs_state(sb, usb1, usb3) == (UFS_FSOK - fs32_to_cpu(sb, usb1->fs_time))))) {
fs/ufs/super.c
107
inode = ufs_iget(sb, ino);
fs/ufs/super.c
1073
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
1077
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
1082
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
1087
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
1093
sb->s_op = &ufs_super_ops;
fs/ufs/super.c
1094
sb->s_export_op = &ufs_export_ops;
fs/ufs/super.c
1096
sb->s_magic = fs32_to_cpu(sb, usb3->fs_magic);
fs/ufs/super.c
1098
uspi->s_sblkno = fs32_to_cpu(sb, usb1->fs_sblkno);
fs/ufs/super.c
1099
uspi->s_cblkno = fs32_to_cpu(sb, usb1->fs_cblkno);
fs/ufs/super.c
1100
uspi->s_iblkno = fs32_to_cpu(sb, usb1->fs_iblkno);
fs/ufs/super.c
1101
uspi->s_dblkno = fs32_to_cpu(sb, usb1->fs_dblkno);
fs/ufs/super.c
1102
uspi->s_cgoffset = fs32_to_cpu(sb, usb1->fs_cgoffset);
fs/ufs/super.c
1103
uspi->s_cgmask = fs32_to_cpu(sb, usb1->fs_cgmask);
fs/ufs/super.c
1106
uspi->s_size = fs64_to_cpu(sb, usb3->fs_un1.fs_u2.fs_size);
fs/ufs/super.c
1107
uspi->s_dsize = fs64_to_cpu(sb, usb3->fs_un1.fs_u2.fs_dsize);
fs/ufs/super.c
1109
uspi->s_size = fs32_to_cpu(sb, usb1->fs_size);
fs/ufs/super.c
1110
uspi->s_dsize = fs32_to_cpu(sb, usb1->fs_dsize);
fs/ufs/super.c
1113
uspi->s_ncg = fs32_to_cpu(sb, usb1->fs_ncg);
fs/ufs/super.c
1116
uspi->s_fpb = fs32_to_cpu(sb, usb1->fs_frag);
fs/ufs/super.c
1117
uspi->s_minfree = fs32_to_cpu(sb, usb1->fs_minfree);
fs/ufs/super.c
1118
uspi->s_bmask = fs32_to_cpu(sb, usb1->fs_bmask);
fs/ufs/super.c
1119
uspi->s_fmask = fs32_to_cpu(sb, usb1->fs_fmask);
fs/ufs/super.c
1120
uspi->s_bshift = fs32_to_cpu(sb, usb1->fs_bshift);
fs/ufs/super.c
1121
uspi->s_fshift = fs32_to_cpu(sb, usb1->fs_fshift);
fs/ufs/super.c
1124
uspi->s_fpbshift = fs32_to_cpu(sb, usb1->fs_fragshift);
fs/ufs/super.c
1125
uspi->s_fsbtodb = fs32_to_cpu(sb, usb1->fs_fsbtodb);
fs/ufs/super.c
1127
uspi->s_csmask = fs32_to_cpu(sb, usb1->fs_csmask);
fs/ufs/super.c
1128
uspi->s_csshift = fs32_to_cpu(sb, usb1->fs_csshift);
fs/ufs/super.c
1129
uspi->s_nindir = fs32_to_cpu(sb, usb1->fs_nindir);
fs/ufs/super.c
1130
uspi->s_inopb = fs32_to_cpu(sb, usb1->fs_inopb);
fs/ufs/super.c
1131
uspi->s_nspf = fs32_to_cpu(sb, usb1->fs_nspf);
fs/ufs/super.c
1132
uspi->s_npsect = ufs_get_fs_npsect(sb, usb1, usb3);
fs/ufs/super.c
1133
uspi->s_interleave = fs32_to_cpu(sb, usb1->fs_interleave);
fs/ufs/super.c
1134
uspi->s_trackskew = fs32_to_cpu(sb, usb1->fs_trackskew);
fs/ufs/super.c
1137
uspi->s_csaddr = fs64_to_cpu(sb, usb3->fs_un1.fs_u2.fs_csaddr);
fs/ufs/super.c
1139
uspi->s_csaddr = fs32_to_cpu(sb, usb1->fs_csaddr);
fs/ufs/super.c
1141
uspi->s_cssize = fs32_to_cpu(sb, usb1->fs_cssize);
fs/ufs/super.c
1142
uspi->s_cgsize = fs32_to_cpu(sb, usb1->fs_cgsize);
fs/ufs/super.c
1143
uspi->s_ntrak = fs32_to_cpu(sb, usb1->fs_ntrak);
fs/ufs/super.c
1144
uspi->s_nsect = fs32_to_cpu(sb, usb1->fs_nsect);
fs/ufs/super.c
1145
uspi->s_spc = fs32_to_cpu(sb, usb1->fs_spc);
fs/ufs/super.c
1146
uspi->s_ipg = fs32_to_cpu(sb, usb1->fs_ipg);
fs/ufs/super.c
1147
uspi->s_fpg = fs32_to_cpu(sb, usb1->fs_fpg);
fs/ufs/super.c
1148
uspi->s_cpc = fs32_to_cpu(sb, usb2->fs_un.fs_u1.fs_cpc);
fs/ufs/super.c
1149
uspi->s_contigsumsize = fs32_to_cpu(sb, usb3->fs_un2.fs_44.fs_contigsumsize);
fs/ufs/super.c
1150
uspi->s_qbmask = ufs_get_fs_qbmask(sb, usb3);
fs/ufs/super.c
1151
uspi->s_qfmask = ufs_get_fs_qfmask(sb, usb3);
fs/ufs/super.c
1152
uspi->s_nrpos = fs32_to_cpu(sb, usb3->fs_nrpos);
fs/ufs/super.c
1153
uspi->s_postbloff = fs32_to_cpu(sb, usb3->fs_postbloff);
fs/ufs/super.c
1154
uspi->s_rotbloff = fs32_to_cpu(sb, usb3->fs_rotbloff);
fs/ufs/super.c
1161
usb1->fs_optim = cpu_to_fs32(sb, UFS_OPTSPACE);
fs/ufs/super.c
117
static struct dentry *ufs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/ufs/super.c
1188
fs32_to_cpu(sb, usb3->fs_un2.fs_44.fs_maxsymlinklen);
fs/ufs/super.c
1195
ufs_warning(sb, __func__, "ufs_read_super: excessive maximum "
fs/ufs/super.c
1199
sb->s_maxbytes = ufs_max_bytes(sb);
fs/ufs/super.c
120
return generic_fh_to_dentry(sb, fid, fh_len, fh_type, ufs_nfs_get_inode);
fs/ufs/super.c
1200
sb->s_max_links = UFS_LINK_MAX;
fs/ufs/super.c
1202
inode = ufs_iget(sb, UFS_ROOTINO);
fs/ufs/super.c
1207
sb->s_root = d_make_root(inode);
fs/ufs/super.c
1208
if (!sb->s_root) {
fs/ufs/super.c
1213
ufs_setup_cstotal(sb);
fs/ufs/super.c
1217
if (!sb_rdonly(sb))
fs/ufs/super.c
1218
if (!ufs_read_cylinder_structures(sb))
fs/ufs/super.c
1229
sb->s_fs_info = NULL;
fs/ufs/super.c
123
static struct dentry *ufs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/ufs/super.c
1244
struct super_block *sb = fc->root->d_sb;
fs/ufs/super.c
1248
sync_filesystem(sb);
fs/ufs/super.c
1249
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1250
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
1251
flags = UFS_SB(sb)->s_flags;
fs/ufs/super.c
1255
ufstype = UFS_SB(sb)->s_flavour;
fs/ufs/super.c
1257
if ((bool)(fc->sb_flags & SB_RDONLY) == sb_rdonly(sb)) {
fs/ufs/super.c
1258
UFS_SB(sb)->s_on_err = ctx->on_err;
fs/ufs/super.c
1259
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
126
return generic_fh_to_parent(sb, fid, fh_len, fh_type, ufs_nfs_get_inode);
fs/ufs/super.c
1267
ufs_put_super_internal(sb);
fs/ufs/super.c
1268
usb1->fs_time = ufs_get_seconds(sb);
fs/ufs/super.c
1272
ufs_set_fs_state(sb, usb1, usb3,
fs/ufs/super.c
1273
UFS_FSOK - fs32_to_cpu(sb, usb1->fs_time));
fs/ufs/super.c
1275
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
1282
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1291
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1294
if (!ufs_read_cylinder_structures(sb)) {
fs/ufs/super.c
1296
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1299
sb->s_flags &= ~SB_RDONLY;
fs/ufs/super.c
1302
UFS_SB(sb)->s_on_err = ctx->on_err;
fs/ufs/super.c
1303
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1329
struct super_block *sb = dentry->d_sb;
fs/ufs/super.c
1330
struct ufs_sb_private_info *uspi= UFS_SB(sb)->s_uspi;
fs/ufs/super.c
1331
unsigned flags = UFS_SB(sb)->s_flags;
fs/ufs/super.c
1332
u64 id = huge_encode_dev(sb->s_bdev->bd_dev);
fs/ufs/super.c
1334
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1344
buf->f_bsize = sb->s_blocksize;
fs/ufs/super.c
1351
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
1358
static struct inode *ufs_alloc_inode(struct super_block *sb)
fs/ufs/super.c
1362
ei = alloc_inode_sb(sb, ufs_inode_cachep, GFP_NOFS);
fs/ufs/super.c
1445
struct super_block *sb = fc->root->d_sb;
fs/ufs/super.c
1446
struct ufs_sb_info *sbi = UFS_SB(sb);
fs/ufs/super.c
150
static void ufs_print_super_stuff(struct super_block *sb,
fs/ufs/super.c
155
u32 magic = fs32_to_cpu(sb, usb3->fs_magic);
fs/ufs/super.c
159
if (fs32_to_cpu(sb, usb3->fs_magic) == UFS2_MAGIC) {
fs/ufs/super.c
161
fs64_to_cpu(sb, usb3->fs_un1.fs_u2.fs_size));
fs/ufs/super.c
163
fs64_to_cpu(sb, usb3->fs_un1.fs_u2.fs_dsize));
fs/ufs/super.c
165
fs32_to_cpu(sb, usb1->fs_bsize));
fs/ufs/super.c
167
fs32_to_cpu(sb, usb1->fs_fsize));
fs/ufs/super.c
170
fs64_to_cpu(sb, usb2->fs_un.fs_u2.fs_sblockloc));
fs/ufs/super.c
172
fs64_to_cpu(sb, usb2->fs_un.fs_u2.cs_ndir));
fs/ufs/super.c
175
fs64_to_cpu(sb, usb2->fs_un.fs_u2.cs_nbfree));
fs/ufs/super.c
178
fs64_to_cpu(sb, usb3->fs_un1.fs_u2.cs_nifree));
fs/ufs/super.c
181
fs64_to_cpu(sb, usb3->fs_un1.fs_u2.cs_nffree));
fs/ufs/super.c
183
fs32_to_cpu(sb, usb3->fs_un2.fs_44.fs_maxsymlinklen));
fs/ufs/super.c
185
pr_debug(" sblkno: %u\n", fs32_to_cpu(sb, usb1->fs_sblkno));
fs/ufs/super.c
186
pr_debug(" cblkno: %u\n", fs32_to_cpu(sb, usb1->fs_cblkno));
fs/ufs/super.c
187
pr_debug(" iblkno: %u\n", fs32_to_cpu(sb, usb1->fs_iblkno));
fs/ufs/super.c
188
pr_debug(" dblkno: %u\n", fs32_to_cpu(sb, usb1->fs_dblkno));
fs/ufs/super.c
190
fs32_to_cpu(sb, usb1->fs_cgoffset));
fs/ufs/super.c
192
~fs32_to_cpu(sb, usb1->fs_cgmask));
fs/ufs/super.c
193
pr_debug(" size: %u\n", fs32_to_cpu(sb, usb1->fs_size));
fs/ufs/super.c
194
pr_debug(" dsize: %u\n", fs32_to_cpu(sb, usb1->fs_dsize));
fs/ufs/super.c
195
pr_debug(" ncg: %u\n", fs32_to_cpu(sb, usb1->fs_ncg));
fs/ufs/super.c
196
pr_debug(" bsize: %u\n", fs32_to_cpu(sb, usb1->fs_bsize));
fs/ufs/super.c
197
pr_debug(" fsize: %u\n", fs32_to_cpu(sb, usb1->fs_fsize));
fs/ufs/super.c
198
pr_debug(" frag: %u\n", fs32_to_cpu(sb, usb1->fs_frag));
fs/ufs/super.c
200
fs32_to_cpu(sb, usb1->fs_fragshift));
fs/ufs/super.c
201
pr_debug(" ~fmask: %u\n", ~fs32_to_cpu(sb, usb1->fs_fmask));
fs/ufs/super.c
202
pr_debug(" fshift: %u\n", fs32_to_cpu(sb, usb1->fs_fshift));
fs/ufs/super.c
203
pr_debug(" sbsize: %u\n", fs32_to_cpu(sb, usb1->fs_sbsize));
fs/ufs/super.c
204
pr_debug(" spc: %u\n", fs32_to_cpu(sb, usb1->fs_spc));
fs/ufs/super.c
205
pr_debug(" cpg: %u\n", fs32_to_cpu(sb, usb1->fs_cpg));
fs/ufs/super.c
206
pr_debug(" ipg: %u\n", fs32_to_cpu(sb, usb1->fs_ipg));
fs/ufs/super.c
207
pr_debug(" fpg: %u\n", fs32_to_cpu(sb, usb1->fs_fpg));
fs/ufs/super.c
208
pr_debug(" csaddr: %u\n", fs32_to_cpu(sb, usb1->fs_csaddr));
fs/ufs/super.c
209
pr_debug(" cssize: %u\n", fs32_to_cpu(sb, usb1->fs_cssize));
fs/ufs/super.c
210
pr_debug(" cgsize: %u\n", fs32_to_cpu(sb, usb1->fs_cgsize));
fs/ufs/super.c
212
fs32_to_cpu(sb, usb1->fs_fsbtodb));
fs/ufs/super.c
213
pr_debug(" nrpos: %u\n", fs32_to_cpu(sb, usb3->fs_nrpos));
fs/ufs/super.c
215
fs32_to_cpu(sb, usb1->fs_cstotal.cs_ndir));
fs/ufs/super.c
217
fs32_to_cpu(sb, usb1->fs_cstotal.cs_nifree));
fs/ufs/super.c
219
fs32_to_cpu(sb, usb1->fs_cstotal.cs_nbfree));
fs/ufs/super.c
221
fs32_to_cpu(sb, usb1->fs_cstotal.cs_nffree));
fs/ufs/super.c
229
static void ufs_print_cylinder_stuff(struct super_block *sb,
fs/ufs/super.c
234
pr_debug(" magic: %x\n", fs32_to_cpu(sb, cg->cg_magic));
fs/ufs/super.c
235
pr_debug(" time: %u\n", fs32_to_cpu(sb, cg->cg_time));
fs/ufs/super.c
236
pr_debug(" cgx: %u\n", fs32_to_cpu(sb, cg->cg_cgx));
fs/ufs/super.c
237
pr_debug(" ncyl: %u\n", fs16_to_cpu(sb, cg->cg_ncyl));
fs/ufs/super.c
238
pr_debug(" niblk: %u\n", fs16_to_cpu(sb, cg->cg_niblk));
fs/ufs/super.c
239
pr_debug(" ndblk: %u\n", fs32_to_cpu(sb, cg->cg_ndblk));
fs/ufs/super.c
240
pr_debug(" cs_ndir: %u\n", fs32_to_cpu(sb, cg->cg_cs.cs_ndir));
fs/ufs/super.c
241
pr_debug(" cs_nbfree: %u\n", fs32_to_cpu(sb, cg->cg_cs.cs_nbfree));
fs/ufs/super.c
242
pr_debug(" cs_nifree: %u\n", fs32_to_cpu(sb, cg->cg_cs.cs_nifree));
fs/ufs/super.c
243
pr_debug(" cs_nffree: %u\n", fs32_to_cpu(sb, cg->cg_cs.cs_nffree));
fs/ufs/super.c
244
pr_debug(" rotor: %u\n", fs32_to_cpu(sb, cg->cg_rotor));
fs/ufs/super.c
245
pr_debug(" frotor: %u\n", fs32_to_cpu(sb, cg->cg_frotor));
fs/ufs/super.c
246
pr_debug(" irotor: %u\n", fs32_to_cpu(sb, cg->cg_irotor));
fs/ufs/super.c
248
fs32_to_cpu(sb, cg->cg_frsum[0]), fs32_to_cpu(sb, cg->cg_frsum[1]),
fs/ufs/super.c
249
fs32_to_cpu(sb, cg->cg_frsum[2]), fs32_to_cpu(sb, cg->cg_frsum[3]),
fs/ufs/super.c
250
fs32_to_cpu(sb, cg->cg_frsum[4]), fs32_to_cpu(sb, cg->cg_frsum[5]),
fs/ufs/super.c
251
fs32_to_cpu(sb, cg->cg_frsum[6]), fs32_to_cpu(sb, cg->cg_frsum[7]));
fs/ufs/super.c
252
pr_debug(" btotoff: %u\n", fs32_to_cpu(sb, cg->cg_btotoff));
fs/ufs/super.c
253
pr_debug(" boff: %u\n", fs32_to_cpu(sb, cg->cg_boff));
fs/ufs/super.c
254
pr_debug(" iuseoff: %u\n", fs32_to_cpu(sb, cg->cg_iusedoff));
fs/ufs/super.c
255
pr_debug(" freeoff: %u\n", fs32_to_cpu(sb, cg->cg_freeoff));
fs/ufs/super.c
256
pr_debug(" nextfreeoff: %u\n", fs32_to_cpu(sb, cg->cg_nextfreeoff));
fs/ufs/super.c
258
fs32_to_cpu(sb, cg->cg_u.cg_44.cg_clustersumoff));
fs/ufs/super.c
260
fs32_to_cpu(sb, cg->cg_u.cg_44.cg_clusteroff));
fs/ufs/super.c
262
fs32_to_cpu(sb, cg->cg_u.cg_44.cg_nclusterblks));
fs/ufs/super.c
266
# define ufs_print_super_stuff(sb, usb1, usb2, usb3) /**/
fs/ufs/super.c
267
# define ufs_print_cylinder_stuff(sb, cg) /**/
fs/ufs/super.c
272
void ufs_error (struct super_block * sb, const char * function,
fs/ufs/super.c
280
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
283
if (!sb_rdonly(sb)) {
fs/ufs/super.c
286
ufs_mark_sb_dirty(sb);
fs/ufs/super.c
287
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
292
switch (UFS_SB(sb)->s_on_err) {
fs/ufs/super.c
295
sb->s_id, function, &vaf);
fs/ufs/super.c
301
sb->s_id, function, &vaf);
fs/ufs/super.c
306
void ufs_panic (struct super_block * sb, const char * function,
fs/ufs/super.c
314
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
317
if (!sb_rdonly(sb)) {
fs/ufs/super.c
320
ufs_mark_sb_dirty(sb);
fs/ufs/super.c
325
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
327
sb->s_id, function, &vaf);
fs/ufs/super.c
331
void ufs_warning (struct super_block * sb, const char * function,
fs/ufs/super.c
341
sb->s_id, function, &vaf);
fs/ufs/super.c
420
static void ufs_setup_cstotal(struct super_block *sb)
fs/ufs/super.c
422
struct ufs_sb_info *sbi = UFS_SB(sb);
fs/ufs/super.c
438
uspi->cs_total.cs_ndir = fs64_to_cpu(sb, usb2->fs_un.fs_u2.cs_ndir);
fs/ufs/super.c
439
uspi->cs_total.cs_nbfree = fs64_to_cpu(sb, usb2->fs_un.fs_u2.cs_nbfree);
fs/ufs/super.c
440
uspi->cs_total.cs_nifree = fs64_to_cpu(sb, usb3->fs_un1.fs_u2.cs_nifree);
fs/ufs/super.c
441
uspi->cs_total.cs_nffree = fs64_to_cpu(sb, usb3->fs_un1.fs_u2.cs_nffree);
fs/ufs/super.c
443
uspi->cs_total.cs_ndir = fs32_to_cpu(sb, usb1->fs_cstotal.cs_ndir);
fs/ufs/super.c
444
uspi->cs_total.cs_nbfree = fs32_to_cpu(sb, usb1->fs_cstotal.cs_nbfree);
fs/ufs/super.c
445
uspi->cs_total.cs_nifree = fs32_to_cpu(sb, usb1->fs_cstotal.cs_nifree);
fs/ufs/super.c
446
uspi->cs_total.cs_nffree = fs32_to_cpu(sb, usb1->fs_cstotal.cs_nffree);
fs/ufs/super.c
454
static int ufs_read_cylinder_structures(struct super_block *sb)
fs/ufs/super.c
456
struct ufs_sb_info *sbi = UFS_SB(sb);
fs/ufs/super.c
474
struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
fs/ufs/super.c
497
if (!(sbi->s_ucg[i] = sb_bread(sb, ufs_cgcmin(i))))
fs/ufs/super.c
499
if (!ufs_cg_chkmagic (sb, (struct ufs_cylinder_group *) sbi->s_ucg[i]->b_data))
fs/ufs/super.c
502
ufs_print_cylinder_stuff(sb, (struct ufs_cylinder_group *) sbi->s_ucg[i]->b_data);
fs/ufs/super.c
530
static void ufs_put_cstotal(struct super_block *sb)
fs/ufs/super.c
532
unsigned mtype = UFS_SB(sb)->s_flavour;
fs/ufs/super.c
533
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
546
cpu_to_fs64(sb, uspi->cs_total.cs_ndir);
fs/ufs/super.c
548
cpu_to_fs64(sb, uspi->cs_total.cs_nbfree);
fs/ufs/super.c
550
cpu_to_fs64(sb, uspi->cs_total.cs_nifree);
fs/ufs/super.c
552
cpu_to_fs64(sb, uspi->cs_total.cs_nffree);
fs/ufs/super.c
560
cpu_to_fs64(sb, uspi->cs_total.cs_ndir);
fs/ufs/super.c
562
cpu_to_fs64(sb, uspi->cs_total.cs_nbfree);
fs/ufs/super.c
564
cpu_to_fs64(sb, uspi->cs_total.cs_nifree);
fs/ufs/super.c
566
cpu_to_fs64(sb, uspi->cs_total.cs_nffree);
fs/ufs/super.c
568
usb1->fs_cstotal.cs_ndir = cpu_to_fs32(sb, uspi->cs_total.cs_ndir);
fs/ufs/super.c
569
usb1->fs_cstotal.cs_nbfree = cpu_to_fs32(sb, uspi->cs_total.cs_nbfree);
fs/ufs/super.c
570
usb1->fs_cstotal.cs_nifree = cpu_to_fs32(sb, uspi->cs_total.cs_nifree);
fs/ufs/super.c
571
usb1->fs_cstotal.cs_nffree = cpu_to_fs32(sb, uspi->cs_total.cs_nffree);
fs/ufs/super.c
574
ufs_print_super_stuff(sb, usb1, usb2, usb3);
fs/ufs/super.c
584
static void ufs_put_super_internal(struct super_block *sb)
fs/ufs/super.c
586
struct ufs_sb_info *sbi = UFS_SB(sb);
fs/ufs/super.c
594
ufs_put_cstotal(sb);
fs/ufs/super.c
599
struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
fs/ufs/super.c
602
ufs_panic(sb, __func__,
fs/ufs/super.c
611
ufs_put_cylinder (sb, i);
fs/ufs/super.c
624
static int ufs_sync_fs(struct super_block *sb, int wait)
fs/ufs/super.c
631
mutex_lock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
635
flags = UFS_SB(sb)->s_flags;
fs/ufs/super.c
636
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
640
usb1->fs_time = ufs_get_seconds(sb);
fs/ufs/super.c
644
ufs_set_fs_state(sb, usb1, usb3,
fs/ufs/super.c
645
UFS_FSOK - fs32_to_cpu(sb, usb1->fs_time));
fs/ufs/super.c
646
ufs_put_cstotal(sb);
fs/ufs/super.c
649
mutex_unlock(&UFS_SB(sb)->s_lock);
fs/ufs/super.c
664
ufs_sync_fs(sbi->sb, 1);
fs/ufs/super.c
667
void ufs_mark_sb_dirty(struct super_block *sb)
fs/ufs/super.c
669
struct ufs_sb_info *sbi = UFS_SB(sb);
fs/ufs/super.c
681
static void ufs_put_super(struct super_block *sb)
fs/ufs/super.c
683
struct ufs_sb_info * sbi = UFS_SB(sb);
fs/ufs/super.c
687
if (!sb_rdonly(sb))
fs/ufs/super.c
688
ufs_put_super_internal(sb);
fs/ufs/super.c
694
sb->s_fs_info = NULL;
fs/ufs/super.c
699
static u64 ufs_max_bytes(struct super_block *sb)
fs/ufs/super.c
701
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
fs/ufs/super.c
716
static int ufs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/ufs/super.c
740
if (!sb_rdonly(sb)) {
fs/ufs/super.c
749
sb->s_fs_info = sbi;
fs/ufs/super.c
750
sbi->sb = sb;
fs/ufs/super.c
752
UFSD("flag %u\n", (int)(sb_rdonly(sb)));
fs/ufs/super.c
778
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/ufs/super.c
780
sb->s_time_gran = NSEC_PER_SEC;
fs/ufs/super.c
781
sb->s_time_min = S32_MIN;
fs/ufs/super.c
782
sb->s_time_max = S32_MAX;
fs/ufs/super.c
802
sb->s_time_gran = 1;
fs/ufs/super.c
803
sb->s_time_min = S64_MIN;
fs/ufs/super.c
804
sb->s_time_max = S64_MAX;
fs/ufs/super.c
850
if (!sb_rdonly(sb)) {
fs/ufs/super.c
853
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
866
if (!sb_rdonly(sb)) {
fs/ufs/super.c
869
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
882
if (!sb_rdonly(sb)) {
fs/ufs/super.c
885
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
898
if (!sb_rdonly(sb)) {
fs/ufs/super.c
901
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
913
if (!sb_rdonly(sb)) {
fs/ufs/super.c
916
sb->s_flags |= SB_RDONLY;
fs/ufs/super.c
926
if (!sb_set_blocksize(sb, block_size)) {
fs/ufs/super.c
935
ubh = ubh_bread_uspi(uspi, sb, uspi->s_sbbase + super_block_offset/block_size, super_block_size);
fs/ufs/super.c
945
uspi->s_postblformat = fs32_to_cpu(sb, usb3->fs_postblformat);
fs/ufs/super.c
963
switch ((uspi->fs_magic = fs32_to_cpu(sb, usb3->fs_magic))) {
fs/ufs/super.c
973
switch ((uspi->fs_magic = fs32_to_cpu(sb, usb3->fs_magic))) {
fs/ufs/super.c
99
static struct inode *ufs_nfs_get_inode(struct super_block *sb, u64 ino, u32 generation)
fs/ufs/ufs.h
136
void ufs_mark_sb_dirty(struct super_block *sb);
fs/ufs/ufs.h
138
static inline struct ufs_sb_info *UFS_SB(struct super_block *sb)
fs/ufs/ufs.h
140
return sb->s_fs_info;
fs/ufs/ufs.h
29
struct super_block *sb;
fs/ufs/ufs_fs.h
240
((UFS_SB(sb)->s_flags & UFS_CG_SUN) ? \
fs/ufs/ufs_fs.h
522
#define ufs_cg_chkmagic(sb, ucg) \
fs/ufs/ufs_fs.h
523
(fs32_to_cpu((sb), (ucg)->cg_magic) == CG_MAGIC)
fs/ufs/ufs_fs.h
527
#define ufs_ocg_blktot(sb, ucg) fs32_to_cpu((sb), ((struct ufs_old_cylinder_group *)(ucg))->cg_btot)
fs/ufs/ufs_fs.h
528
#define ufs_ocg_blks(sb, ucg, cylno) fs32_to_cpu((sb), ((struct ufs_old_cylinder_group *)(ucg))->cg_b[cylno])
fs/ufs/ufs_fs.h
529
#define ufs_ocg_inosused(sb, ucg) fs32_to_cpu((sb), ((struct ufs_old_cylinder_group *)(ucg))->cg_iused)
fs/ufs/ufs_fs.h
530
#define ufs_ocg_blksfree(sb, ucg) fs32_to_cpu((sb), ((struct ufs_old_cylinder_group *)(ucg))->cg_free)
fs/ufs/ufs_fs.h
531
#define ufs_ocg_chkmagic(sb, ucg) \
fs/ufs/ufs_fs.h
532
(fs32_to_cpu((sb), ((struct ufs_old_cylinder_group *)(ucg))->cg_magic) == CG_MAGIC)
fs/ufs/util.c
136
ufs_get_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi)
fs/ufs/util.c
141
if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
fs/ufs/util.c
142
fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[1]);
fs/ufs/util.c
144
fs32 = fs32_to_cpu(sb, ufsi->i_u1.i_data[0]);
fs/ufs/util.c
145
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/ufs/util.c
163
ufs_set_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi, dev_t dev)
fs/ufs/util.c
167
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/ufs/util.c
180
if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
fs/ufs/util.c
181
ufsi->i_u1.i_data[1] = cpu_to_fs32(sb, fs32);
fs/ufs/util.c
183
ufsi->i_u1.i_data[0] = cpu_to_fs32(sb, fs32);
fs/ufs/util.c
20
struct super_block *sb, u64 fragment, u64 size)
fs/ufs/util.c
36
if (!(ubh->bh[i] = sb_bread(sb, fragment + i)))
fs/ufs/util.c
49
struct super_block *sb, u64 fragment, u64 size)
fs/ufs/util.c
61
if (!(USPI_UBH(uspi)->bh[i] = sb_bread(sb, fragment + i)))
fs/ufs/util.h
104
return fs64_to_cpu(sb, tmp);
fs/ufs/util.h
108
ufs_get_fs_qfmask(struct super_block *sb, struct ufs_super_block_third *usb3)
fs/ufs/util.h
112
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/ufs/util.h
128
return fs64_to_cpu(sb, tmp);
fs/ufs/util.h
132
ufs_get_de_namlen(struct super_block *sb, struct ufs_dir_entry *de)
fs/ufs/util.h
134
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) == UFS_DE_OLD)
fs/ufs/util.h
135
return fs16_to_cpu(sb, de->d_u.d_namlen);
fs/ufs/util.h
141
ufs_set_de_namlen(struct super_block *sb, struct ufs_dir_entry *de, u16 value)
fs/ufs/util.h
143
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) == UFS_DE_OLD)
fs/ufs/util.h
144
de->d_u.d_namlen = cpu_to_fs16(sb, value);
fs/ufs/util.h
150
ufs_set_de_type(struct super_block *sb, struct ufs_dir_entry *de, int mode)
fs/ufs/util.h
152
if ((UFS_SB(sb)->s_flags & UFS_DE_MASK) != UFS_DE_44BSD)
fs/ufs/util.h
186
ufs_get_inode_uid(struct super_block *sb, struct ufs_inode *inode)
fs/ufs/util.h
188
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
fs/ufs/util.h
190
return fs32_to_cpu(sb, inode->ui_u3.ui_44.ui_uid);
fs/ufs/util.h
193
return fs32_to_cpu(sb, inode->ui_u3.ui_sun.ui_uid);
fs/ufs/util.h
196
return fs16_to_cpu(sb, inode->ui_u1.oldids.ui_suid);
fs/ufs/util.h
201
ufs_set_inode_uid(struct super_block *sb, struct ufs_inode *inode, u32 value)
fs/ufs/util.h
203
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
fs/ufs/util.h
205
inode->ui_u3.ui_44.ui_uid = cpu_to_fs32(sb, value);
fs/ufs/util.h
206
inode->ui_u1.oldids.ui_suid = cpu_to_fs16(sb, value);
fs/ufs/util.h
209
inode->ui_u3.ui_sun.ui_uid = cpu_to_fs32(sb, value);
fs/ufs/util.h
214
inode->ui_u1.oldids.ui_suid = cpu_to_fs16(sb, value);
fs/ufs/util.h
220
ufs_get_inode_gid(struct super_block *sb, struct ufs_inode *inode)
fs/ufs/util.h
222
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
fs/ufs/util.h
224
return fs32_to_cpu(sb, inode->ui_u3.ui_44.ui_gid);
fs/ufs/util.h
227
return fs32_to_cpu(sb, inode->ui_u3.ui_sun.ui_gid);
fs/ufs/util.h
230
return fs16_to_cpu(sb, inode->ui_u1.oldids.ui_sgid);
fs/ufs/util.h
235
ufs_set_inode_gid(struct super_block *sb, struct ufs_inode *inode, u32 value)
fs/ufs/util.h
237
switch (UFS_SB(sb)->s_flags & UFS_UID_MASK) {
fs/ufs/util.h
239
inode->ui_u3.ui_44.ui_gid = cpu_to_fs32(sb, value);
fs/ufs/util.h
240
inode->ui_u1.oldids.ui_sgid = cpu_to_fs16(sb, value);
fs/ufs/util.h
243
inode->ui_u3.ui_sun.ui_gid = cpu_to_fs32(sb, value);
fs/ufs/util.h
248
inode->ui_u1.oldids.ui_sgid = cpu_to_fs16(sb, value);
fs/ufs/util.h
260
#define ubh_bread(sb,fragment,size) _ubh_bread_(uspi,sb,fragment,size)
fs/ufs/util.h
32
ufs_get_fs_state(struct super_block *sb, struct ufs_super_block_first *usb1,
fs/ufs/util.h
35
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/ufs/util.h
37
if (fs32_to_cpu(sb, usb3->fs_postblformat) == UFS_42POSTBLFMT)
fs/ufs/util.h
38
return fs32_to_cpu(sb, usb1->fs_u0.fs_sun.fs_state);
fs/ufs/util.h
41
return fs32_to_cpu(sb, usb3->fs_un2.fs_sun.fs_state);
fs/ufs/util.h
43
return fs32_to_cpu(sb, usb1->fs_u1.fs_sunx86.fs_state);
fs/ufs/util.h
46
return fs32_to_cpu(sb, usb3->fs_un2.fs_44.fs_state);
fs/ufs/util.h
51
ufs_set_fs_state(struct super_block *sb, struct ufs_super_block_first *usb1,
fs/ufs/util.h
520
static inline void ufs_fragacct (struct super_block * sb, unsigned blockmap,
fs/ufs/util.h
526
uspi = UFS_SB(sb)->s_uspi;
fs/ufs/util.h
534
fs32_add(sb, &fraglist[fragsize], cnt);
fs/ufs/util.h
539
fs32_add(sb, &fraglist[fragsize], cnt);
fs/ufs/util.h
54
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/ufs/util.h
552
static inline u64 ufs_data_ptr_to_cpu(struct super_block *sb, void *p)
fs/ufs/util.h
554
return UFS_SB(sb)->s_uspi->fs_magic == UFS2_MAGIC ?
fs/ufs/util.h
555
fs64_to_cpu(sb, *(__fs64 *)p) :
fs/ufs/util.h
556
fs32_to_cpu(sb, *(__fs32 *)p);
fs/ufs/util.h
559
static inline void ufs_cpu_to_data_ptr(struct super_block *sb, void *p, u64 val)
fs/ufs/util.h
56
if (fs32_to_cpu(sb, usb3->fs_postblformat) == UFS_42POSTBLFMT) {
fs/ufs/util.h
561
if (UFS_SB(sb)->s_uspi->fs_magic == UFS2_MAGIC)
fs/ufs/util.h
562
*(__fs64 *)p = cpu_to_fs64(sb, val);
fs/ufs/util.h
564
*(__fs32 *)p = cpu_to_fs32(sb, val);
fs/ufs/util.h
57
usb1->fs_u0.fs_sun.fs_state = cpu_to_fs32(sb, value);
fs/ufs/util.h
62
usb3->fs_un2.fs_sun.fs_state = cpu_to_fs32(sb, value);
fs/ufs/util.h
65
usb1->fs_u1.fs_sunx86.fs_state = cpu_to_fs32(sb, value);
fs/ufs/util.h
68
usb3->fs_un2.fs_44.fs_state = cpu_to_fs32(sb, value);
fs/ufs/util.h
74
ufs_get_fs_npsect(struct super_block *sb, struct ufs_super_block_first *usb1,
fs/ufs/util.h
77
if ((UFS_SB(sb)->s_flags & UFS_ST_MASK) == UFS_ST_SUNx86)
fs/ufs/util.h
78
return fs32_to_cpu(sb, usb3->fs_un2.fs_sunx86.fs_npsect);
fs/ufs/util.h
80
return fs32_to_cpu(sb, usb1->fs_u1.fs_sun.fs_npsect);
fs/ufs/util.h
84
ufs_get_fs_qbmask(struct super_block *sb, struct ufs_super_block_third *usb3)
fs/ufs/util.h
88
switch (UFS_SB(sb)->s_flags & UFS_ST_MASK) {
fs/unicode/mkutf8data.c
2429
unsigned int sb = 0xAC00;
fs/unicode/mkutf8data.c
2449
unsigned int si = unichar - sb;
fs/vboxsf/super.c
111
static int vboxsf_fill_super(struct super_block *sb, struct fs_context *fc)
fs/vboxsf/super.c
156
err = super_setup_bdi_name(sb, "vboxsf-%d", sbi->bdi_id);
fs/vboxsf/super.c
159
sb->s_bdi->ra_pages = 0;
fs/vboxsf/super.c
160
sb->s_bdi->io_pages = 0;
fs/vboxsf/super.c
188
sb->s_magic = VBOXSF_SUPER_MAGIC;
fs/vboxsf/super.c
189
sb->s_blocksize = 1024;
fs/vboxsf/super.c
190
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/vboxsf/super.c
191
sb->s_op = &vboxsf_super_ops;
fs/vboxsf/super.c
192
set_default_d_op(sb, &vboxsf_dentry_ops);
fs/vboxsf/super.c
194
iroot = iget_locked(sb, 0);
fs/vboxsf/super.c
208
sb->s_root = droot;
fs/vboxsf/super.c
209
sb->s_fs_info = sbi;
fs/vboxsf/super.c
233
static struct inode *vboxsf_alloc_inode(struct super_block *sb)
fs/vboxsf/super.c
237
sf_i = alloc_inode_sb(sb, vboxsf_inode_cachep, GFP_NOFS);
fs/vboxsf/super.c
258
static void vboxsf_put_super(struct super_block *sb)
fs/vboxsf/super.c
260
struct vboxsf_sbi *sbi = VBOXSF_SBI(sb);
fs/vboxsf/super.c
279
struct super_block *sb = dentry->d_sb;
fs/vboxsf/super.c
285
sbi = VBOXSF_SBI(sb);
fs/vboxsf/utils.c
16
struct inode *vboxsf_new_inode(struct super_block *sb)
fs/vboxsf/utils.c
18
struct vboxsf_sbi *sbi = VBOXSF_SBI(sb);
fs/vboxsf/utils.c
24
inode = new_inode(sb);
fs/vboxsf/vfsmod.h
18
#define VBOXSF_SBI(sb) ((struct vboxsf_sbi *)(sb)->s_fs_info)
fs/vboxsf/vfsmod.h
91
struct inode *vboxsf_new_inode(struct super_block *sb);
fs/xfs/libxfs/xfs_log_rlimit.c
41
struct xfs_sb *sb)
fs/xfs/libxfs/xfs_log_rlimit.c
43
return xfs_sb_is_v5(sb) &&
fs/xfs/libxfs/xfs_log_rlimit.c
44
xfs_sb_has_incompat_feature(sb, XFS_SB_FEAT_INCOMPAT_PARENT);
fs/xfs/libxfs/xfs_sb.c
1092
struct xfs_sb sb;
fs/xfs/libxfs/xfs_sb.c
1120
__xfs_sb_from_disk(&sb, dsb, false);
fs/xfs/libxfs/xfs_sb.c
1121
error = xfs_validate_sb_common(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1124
error = xfs_validate_sb_read(mp, &sb);
fs/xfs/libxfs/xfs_sb.c
1158
struct xfs_sb sb;
fs/xfs/libxfs/xfs_sb.c
1168
__xfs_sb_from_disk(&sb, dsb, false);
fs/xfs/libxfs/xfs_sb.c
1169
error = xfs_validate_sb_common(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1172
error = xfs_validate_sb_write(mp, bp, &sb);
fs/xfs/libxfs/xfs_sb.c
1176
if (!xfs_sb_is_v5(&sb))
fs/xfs/scrub/agheader.c
104
struct xfs_dsb *sb;
fs/xfs/scrub/agheader.c
147
sb = bp->b_addr;
fs/xfs/scrub/agheader.c
155
if (sb->sb_blocksize != cpu_to_be32(mp->m_sb.sb_blocksize))
fs/xfs/scrub/agheader.c
158
if (sb->sb_dblocks != cpu_to_be64(mp->m_sb.sb_dblocks))
fs/xfs/scrub/agheader.c
161
if (sb->sb_rblocks != cpu_to_be64(mp->m_sb.sb_rblocks))
fs/xfs/scrub/agheader.c
164
if (sb->sb_rextents != cpu_to_be64(mp->m_sb.sb_rextents))
fs/xfs/scrub/agheader.c
167
if (!uuid_equal(&sb->sb_uuid, &mp->m_sb.sb_uuid))
fs/xfs/scrub/agheader.c
170
if (sb->sb_logstart != cpu_to_be64(mp->m_sb.sb_logstart))
fs/xfs/scrub/agheader.c
173
if (sb->sb_rootino != cpu_to_be64(mp->m_sb.sb_rootino))
fs/xfs/scrub/agheader.c
177
if (sb->sb_rbmino != cpu_to_be64(0))
fs/xfs/scrub/agheader.c
180
if (sb->sb_rsumino != cpu_to_be64(0))
fs/xfs/scrub/agheader.c
183
if (sb->sb_rbmino != cpu_to_be64(mp->m_sb.sb_rbmino))
fs/xfs/scrub/agheader.c
186
if (sb->sb_rsumino != cpu_to_be64(mp->m_sb.sb_rsumino))
fs/xfs/scrub/agheader.c
190
if (sb->sb_rextsize != cpu_to_be32(mp->m_sb.sb_rextsize))
fs/xfs/scrub/agheader.c
193
if (sb->sb_agblocks != cpu_to_be32(mp->m_sb.sb_agblocks))
fs/xfs/scrub/agheader.c
196
if (sb->sb_agcount != cpu_to_be32(mp->m_sb.sb_agcount))
fs/xfs/scrub/agheader.c
199
if (sb->sb_rbmblocks != cpu_to_be32(mp->m_sb.sb_rbmblocks))
fs/xfs/scrub/agheader.c
202
if (sb->sb_logblocks != cpu_to_be32(mp->m_sb.sb_logblocks))
fs/xfs/scrub/agheader.c
214
if ((sb->sb_versionnum & vernum_mask) !=
fs/xfs/scrub/agheader.c
222
if ((sb->sb_versionnum & vernum_mask) !=
fs/xfs/scrub/agheader.c
226
if (sb->sb_sectsize != cpu_to_be16(mp->m_sb.sb_sectsize))
fs/xfs/scrub/agheader.c
229
if (sb->sb_inodesize != cpu_to_be16(mp->m_sb.sb_inodesize))
fs/xfs/scrub/agheader.c
232
if (sb->sb_inopblock != cpu_to_be16(mp->m_sb.sb_inopblock))
fs/xfs/scrub/agheader.c
235
if (memcmp(sb->sb_fname, mp->m_sb.sb_fname, sizeof(sb->sb_fname)))
fs/xfs/scrub/agheader.c
238
if (sb->sb_blocklog != mp->m_sb.sb_blocklog)
fs/xfs/scrub/agheader.c
241
if (sb->sb_sectlog != mp->m_sb.sb_sectlog)
fs/xfs/scrub/agheader.c
244
if (sb->sb_inodelog != mp->m_sb.sb_inodelog)
fs/xfs/scrub/agheader.c
247
if (sb->sb_inopblog != mp->m_sb.sb_inopblog)
fs/xfs/scrub/agheader.c
250
if (sb->sb_agblklog != mp->m_sb.sb_agblklog)
fs/xfs/scrub/agheader.c
253
if (sb->sb_rextslog != mp->m_sb.sb_rextslog)
fs/xfs/scrub/agheader.c
256
if (sb->sb_imax_pct != mp->m_sb.sb_imax_pct)
fs/xfs/scrub/agheader.c
265
if (sb->sb_uquotino != cpu_to_be64(0))
fs/xfs/scrub/agheader.c
268
if (sb->sb_gquotino != cpu_to_be64(0))
fs/xfs/scrub/agheader.c
271
if (sb->sb_uquotino != cpu_to_be64(mp->m_sb.sb_uquotino))
fs/xfs/scrub/agheader.c
274
if (sb->sb_gquotino != cpu_to_be64(mp->m_sb.sb_gquotino))
fs/xfs/scrub/agheader.c
283
if (sb->sb_flags != mp->m_sb.sb_flags)
fs/xfs/scrub/agheader.c
286
if (sb->sb_shared_vn != mp->m_sb.sb_shared_vn)
fs/xfs/scrub/agheader.c
289
if (sb->sb_inoalignmt != cpu_to_be32(mp->m_sb.sb_inoalignmt))
fs/xfs/scrub/agheader.c
292
if (sb->sb_unit != cpu_to_be32(mp->m_sb.sb_unit))
fs/xfs/scrub/agheader.c
295
if (sb->sb_width != cpu_to_be32(mp->m_sb.sb_width))
fs/xfs/scrub/agheader.c
298
if (sb->sb_dirblklog != mp->m_sb.sb_dirblklog)
fs/xfs/scrub/agheader.c
301
if (sb->sb_logsectlog != mp->m_sb.sb_logsectlog)
fs/xfs/scrub/agheader.c
304
if (sb->sb_logsectsize != cpu_to_be16(mp->m_sb.sb_logsectsize))
fs/xfs/scrub/agheader.c
307
if (sb->sb_logsunit != cpu_to_be32(mp->m_sb.sb_logsunit))
fs/xfs/scrub/agheader.c
312
if (sb->sb_features2 != 0)
fs/xfs/scrub/agheader.c
319
if (!!(sb->sb_features2 & cpu_to_be32(~v2_ok)))
fs/xfs/scrub/agheader.c
322
if (sb->sb_features2 != sb->sb_bad_features2)
fs/xfs/scrub/agheader.c
331
if ((sb->sb_features2 & features_mask) !=
fs/xfs/scrub/agheader.c
337
if ((sb->sb_features2 & features_mask) !=
fs/xfs/scrub/agheader.c
343
if (memchr_inv(&sb->sb_features_compat, 0,
fs/xfs/scrub/agheader.c
349
if (sb->sb_features_compat !=
fs/xfs/scrub/agheader.c
354
if (sb->sb_features_ro_compat !=
fs/xfs/scrub/agheader.c
364
sb->sb_features_incompat) & features_mask)
fs/xfs/scrub/agheader.c
369
sb->sb_features_incompat) & ~features_mask)
fs/xfs/scrub/agheader.c
377
if (sb->sb_features_log_incompat)
fs/xfs/scrub/agheader.c
382
if (sb->sb_spino_align != cpu_to_be32(mp->m_sb.sb_spino_align))
fs/xfs/scrub/agheader.c
386
if (sb->sb_pquotino != cpu_to_be64(0))
fs/xfs/scrub/agheader.c
389
if (sb->sb_pquotino != cpu_to_be64(mp->m_sb.sb_pquotino))
fs/xfs/scrub/agheader.c
398
if (!uuid_equal(&sb->sb_meta_uuid, &mp->m_sb.sb_meta_uuid))
fs/xfs/scrub/agheader.c
403
if (sb->sb_metadirino != cpu_to_be64(mp->m_sb.sb_metadirino))
fs/xfs/scrub/agheader.c
406
if (sb->sb_rgcount != cpu_to_be32(mp->m_sb.sb_rgcount))
fs/xfs/scrub/agheader.c
409
if (sb->sb_rgextents != cpu_to_be32(mp->m_sb.sb_rgextents))
fs/xfs/scrub/agheader.c
412
if (sb->sb_rgblklog != mp->m_sb.sb_rgblklog)
fs/xfs/scrub/agheader.c
415
if (memchr_inv(sb->sb_pad, 0, sizeof(sb->sb_pad)))
fs/xfs/scrub/agheader.c
421
if (memchr_inv((char *)sb + sblen, 0, BBTOB(bp->b_length) - sblen))
fs/xfs/scrub/agheader_repair.c
71
struct xfs_dsb *sb = bp->b_addr;
fs/xfs/scrub/agheader_repair.c
73
sb->sb_features_incompat &=
fs/xfs/scrub/agheader_repair.c
75
sb->sb_features_log_incompat = 0;
fs/xfs/scrub/refcount_repair.c
305
uint32_t sa, sb;
fs/xfs/scrub/refcount_repair.c
308
sb = xrep_refc_encode_startblock(bp);
fs/xfs/scrub/refcount_repair.c
310
if (sa > sb)
fs/xfs/scrub/refcount_repair.c
312
if (sa < sb)
fs/xfs/scrub/rtrefcount_repair.c
293
uint32_t sa, sb;
fs/xfs/scrub/rtrefcount_repair.c
296
sb = xrep_rtrefc_encode_startblock(bp);
fs/xfs/scrub/rtrefcount_repair.c
298
if (sa > sb)
fs/xfs/scrub/rtrefcount_repair.c
300
if (sa < sb)
fs/xfs/xfs_attr_list.c
30
xfs_attr_sf_sort_t *sa, *sb;
fs/xfs/xfs_attr_list.c
33
sb = (xfs_attr_sf_sort_t *)b;
fs/xfs/xfs_attr_list.c
34
if (sa->hash < sb->hash) {
fs/xfs/xfs_attr_list.c
36
} else if (sa->hash > sb->hash) {
fs/xfs/xfs_attr_list.c
39
return sa->entno - sb->entno;
fs/xfs/xfs_export.c
107
struct super_block *sb,
fs/xfs/xfs_export.c
111
xfs_mount_t *mp = XFS_M(sb);
fs/xfs/xfs_export.c
172
xfs_fs_fh_to_dentry(struct super_block *sb, struct fid *fid,
fs/xfs/xfs_export.c
184
inode = xfs_nfs_get_inode(sb, fid->i32.ino, fid->i32.gen);
fs/xfs/xfs_export.c
188
inode = xfs_nfs_get_inode(sb, fid64->ino, fid64->gen);
fs/xfs/xfs_export.c
196
xfs_fs_fh_to_parent(struct super_block *sb, struct fid *fid,
fs/xfs/xfs_export.c
207
inode = xfs_nfs_get_inode(sb, fid->i32.parent_ino,
fs/xfs/xfs_export.c
211
inode = xfs_nfs_get_inode(sb, fid64->parent_ino,
fs/xfs/xfs_export.h
60
struct inode *xfs_nfs_get_inode(struct super_block *sb, u64 ino, u32 gen);
fs/xfs/xfs_notify_failure.c
132
struct super_block *sb = mp->m_super;
fs/xfs/xfs_notify_failure.c
135
error = freeze_super(sb, FREEZE_HOLDER_KERNEL, NULL);
fs/xfs/xfs_notify_failure.c
147
struct super_block *sb = mp->m_super;
fs/xfs/xfs_notify_failure.c
151
error = thaw_super(sb, FREEZE_HOLDER_KERNEL, NULL);
fs/xfs/xfs_notify_failure.c
161
thaw_super(sb, FREEZE_HOLDER_USERSPACE, NULL);
fs/xfs/xfs_pnfs.c
54
struct super_block *sb,
fs/xfs/xfs_pnfs.c
59
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_pnfs.h
6
int xfs_fs_get_uuid(struct super_block *sb, u8 *buf, u32 *len, u64 *offset);
fs/xfs/xfs_quotaops.c
117
struct super_block *sb,
fs/xfs/xfs_quotaops.c
121
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
124
if (sb_rdonly(sb))
fs/xfs/xfs_quotaops.c
167
struct super_block *sb,
fs/xfs/xfs_quotaops.c
170
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
172
if (sb_rdonly(sb))
fs/xfs/xfs_quotaops.c
182
struct super_block *sb,
fs/xfs/xfs_quotaops.c
185
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
187
if (sb_rdonly(sb))
fs/xfs/xfs_quotaops.c
197
struct super_block *sb,
fs/xfs/xfs_quotaops.c
200
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
203
if (sb_rdonly(sb))
fs/xfs/xfs_quotaops.c
224
struct super_block *sb,
fs/xfs/xfs_quotaops.c
228
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
241
struct super_block *sb,
fs/xfs/xfs_quotaops.c
246
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
265
struct super_block *sb,
fs/xfs/xfs_quotaops.c
269
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_quotaops.c
271
if (sb_rdonly(sb))
fs/xfs/xfs_quotaops.c
58
struct super_block *sb,
fs/xfs/xfs_quotaops.c
61
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
1027
struct super_block *sb)
fs/xfs/xfs_super.c
1029
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
1219
struct super_block *sb)
fs/xfs/xfs_super.c
1221
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
1239
struct super_block *sb,
fs/xfs/xfs_super.c
1243
if (WARN_ON_ONCE(!sb->s_fs_info))
fs/xfs/xfs_super.c
1245
return xfs_reclaim_inodes_count(XFS_M(sb));
fs/xfs/xfs_super.c
1250
struct super_block *sb,
fs/xfs/xfs_super.c
1253
return xfs_reclaim_inodes_nr(XFS_M(sb), sc->nr_to_scan);
fs/xfs/xfs_super.c
1258
struct super_block *sb)
fs/xfs/xfs_super.c
1260
xfs_force_shutdown(XFS_M(sb), SHUTDOWN_DEVICE_REMOVED);
fs/xfs/xfs_super.c
1653
struct super_block *sb,
fs/xfs/xfs_super.c
1656
struct xfs_mount *mp = sb->s_fs_info;
fs/xfs/xfs_super.c
1660
mp->m_super = sb;
fs/xfs/xfs_super.c
1678
if (!sb_min_blocksize(sb, BBSIZE)) {
fs/xfs/xfs_super.c
1682
sb->s_xattr = xfs_xattr_handlers;
fs/xfs/xfs_super.c
1683
sb->s_export_op = &xfs_export_operations;
fs/xfs/xfs_super.c
1685
sb->s_qcop = &xfs_quotactl_operations;
fs/xfs/xfs_super.c
1686
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP | QTYPE_MASK_PRJ;
fs/xfs/xfs_super.c
1688
sb->s_op = &xfs_super_operations;
fs/xfs/xfs_super.c
1863
sb->s_magic = XFS_SUPER_MAGIC;
fs/xfs/xfs_super.c
1864
sb->s_blocksize = mp->m_sb.sb_blocksize;
fs/xfs/xfs_super.c
1865
sb->s_blocksize_bits = ffs(sb->s_blocksize) - 1;
fs/xfs/xfs_super.c
1866
sb->s_maxbytes = MAX_LFS_FILESIZE;
fs/xfs/xfs_super.c
1867
sb->s_max_links = XFS_MAXLINK;
fs/xfs/xfs_super.c
1868
sb->s_time_gran = 1;
fs/xfs/xfs_super.c
1870
sb->s_time_min = xfs_bigtime_to_unix(XFS_BIGTIME_TIME_MIN);
fs/xfs/xfs_super.c
1871
sb->s_time_max = xfs_bigtime_to_unix(XFS_BIGTIME_TIME_MAX);
fs/xfs/xfs_super.c
1873
sb->s_time_min = XFS_LEGACY_TIME_MIN;
fs/xfs/xfs_super.c
1874
sb->s_time_max = XFS_LEGACY_TIME_MAX;
fs/xfs/xfs_super.c
1876
trace_xfs_inode_timestamp_range(mp, sb->s_time_min, sb->s_time_max);
fs/xfs/xfs_super.c
1877
sb->s_iflags |= SB_I_CGROUPWB | SB_I_ALLOW_HSM;
fs/xfs/xfs_super.c
1879
set_posix_acl_flag(sb);
fs/xfs/xfs_super.c
1883
sb->s_flags |= SB_I_VERSION;
fs/xfs/xfs_super.c
1891
if (xfs_has_discard(mp) && !bdev_max_discard_sectors(sb->s_bdev)) {
fs/xfs/xfs_super.c
1947
sb->s_root = d_make_root(root);
fs/xfs/xfs_super.c
1948
if (!sb->s_root) {
fs/xfs/xfs_super.c
2282
struct super_block *sb)
fs/xfs/xfs_super.c
2284
kill_block_super(sb);
fs/xfs/xfs_super.c
2285
xfs_mount_free(XFS_M(sb));
fs/xfs/xfs_super.c
470
struct super_block *sb = mp->m_super;
fs/xfs/xfs_super.c
471
struct block_device *ddev = sb->s_bdev;
fs/xfs/xfs_super.c
502
mp->m_ddev_targp = xfs_alloc_buftarg(mp, sb->s_bdev_file);
fs/xfs/xfs_super.c
664
struct super_block *sb = mp->m_super;
fs/xfs/xfs_super.c
666
if (down_read_trylock(&sb->s_umount)) {
fs/xfs/xfs_super.c
667
sync_inodes_sb(sb);
fs/xfs/xfs_super.c
668
up_read(&sb->s_umount);
fs/xfs/xfs_super.c
696
struct super_block *sb)
fs/xfs/xfs_super.c
810
struct super_block *sb,
fs/xfs/xfs_super.c
813
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.c
842
if (sb->s_writers.frozen == SB_FREEZE_PAGEFAULT) {
fs/xfs/xfs_super.c
993
struct super_block *sb)
fs/xfs/xfs_super.c
995
struct xfs_mount *mp = XFS_M(sb);
fs/xfs/xfs_super.h
100
#define XFS_M(sb) ((struct xfs_mount *)((sb)->s_fs_info))
fs/xfs/xfs_super.h
23
# define set_posix_acl_flag(sb) ((sb)->s_flags |= SB_POSIXACL)
fs/xfs/xfs_super.h
26
# define set_posix_acl_flag(sb) do { } while (0)
fs/zonefs/file.c
33
struct super_block *sb = inode->i_sb;
fs/zonefs/file.c
42
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
fs/zonefs/file.c
475
struct super_block *sb = inode->i_sb;
fs/zonefs/file.c
500
if ((iocb->ki_pos | count) & (sb->s_blocksize - 1)) {
fs/zonefs/file.c
637
struct super_block *sb = inode->i_sb;
fs/zonefs/file.c
669
if ((iocb->ki_pos | count) & (sb->s_blocksize - 1)) {
fs/zonefs/file.c
70
struct super_block *sb = inode->i_sb;
fs/zonefs/file.c
802
struct super_block *sb = inode->i_sb;
fs/zonefs/file.c
803
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/file.c
828
!(sb->s_flags & SB_RDONLY)) {
fs/zonefs/file.c
829
zonefs_warn(sb,
fs/zonefs/file.c
832
zonefs_warn(sb,
fs/zonefs/file.c
834
sb->s_flags |= SB_RDONLY;
fs/zonefs/file.c
92
iomap->offset = ALIGN_DOWN(offset, sb->s_blocksize);
fs/zonefs/super.c
1002
bdev_zone_sectors(sb->s_bdev) << SECTOR_SHIFT);
fs/zonefs/super.c
1008
z->z_wpoffset = zonefs_check_zone_condition(sb, z, zone);
fs/zonefs/super.c
1021
sb->s_maxbytes = max(z->z_capacity, sb->s_maxbytes);
fs/zonefs/super.c
1022
sbi->s_blocks += z->z_capacity >> sb->s_blocksize_bits;
fs/zonefs/super.c
1023
sbi->s_used_blocks += z->z_wpoffset >> sb->s_blocksize_bits;
fs/zonefs/super.c
1034
ret = zonefs_zone_mgmt(sb, z, REQ_OP_ZONE_CLOSE);
fs/zonefs/super.c
1039
zonefs_account_active(sb, z);
fs/zonefs/super.c
1047
zonefs_info(sb, "Zone group \"%s\" has %u file%s\n",
fs/zonefs/super.c
1055
static void zonefs_free_zgroups(struct super_block *sb)
fs/zonefs/super.c
1057
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
1072
static int zonefs_init_zgroups(struct super_block *sb)
fs/zonefs/super.c
1080
zd.sb = sb;
fs/zonefs/super.c
1087
ret = zonefs_init_zgroup(sb, &zd, ztype);
fs/zonefs/super.c
1089
zonefs_info(sb,
fs/zonefs/super.c
1099
zonefs_free_zgroups(sb);
fs/zonefs/super.c
1107
static int zonefs_read_super(struct super_block *sb)
fs/zonefs/super.c
1109
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
1118
ret = bdev_rw_virt(sb->s_bdev, 0, super, ZONEFS_SUPER_SIZE,
fs/zonefs/super.c
1131
zonefs_err(sb, "Invalid checksum (Expected 0x%08x, got 0x%08x)",
fs/zonefs/super.c
1138
zonefs_err(sb, "Unknown features set 0x%llx\n",
fs/zonefs/super.c
1147
zonefs_err(sb, "Invalid UID feature\n");
fs/zonefs/super.c
115
trace_zonefs_zone_mgmt(sb, z, op);
fs/zonefs/super.c
1156
zonefs_err(sb, "Invalid GID feature\n");
fs/zonefs/super.c
116
ret = blkdev_zone_mgmt(sb->s_bdev, op, z->z_sector,
fs/zonefs/super.c
1165
zonefs_err(sb, "Reserved area is being used\n");
fs/zonefs/super.c
1184
static int zonefs_get_zgroup_inodes(struct super_block *sb)
fs/zonefs/super.c
1186
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
119
zonefs_err(sb,
fs/zonefs/super.c
1194
dir_inode = zonefs_get_zgroup_inode(sb, ztype);
fs/zonefs/super.c
1204
static void zonefs_release_zgroup_inodes(struct super_block *sb)
fs/zonefs/super.c
1206
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
1225
static int zonefs_fill_super(struct super_block *sb, struct fs_context *fc)
fs/zonefs/super.c
1233
if (!bdev_is_zoned(sb->s_bdev)) {
fs/zonefs/super.c
1234
zonefs_err(sb, "Not a zoned block device\n");
fs/zonefs/super.c
1249
sb->s_fs_info = sbi;
fs/zonefs/super.c
1250
sb->s_magic = ZONEFS_MAGIC;
fs/zonefs/super.c
1251
sb->s_maxbytes = 0;
fs/zonefs/super.c
1252
sb->s_op = &zonefs_sops;
fs/zonefs/super.c
1253
sb->s_time_gran = 1;
fs/zonefs/super.c
1260
sb_set_blocksize(sb, bdev_zone_write_granularity(sb->s_bdev));
fs/zonefs/super.c
1261
sbi->s_zone_sectors_shift = ilog2(bdev_zone_sectors(sb->s_bdev));
fs/zonefs/super.c
1268
sbi->s_max_wro_seq_files = bdev_max_open_zones(sb->s_bdev);
fs/zonefs/super.c
1270
sbi->s_max_active_seq_files = bdev_max_active_zones(sb->s_bdev);
fs/zonefs/super.c
1272
ret = zonefs_read_super(sb);
fs/zonefs/super.c
1276
zonefs_info(sb, "Mounting %u zones", bdev_nr_zones(sb->s_bdev));
fs/zonefs/super.c
1281
zonefs_info(sb,
fs/zonefs/super.c
1287
ret = zonefs_init_zgroups(sb);
fs/zonefs/super.c
1293
inode = new_inode(sb);
fs/zonefs/super.c
1297
inode->i_ino = bdev_nr_zones(sb->s_bdev);
fs/zonefs/super.c
1311
sb->s_root = d_make_root(inode);
fs/zonefs/super.c
1312
if (!sb->s_root)
fs/zonefs/super.c
1319
ret = zonefs_get_zgroup_inodes(sb);
fs/zonefs/super.c
1323
ret = zonefs_sysfs_register(sb);
fs/zonefs/super.c
1330
zonefs_release_zgroup_inodes(sb);
fs/zonefs/super.c
1331
zonefs_free_zgroups(sb);
fs/zonefs/super.c
1336
static void zonefs_kill_super(struct super_block *sb)
fs/zonefs/super.c
1338
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
1341
zonefs_release_zgroup_inodes(sb);
fs/zonefs/super.c
1343
kill_block_super(sb);
fs/zonefs/super.c
1345
zonefs_sysfs_unregister(sb);
fs/zonefs/super.c
1346
zonefs_free_zgroups(sb);
fs/zonefs/super.c
1365
struct super_block *sb = fc->root->d_sb;
fs/zonefs/super.c
1366
struct zonefs_sb_info *sbi = sb->s_fs_info;
fs/zonefs/super.c
156
struct super_block *sb = inode->i_sb;
fs/zonefs/super.c
157
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
171
nr_blocks = (old_isize - new_isize) >> sb->s_blocksize_bits;
fs/zonefs/super.c
178
(new_isize - old_isize) >> sb->s_blocksize_bits;
fs/zonefs/super.c
190
static loff_t zonefs_check_zone_condition(struct super_block *sb,
fs/zonefs/super.c
196
zonefs_warn(sb, "Zone %llu: offline zone\n",
fs/zonefs/super.c
209
zonefs_warn(sb, "Zone %llu: read-only zone\n",
fs/zonefs/super.c
263
struct super_block *sb = inode->i_sb;
fs/zonefs/super.c
264
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
273
data_size = zonefs_check_zone_condition(sb, z, zone);
fs/zonefs/super.c
299
zonefs_warn(sb,
fs/zonefs/super.c
311
zonefs_warn(sb, "inode %lu: read/write access disabled\n",
fs/zonefs/super.c
319
zonefs_warn(sb, "inode %lu: write access disabled\n",
fs/zonefs/super.c
345
if ((sbi->s_mount_opts & ZONEFS_MNTOPT_ERRORS_RO) && !sb_rdonly(sb)) {
fs/zonefs/super.c
346
zonefs_warn(sb, "remounting filesystem read-only\n");
fs/zonefs/super.c
347
sb->s_flags |= SB_RDONLY;
fs/zonefs/super.c
370
struct super_block *sb = inode->i_sb;
fs/zonefs/super.c
400
ret = blkdev_report_zones(sb->s_bdev, z->z_sector, 1,
fs/zonefs/super.c
405
zonefs_err(sb, "Get inode %lu zone information failed %d\n",
fs/zonefs/super.c
407
zonefs_warn(sb, "remounting filesystem read-only\n");
fs/zonefs/super.c
408
sb->s_flags |= SB_RDONLY;
fs/zonefs/super.c
418
static struct inode *zonefs_alloc_inode(struct super_block *sb)
fs/zonefs/super.c
422
zi = alloc_inode_sb(sb, zonefs_inode_cachep, GFP_KERNEL);
fs/zonefs/super.c
443
struct super_block *sb = dentry->d_sb;
fs/zonefs/super.c
444
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
448
buf->f_bsize = sb->s_blocksize;
fs/zonefs/super.c
50
static void zonefs_account_active(struct super_block *sb,
fs/zonefs/super.c
53
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
627
struct super_block *sb = dir->i_sb;
fs/zonefs/super.c
628
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
644
inode = iget_locked(sb, ino);
fs/zonefs/super.c
675
static struct inode *zonefs_get_zgroup_inode(struct super_block *sb,
fs/zonefs/super.c
678
struct inode *root = d_inode(sb->s_root);
fs/zonefs/super.c
679
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
681
ino_t ino = bdev_nr_zones(sb->s_bdev) + ztype + 1;
fs/zonefs/super.c
683
inode = iget_locked(sb, ino);
fs/zonefs/super.c
709
struct super_block *sb = dir->i_sb;
fs/zonefs/super.c
710
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
729
return zonefs_get_zgroup_inode(sb, ztype);
fs/zonefs/super.c
751
struct super_block *sb = inode->i_sb;
fs/zonefs/super.c
752
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
754
ino_t base_ino = bdev_nr_zones(sb->s_bdev) + 1;
fs/zonefs/super.c
788
struct super_block *sb = inode->i_sb;
fs/zonefs/super.c
789
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
847
struct super_block *sb;
fs/zonefs/super.c
857
struct super_block *sb = zd->sb;
fs/zonefs/super.c
858
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
891
zonefs_err(zd->sb, "Unsupported zone type 0x%x\n",
fs/zonefs/super.c
903
struct block_device *bdev = zd->sb->s_bdev;
fs/zonefs/super.c
914
zonefs_err(zd->sb, "Zone report failed %d\n", ret);
fs/zonefs/super.c
919
zonefs_err(zd->sb, "Invalid zone report (%d/%u zones)\n",
fs/zonefs/super.c
935
static int zonefs_init_zgroup(struct super_block *sb,
fs/zonefs/super.c
939
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/super.c
959
end = zd->zones + bdev_nr_zones(sb->s_bdev);
fs/zonefs/super.c
99
static int zonefs_zone_mgmt(struct super_block *sb,
fs/zonefs/super.c
997
if (z->z_size > bdev_zone_sectors(sb->s_bdev) << SECTOR_SHIFT &&
fs/zonefs/super.c
999
zonefs_err(sb,
fs/zonefs/sysfs.c
110
void zonefs_sysfs_unregister(struct super_block *sb)
fs/zonefs/sysfs.c
112
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/sysfs.c
90
int zonefs_sysfs_register(struct super_block *sb)
fs/zonefs/sysfs.c
92
struct zonefs_sb_info *sbi = ZONEFS_SB(sb);
fs/zonefs/sysfs.c
95
super_set_sysfs_name_id(sb);
fs/zonefs/sysfs.c
98
zonefs_sysfs_root, "%s", sb->s_id);
fs/zonefs/trace.h
23
TP_PROTO(struct super_block *sb, struct zonefs_zone *z,
fs/zonefs/trace.h
25
TP_ARGS(sb, z, op),
fs/zonefs/trace.h
34
__entry->dev = sb->s_dev;
fs/zonefs/trace.h
36
z->z_sector >> ZONEFS_SB(sb)->s_zone_sectors_shift;
fs/zonefs/zonefs.h
246
static inline struct zonefs_sb_info *ZONEFS_SB(struct super_block *sb)
fs/zonefs/zonefs.h
248
return sb->s_fs_info;
fs/zonefs/zonefs.h
251
#define zonefs_info(sb, format, args...) \
fs/zonefs/zonefs.h
252
pr_info("zonefs (%s): " format, sb->s_id, ## args)
fs/zonefs/zonefs.h
253
#define zonefs_err(sb, format, args...) \
fs/zonefs/zonefs.h
254
pr_err("zonefs (%s) ERROR: " format, sb->s_id, ## args)
fs/zonefs/zonefs.h
255
#define zonefs_warn(sb, format, args...) \
fs/zonefs/zonefs.h
256
pr_warn("zonefs (%s) WARNING: " format, sb->s_id, ## args)
fs/zonefs/zonefs.h
284
int zonefs_sysfs_register(struct super_block *sb);
fs/zonefs/zonefs.h
285
void zonefs_sysfs_unregister(struct super_block *sb);
include/drm/drm_fb_dma_helper.h
25
struct drm_scanout_buffer *sb);
include/drm/drm_modeset_helper_vtables.h
1463
struct drm_scanout_buffer *sb);
include/drm/drm_panic.h
72
void (*set_pixel)(struct drm_scanout_buffer *sb, unsigned int x,
include/drm/intel/display_parent_interface.h
49
int (*setup)(struct intel_panic *panic, struct drm_scanout_buffer *sb);
include/linux/blkdev.h
1277
static inline int sb_issue_discard(struct super_block *sb, sector_t block,
include/linux/blkdev.h
1280
return blkdev_issue_discard(sb->s_bdev,
include/linux/blkdev.h
1281
block << (sb->s_blocksize_bits -
include/linux/blkdev.h
1283
nr_blocks << (sb->s_blocksize_bits -
include/linux/blkdev.h
1287
static inline int sb_issue_zeroout(struct super_block *sb, sector_t block,
include/linux/blkdev.h
1290
return blkdev_issue_zeroout(sb->s_bdev,
include/linux/blkdev.h
1291
block << (sb->s_blocksize_bits -
include/linux/blkdev.h
1293
nr_blocks << (sb->s_blocksize_bits -
include/linux/blkdev.h
850
static inline u64 sb_bdev_nr_blocks(struct super_block *sb)
include/linux/blkdev.h
852
return bdev_nr_sectors(sb->s_bdev) >>
include/linux/blkdev.h
853
(sb->s_blocksize_bits - SECTOR_SHIFT);
include/linux/bpf.h
2793
struct inode *bpf_get_inode(struct super_block *sb, const struct inode *dir,
include/linux/buffer_head.h
344
sb_bread(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
346
return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, __GFP_MOVABLE);
include/linux/buffer_head.h
350
sb_bread_unmovable(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
352
return __bread_gfp(sb->s_bdev, block, sb->s_blocksize, 0);
include/linux/buffer_head.h
356
sb_breadahead(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
358
__breadahead(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
383
static inline struct buffer_head *sb_getblk(struct super_block *sb,
include/linux/buffer_head.h
386
return __getblk(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
389
static inline struct buffer_head *sb_getblk_gfp(struct super_block *sb,
include/linux/buffer_head.h
392
return bdev_getblk(sb->s_bdev, block, sb->s_blocksize, gfp);
include/linux/buffer_head.h
396
sb_find_get_block(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
398
return __find_get_block(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
402
sb_find_get_block_nonatomic(struct super_block *sb, sector_t block)
include/linux/buffer_head.h
404
return __find_get_block_nonatomic(sb->s_bdev, block, sb->s_blocksize);
include/linux/buffer_head.h
408
map_bh(struct buffer_head *bh, struct super_block *sb, sector_t block)
include/linux/buffer_head.h
411
bh->b_bdev = sb->s_bdev;
include/linux/buffer_head.h
413
bh->b_size = sb->s_blocksize;
include/linux/exportfs.h
284
struct dentry * (*fh_to_dentry)(struct super_block *sb, struct fid *fid,
include/linux/exportfs.h
286
struct dentry * (*fh_to_parent)(struct super_block *sb, struct fid *fid,
include/linux/exportfs.h
293
int (*get_uuid)(struct super_block *sb, u8 *buf, u32 *len, u64 *offset);
include/linux/exportfs.h
400
struct dentry *generic_fh_to_dentry(struct super_block *sb,
include/linux/exportfs.h
402
struct inode *(*get_inode) (struct super_block *sb, u64 ino, u32 gen));
include/linux/exportfs.h
403
struct dentry *generic_fh_to_parent(struct super_block *sb,
include/linux/exportfs.h
405
struct inode *(*get_inode) (struct super_block *sb, u64 ino, u32 gen));
include/linux/fs.h
1581
static inline bool fsuidgid_has_mapping(struct super_block *sb,
include/linux/fs.h
1584
struct user_namespace *fs_userns = sb->s_user_ns;
include/linux/fs.h
2316
void retire_super(struct super_block *sb);
include/linux/fs.h
2317
void generic_shutdown_super(struct super_block *sb);
include/linux/fs.h
2318
void kill_block_super(struct super_block *sb);
include/linux/fs.h
2319
void kill_anon_super(struct super_block *sb);
include/linux/fs.h
2320
void deactivate_super(struct super_block *sb);
include/linux/fs.h
2321
void deactivate_locked_super(struct super_block *sb);
include/linux/fs.h
2365
int super_setup_bdi_name(struct super_block *sb, char *fmt, ...);
include/linux/fs.h
2366
extern int super_setup_bdi(struct super_block *sb);
include/linux/fs.h
2368
static inline void super_set_uuid(struct super_block *sb, const u8 *uuid, unsigned len)
include/linux/fs.h
2370
if (WARN_ON(len > sizeof(sb->s_uuid)))
include/linux/fs.h
2371
len = sizeof(sb->s_uuid);
include/linux/fs.h
2372
sb->s_uuid_len = len;
include/linux/fs.h
2373
memcpy(&sb->s_uuid, uuid, len);
include/linux/fs.h
2377
static inline void super_set_sysfs_name_bdev(struct super_block *sb)
include/linux/fs.h
2379
snprintf(sb->s_sysfs_name, sizeof(sb->s_sysfs_name), "%pg", sb->s_bdev);
include/linux/fs.h
2383
static inline void super_set_sysfs_name_uuid(struct super_block *sb)
include/linux/fs.h
2385
WARN_ON(sb->s_uuid_len != sizeof(sb->s_uuid));
include/linux/fs.h
2386
snprintf(sb->s_sysfs_name, sizeof(sb->s_sysfs_name), "%pU", sb->s_uuid.b);
include/linux/fs.h
2390
static inline void super_set_sysfs_name_id(struct super_block *sb)
include/linux/fs.h
2392
strscpy(sb->s_sysfs_name, sb->s_id, sizeof(sb->s_sysfs_name));
include/linux/fs.h
2397
static inline void super_set_sysfs_name_generic(struct super_block *sb, const char *fmt, ...)
include/linux/fs.h
2402
vsnprintf(sb->s_sysfs_name, sizeof(sb->s_sysfs_name), fmt, args);
include/linux/fs.h
2920
static inline int inode_init_always(struct super_block *sb, struct inode *inode)
include/linux/fs.h
2922
return inode_init_always_gfp(sb, inode, GFP_NOFS);
include/linux/fs.h
2937
extern struct inode *ilookup5_nowait(struct super_block *sb,
include/linux/fs.h
2940
extern struct inode *ilookup5(struct super_block *sb, unsigned long hashval,
include/linux/fs.h
2942
extern struct inode *ilookup(struct super_block *sb, unsigned long ino);
include/linux/fs.h
2973
extern void evict_inodes(struct super_block *sb);
include/linux/fs.h
3000
struct inode *alloc_inode(struct super_block *sb);
include/linux/fs.h
3001
static inline struct inode *new_inode_pseudo(struct super_block *sb)
include/linux/fs.h
3003
return alloc_inode(sb);
include/linux/fs.h
3005
extern struct inode *new_inode(struct super_block *sb);
include/linux/fs.h
3209
extern void drop_super(struct super_block *sb);
include/linux/fs.h
3210
extern void drop_super_exclusive(struct super_block *sb);
include/linux/fs.h
3257
struct inode *anon_inode_make_secure_inode(struct super_block *sb, const char *name,
include/linux/fs.h
3303
extern void generic_set_sb_d_ops(struct super_block *sb);
include/linux/fs/super.h
101
__sb_end_write(sb, SB_FREEZE_FS);
include/linux/fs/super.h
12
static inline void __sb_end_write(struct super_block *sb, int level)
include/linux/fs/super.h
123
static inline void sb_start_write(struct super_block *sb)
include/linux/fs/super.h
125
__sb_start_write(sb, SB_FREEZE_WRITE);
include/linux/fs/super.h
133
static inline bool sb_start_write_trylock(struct super_block *sb)
include/linux/fs/super.h
135
return __sb_start_write_trylock(sb, SB_FREEZE_WRITE);
include/linux/fs/super.h
14
percpu_up_read(sb->s_writers.rw_sem + level - 1);
include/linux/fs/super.h
157
static inline void sb_start_pagefault(struct super_block *sb)
include/linux/fs/super.h
159
__sb_start_write(sb, SB_FREEZE_PAGEFAULT);
include/linux/fs/super.h
17
static inline void __sb_start_write(struct super_block *sb, int level)
include/linux/fs/super.h
175
static inline void sb_start_intwrite(struct super_block *sb)
include/linux/fs/super.h
177
__sb_start_write(sb, SB_FREEZE_FS);
include/linux/fs/super.h
180
static inline bool sb_start_intwrite_trylock(struct super_block *sb)
include/linux/fs/super.h
182
return __sb_start_write_trylock(sb, SB_FREEZE_FS);
include/linux/fs/super.h
185
static inline bool sb_rdonly(const struct super_block *sb)
include/linux/fs/super.h
187
return sb->s_flags & SB_RDONLY;
include/linux/fs/super.h
19
percpu_down_read_freezable(sb->s_writers.rw_sem + level - 1, true);
include/linux/fs/super.h
190
static inline bool sb_is_blkdev_sb(struct super_block *sb)
include/linux/fs/super.h
192
return IS_ENABLED(CONFIG_BLOCK) && sb == blockdev_superblock;
include/linux/fs/super.h
196
static inline struct unicode_map *sb_encoding(const struct super_block *sb)
include/linux/fs/super.h
198
return sb->s_encoding;
include/linux/fs/super.h
213
static inline struct unicode_map *sb_encoding(const struct super_block *sb)
include/linux/fs/super.h
22
static inline bool __sb_start_write_trylock(struct super_block *sb, int level)
include/linux/fs/super.h
225
static inline bool sb_has_encoding(const struct super_block *sb)
include/linux/fs/super.h
227
return !!sb_encoding(sb);
include/linux/fs/super.h
230
int sb_set_blocksize(struct super_block *sb, int size);
include/linux/fs/super.h
231
int __must_check sb_min_blocksize(struct super_block *sb, int size);
include/linux/fs/super.h
24
return percpu_down_read_trylock(sb->s_writers.rw_sem + level - 1);
include/linux/fs/super.h
27
#define __sb_writers_acquired(sb, lev) \
include/linux/fs/super.h
28
percpu_rwsem_acquire(&(sb)->s_writers.rw_sem[(lev) - 1], 1, _THIS_IP_)
include/linux/fs/super.h
29
#define __sb_writers_release(sb, lev) \
include/linux/fs/super.h
30
percpu_rwsem_release(&(sb)->s_writers.rw_sem[(lev) - 1], _THIS_IP_)
include/linux/fs/super.h
41
static inline int __sb_write_started(const struct super_block *sb, int level)
include/linux/fs/super.h
43
return lockdep_is_held_type(sb->s_writers.rw_sem + level - 1, 1);
include/linux/fs/super.h
52
static inline bool sb_write_started(const struct super_block *sb)
include/linux/fs/super.h
54
return __sb_write_started(sb, SB_FREEZE_WRITE);
include/linux/fs/super.h
63
static inline bool sb_write_not_started(const struct super_block *sb)
include/linux/fs/super.h
65
return __sb_write_started(sb, SB_FREEZE_WRITE) <= 0;
include/linux/fs/super.h
75
static inline void sb_end_write(struct super_block *sb)
include/linux/fs/super.h
77
__sb_end_write(sb, SB_FREEZE_WRITE);
include/linux/fs/super.h
87
static inline void sb_end_pagefault(struct super_block *sb)
include/linux/fs/super.h
89
__sb_end_write(sb, SB_FREEZE_PAGEFAULT);
include/linux/fs/super.h
99
static inline void sb_end_intwrite(struct super_block *sb)
include/linux/fs/super_types.h
100
void (*umount_begin)(struct super_block *sb);
include/linux/fs/super_types.h
107
ssize_t (*quota_read)(struct super_block *sb, int type, char *data,
include/linux/fs/super_types.h
109
ssize_t (*quota_write)(struct super_block *sb, int type,
include/linux/fs/super_types.h
113
long (*nr_cached_objects)(struct super_block *sb,
include/linux/fs/super_types.h
115
long (*free_cached_objects)(struct super_block *sb,
include/linux/fs/super_types.h
125
int (*remove_bdev)(struct super_block *sb, struct block_device *bdev);
include/linux/fs/super_types.h
126
void (*shutdown)(struct super_block *sb);
include/linux/fs/super_types.h
312
#define sb_has_strict_encoding(sb) \
include/linux/fs/super_types.h
313
(sb->s_encoding_flags & SB_ENC_STRICT_MODE_FL)
include/linux/fs/super_types.h
316
#define sb_no_casefold_compat_fallback(sb) \
include/linux/fs/super_types.h
317
(sb->s_encoding_flags & SB_ENC_NO_COMPAT_FALLBACK_FL)
include/linux/fs/super_types.h
319
#define sb_no_casefold_compat_fallback(sb) (1)
include/linux/fs/super_types.h
84
struct inode *(*alloc_inode)(struct super_block *sb);
include/linux/fs/super_types.h
91
void (*put_super)(struct super_block *sb);
include/linux/fs/super_types.h
92
int (*sync_fs)(struct super_block *sb, int wait);
include/linux/fs/super_types.h
93
int (*freeze_super)(struct super_block *sb, enum freeze_holder who,
include/linux/fs/super_types.h
95
int (*freeze_fs)(struct super_block *sb);
include/linux/fs/super_types.h
96
int (*thaw_super)(struct super_block *sb, enum freeze_holder who,
include/linux/fs/super_types.h
98
int (*unfreeze_fs)(struct super_block *sb);
include/linux/fs_context.h
154
int (*fill_super)(struct super_block *sb,
include/linux/fs_context.h
157
int (*fill_super)(struct super_block *sb,
include/linux/fs_context.h
160
int (*fill_super)(struct super_block *sb,
include/linux/fs_context.h
164
int setup_bdev_super(struct super_block *sb, int sb_flags,
include/linux/fs_context.h
169
int (*fill_super)(struct super_block *sb,
include/linux/fs_context.h
173
int (*fill_super)(struct super_block *sb,
include/linux/fscrypt.h
160
const union fscrypt_policy *(*get_dummy_policy)(struct super_block *sb);
include/linux/fscrypt.h
181
bool (*has_stable_inodes)(struct super_block *sb);
include/linux/fscrypt.h
197
struct block_device **(*get_devices)(struct super_block *sb,
include/linux/fscrypt.h
400
struct super_block *sb);
include/linux/fscrypt.h
414
void fscrypt_destroy_keyring(struct super_block *sb);
include/linux/fscrypt.h
479
static inline void fscrypt_set_ops(struct super_block *sb,
include/linux/fscrypt.h
482
sb->s_cop = s_cop;
include/linux/fscrypt.h
624
struct super_block *sb)
include/linux/fscrypt.h
640
static inline void fscrypt_destroy_keyring(struct super_block *sb)
include/linux/fscrypt.h
856
static inline void fscrypt_set_ops(struct super_block *sb,
include/linux/fserror.h
10
void fserror_unmount(struct super_block *sb);
include/linux/fserror.h
30
struct super_block *sb;
include/linux/fserror.h
40
void fserror_report(struct super_block *sb, struct inode *inode,
include/linux/fserror.h
64
static inline void fserror_report_metadata(struct super_block *sb, int error,
include/linux/fserror.h
67
fserror_report(sb, NULL, FSERR_METADATA, 0, 0, error, gfp);
include/linux/fserror.h
70
static inline void fserror_report_shutdown(struct super_block *sb, gfp_t gfp)
include/linux/fserror.h
72
fserror_report(sb, NULL, FSERR_METADATA, 0, 0, -ESHUTDOWN, gfp);
include/linux/fserror.h
9
void fserror_mount(struct super_block *sb);
include/linux/fsnotify.h
21
static inline bool fsnotify_sb_has_priority_watchers(struct super_block *sb,
include/linux/fsnotify.h
24
struct fsnotify_sb_info *sbinfo = fsnotify_sb_info(sb);
include/linux/fsnotify.h
34
static inline bool fsnotify_sb_has_watchers(struct super_block *sb)
include/linux/fsnotify.h
36
return fsnotify_sb_has_priority_watchers(sb, 0);
include/linux/fsnotify_backend.h
319
struct super_block *sb;
include/linux/fsnotify_backend.h
397
return ((struct fs_error_report *) data)->sb;
include/linux/fsnotify_backend.h
527
FSNOTIFY_ITER_FUNCS(sb, SB)
include/linux/fsnotify_backend.h
579
static inline struct fsnotify_sb_info *fsnotify_sb_info(struct super_block *sb)
include/linux/fsnotify_backend.h
582
return READ_ONCE(sb->s_fsnotify_info);
include/linux/fsnotify_backend.h
588
static inline atomic_long_t *fsnotify_sb_watched_objects(struct super_block *sb)
include/linux/fsnotify_backend.h
590
return &fsnotify_sb_info(sb)->watched_objects[0];
include/linux/fsnotify_backend.h
655
extern void fsnotify_sb_delete(struct super_block *sb);
include/linux/fsnotify_backend.h
657
extern void fsnotify_sb_free(struct super_block *sb);
include/linux/fsnotify_backend.h
955
static inline void fsnotify_sb_delete(struct super_block *sb)
include/linux/fsnotify_backend.h
961
static inline void fsnotify_sb_free(struct super_block *sb)
include/linux/fsnotify_backend.h
972
static inline void fsnotify_unmount_inodes(struct super_block *sb)
include/linux/hugetlb.h
514
static inline struct hugetlbfs_sb_info *HUGETLBFS_SB(struct super_block *sb)
include/linux/hugetlb.h
516
return sb->s_fs_info;
include/linux/kernfs.h
416
struct kernfs_root *kernfs_root_from_sb(struct super_block *sb);
include/linux/kernfs.h
417
struct inode *kernfs_get_inode(struct super_block *sb, struct kernfs_node *kn);
include/linux/kernfs.h
420
struct super_block *sb);
include/linux/kernfs.h
462
const void *kernfs_super_ns(struct super_block *sb);
include/linux/kernfs.h
465
void kernfs_kill_sb(struct super_block *sb);
include/linux/kernfs.h
510
static inline struct kernfs_root *kernfs_root_from_sb(struct super_block *sb)
include/linux/kernfs.h
514
kernfs_get_inode(struct super_block *sb, struct kernfs_node *kn)
include/linux/kernfs.h
578
static inline const void *kernfs_super_ns(struct super_block *sb)
include/linux/kernfs.h
586
static inline void kernfs_kill_sb(struct super_block *sb) { }
include/linux/lockd/lockd.h
308
int nlmsvc_unlock_all_by_sb(struct super_block *sb);
include/linux/lsm_hook_defs.h
46
LSM_HOOK(int, 0, quotactl, int cmds, int type, int id, const struct super_block *sb)
include/linux/lsm_hook_defs.h
62
LSM_HOOK(int, 0, sb_alloc_security, struct super_block *sb)
include/linux/lsm_hook_defs.h
63
LSM_HOOK(void, LSM_RET_VOID, sb_delete, struct super_block *sb)
include/linux/lsm_hook_defs.h
64
LSM_HOOK(void, LSM_RET_VOID, sb_free_security, struct super_block *sb)
include/linux/lsm_hook_defs.h
67
LSM_HOOK(int, 0, sb_mnt_opts_compat, struct super_block *sb, void *mnt_opts)
include/linux/lsm_hook_defs.h
68
LSM_HOOK(int, 0, sb_remount, struct super_block *sb, void *mnt_opts)
include/linux/lsm_hook_defs.h
69
LSM_HOOK(int, 0, sb_kern_mount, const struct super_block *sb)
include/linux/lsm_hook_defs.h
70
LSM_HOOK(int, 0, sb_show_options, struct seq_file *m, struct super_block *sb)
include/linux/lsm_hook_defs.h
77
LSM_HOOK(int, 0, sb_set_mnt_opts, struct super_block *sb, void *mnt_opts,
include/linux/mtd/super.h
18
int (*fill_super)(struct super_block *sb,
include/linux/mtd/super.h
20
extern void kill_mtd_super(struct super_block *sb);
include/linux/nfs_fs.h
445
struct inode *nfs_ilookup(struct super_block *sb, struct nfs_fattr *, struct nfs_fh *);
include/linux/nfs_ssc.h
28
void (*sco_sb_deactive)(struct super_block *sb);
include/linux/nfs_ssc.h
77
static inline void nfs_do_sb_deactive(struct super_block *sb)
include/linux/nfs_ssc.h
80
(*nfs_ssc_client_tbl.ssc_nfs_ops->sco_sb_deactive)(sb);
include/linux/proc_fs.h
244
static inline struct pid_namespace *proc_pid_ns(struct super_block *sb)
include/linux/proc_fs.h
246
return proc_sb_info(sb)->pid_ns;
include/linux/proc_fs.h
75
static inline struct proc_fs_info *proc_sb_info(struct super_block *sb)
include/linux/proc_fs.h
77
return sb->s_fs_info;
include/linux/qed/qed_if.h
1207
int (*get_sb_info)(struct qed_dev *cdev, struct qed_sb_info *sb,
include/linux/quota.h
247
extern void mark_info_dirty(struct super_block *sb, int type);
include/linux/quota.h
313
int (*check_quota_file)(struct super_block *sb, int type); /* Detect whether file is in our format */
include/linux/quota.h
314
int (*read_file_info)(struct super_block *sb, int type); /* Read main info about file - called on quotaon() */
include/linux/quota.h
315
int (*write_file_info)(struct super_block *sb, int type); /* Write main info about file */
include/linux/quota.h
316
int (*free_file_info)(struct super_block *sb, int type); /* Called on quotaoff() */
include/linux/quota.h
320
int (*get_next_id)(struct super_block *sb, struct kqid *qid); /* Get next ID with existing structure in the quota file */
include/linux/quota.h
339
int (*get_next_id) (struct super_block *sb, struct kqid *qid);
include/linux/quotaops.h
102
int dquot_quota_on(struct super_block *sb, int type, int format_id,
include/linux/quotaops.h
104
int dquot_quota_on_mount(struct super_block *sb, char *qf_name,
include/linux/quotaops.h
106
int dquot_quota_off(struct super_block *sb, int type);
include/linux/quotaops.h
107
int dquot_writeback_dquots(struct super_block *sb, int type);
include/linux/quotaops.h
108
int dquot_quota_sync(struct super_block *sb, int type);
include/linux/quotaops.h
109
int dquot_get_state(struct super_block *sb, struct qc_state *state);
include/linux/quotaops.h
110
int dquot_set_dqinfo(struct super_block *sb, int type, struct qc_info *ii);
include/linux/quotaops.h
111
int dquot_get_dqblk(struct super_block *sb, struct kqid id,
include/linux/quotaops.h
113
int dquot_get_next_dqblk(struct super_block *sb, struct kqid *id,
include/linux/quotaops.h
115
int dquot_set_dqblk(struct super_block *sb, struct kqid id,
include/linux/quotaops.h
122
static inline struct mem_dqinfo *sb_dqinfo(struct super_block *sb, int type)
include/linux/quotaops.h
124
return sb_dqopt(sb)->info + type;
include/linux/quotaops.h
131
static inline bool sb_has_quota_usage_enabled(struct super_block *sb, int type)
include/linux/quotaops.h
133
return sb_dqopt(sb)->flags &
include/linux/quotaops.h
137
static inline bool sb_has_quota_limits_enabled(struct super_block *sb, int type)
include/linux/quotaops.h
139
return sb_dqopt(sb)->flags &
include/linux/quotaops.h
143
static inline bool sb_has_quota_suspended(struct super_block *sb, int type)
include/linux/quotaops.h
145
return sb_dqopt(sb)->flags &
include/linux/quotaops.h
149
static inline unsigned sb_any_quota_suspended(struct super_block *sb)
include/linux/quotaops.h
151
return dquot_state_types(sb_dqopt(sb)->flags, DQUOT_SUSPENDED);
include/linux/quotaops.h
155
static inline bool sb_has_quota_loaded(struct super_block *sb, int type)
include/linux/quotaops.h
158
return sb_has_quota_usage_enabled(sb, type);
include/linux/quotaops.h
161
static inline unsigned sb_any_quota_loaded(struct super_block *sb)
include/linux/quotaops.h
163
return dquot_state_types(sb_dqopt(sb)->flags, DQUOT_USAGE_ENABLED);
include/linux/quotaops.h
166
static inline bool sb_has_quota_active(struct super_block *sb, int type)
include/linux/quotaops.h
168
return sb_has_quota_loaded(sb, type) &&
include/linux/quotaops.h
169
!sb_has_quota_suspended(sb, type);
include/linux/quotaops.h
17
static inline struct quota_info *sb_dqopt(struct super_block *sb)
include/linux/quotaops.h
180
static inline int sb_has_quota_usage_enabled(struct super_block *sb, int type)
include/linux/quotaops.h
185
static inline int sb_has_quota_limits_enabled(struct super_block *sb, int type)
include/linux/quotaops.h
19
return &sb->s_dquot;
include/linux/quotaops.h
190
static inline int sb_has_quota_suspended(struct super_block *sb, int type)
include/linux/quotaops.h
195
static inline int sb_any_quota_suspended(struct super_block *sb)
include/linux/quotaops.h
201
static inline int sb_has_quota_loaded(struct super_block *sb, int type)
include/linux/quotaops.h
206
static inline int sb_any_quota_loaded(struct super_block *sb)
include/linux/quotaops.h
211
static inline int sb_has_quota_active(struct super_block *sb, int type)
include/linux/quotaops.h
272
static inline int dquot_disable(struct super_block *sb, int type,
include/linux/quotaops.h
278
static inline int dquot_suspend(struct super_block *sb, int type)
include/linux/quotaops.h
283
static inline int dquot_resume(struct super_block *sb, int type)
include/linux/quotaops.h
290
static inline int dquot_writeback_dquots(struct super_block *sb, int type)
include/linux/quotaops.h
33
#define quota_error(sb, fmt, args...) \
include/linux/quotaops.h
34
__quota_error((sb), __func__, fmt , ## args)
include/linux/quotaops.h
37
void __quota_error(struct super_block *sb, const char *func,
include/linux/quotaops.h
46
struct dquot *dqget(struct super_block *sb, struct kqid qid);
include/linux/quotaops.h
66
int dquot_scan_active(struct super_block *sb,
include/linux/quotaops.h
69
struct dquot *dquot_alloc(struct super_block *sb, int type);
include/linux/quotaops.h
81
int dquot_disable(struct super_block *sb, int type, unsigned int flags);
include/linux/quotaops.h
83
static inline int dquot_suspend(struct super_block *sb, int type)
include/linux/quotaops.h
85
return dquot_disable(sb, type, DQUOT_SUSPENDED);
include/linux/quotaops.h
87
int dquot_resume(struct super_block *sb, int type);
include/linux/quotaops.h
92
int dquot_commit_info(struct super_block *sb, int type);
include/linux/quotaops.h
93
int dquot_get_next_id(struct super_block *sb, struct kqid *qid);
include/linux/quotaops.h
98
int dquot_load_quota_sb(struct super_block *sb, int type, int format_id,
include/linux/ramfs.h
10
extern void ramfs_kill_sb(struct super_block *sb);
include/linux/ramfs.h
7
struct inode *ramfs_get_inode(struct super_block *sb, const struct inode *dir,
include/linux/sbitmap.h
113
struct sbitmap sb;
include/linux/sbitmap.h
170
int sbitmap_init_node(struct sbitmap *sb, unsigned int depth, int shift,
include/linux/sbitmap.h
174
static inline unsigned int __map_depth(const struct sbitmap *sb, int index)
include/linux/sbitmap.h
176
if (index == sb->map_nr - 1)
include/linux/sbitmap.h
177
return sb->depth - (index << sb->shift);
include/linux/sbitmap.h
178
return 1U << sb->shift;
include/linux/sbitmap.h
185
static inline void sbitmap_free(struct sbitmap *sb)
include/linux/sbitmap.h
187
free_percpu(sb->alloc_hint);
include/linux/sbitmap.h
188
kvfree(sb->map);
include/linux/sbitmap.h
189
sb->map = NULL;
include/linux/sbitmap.h
200
void sbitmap_resize(struct sbitmap *sb, unsigned int depth);
include/linux/sbitmap.h
210
int sbitmap_get(struct sbitmap *sb);
include/linux/sbitmap.h
218
bool sbitmap_any_bit_set(const struct sbitmap *sb);
include/linux/sbitmap.h
220
#define SB_NR_TO_INDEX(sb, bitnr) ((bitnr) >> (sb)->shift)
include/linux/sbitmap.h
221
#define SB_NR_TO_BIT(sb, bitnr) ((bitnr) & ((1U << (sb)->shift) - 1U))
include/linux/sbitmap.h
235
static inline void __sbitmap_for_each_set(struct sbitmap *sb,
include/linux/sbitmap.h
243
if (start >= sb->depth)
include/linux/sbitmap.h
245
index = SB_NR_TO_INDEX(sb, start);
include/linux/sbitmap.h
246
nr = SB_NR_TO_BIT(sb, start);
include/linux/sbitmap.h
248
while (scanned < sb->depth) {
include/linux/sbitmap.h
251
__map_depth(sb, index) - nr,
include/linux/sbitmap.h
252
sb->depth - scanned);
include/linux/sbitmap.h
255
word = sb->map[index].word & ~sb->map[index].cleared;
include/linux/sbitmap.h
269
if (!fn(sb, (index << sb->shift) + nr, data))
include/linux/sbitmap.h
276
if (++index >= sb->map_nr)
include/linux/sbitmap.h
287
static inline void sbitmap_for_each_set(struct sbitmap *sb, sb_for_each_fn fn,
include/linux/sbitmap.h
290
__sbitmap_for_each_set(sb, 0, fn, data);
include/linux/sbitmap.h
293
static inline unsigned long *__sbitmap_word(struct sbitmap *sb,
include/linux/sbitmap.h
296
return &sb->map[SB_NR_TO_INDEX(sb, bitnr)].word;
include/linux/sbitmap.h
301
static inline void sbitmap_set_bit(struct sbitmap *sb, unsigned int bitnr)
include/linux/sbitmap.h
303
set_bit(SB_NR_TO_BIT(sb, bitnr), __sbitmap_word(sb, bitnr));
include/linux/sbitmap.h
306
static inline void sbitmap_clear_bit(struct sbitmap *sb, unsigned int bitnr)
include/linux/sbitmap.h
308
clear_bit(SB_NR_TO_BIT(sb, bitnr), __sbitmap_word(sb, bitnr));
include/linux/sbitmap.h
317
static inline void sbitmap_deferred_clear_bit(struct sbitmap *sb, unsigned int bitnr)
include/linux/sbitmap.h
319
unsigned long *addr = &sb->map[SB_NR_TO_INDEX(sb, bitnr)].cleared;
include/linux/sbitmap.h
321
set_bit(SB_NR_TO_BIT(sb, bitnr), addr);
include/linux/sbitmap.h
328
static inline void sbitmap_put(struct sbitmap *sb, unsigned int bitnr)
include/linux/sbitmap.h
330
sbitmap_deferred_clear_bit(sb, bitnr);
include/linux/sbitmap.h
332
if (likely(sb->alloc_hint && !sb->round_robin && bitnr < sb->depth))
include/linux/sbitmap.h
333
*raw_cpu_ptr(sb->alloc_hint) = bitnr;
include/linux/sbitmap.h
336
static inline int sbitmap_test_bit(struct sbitmap *sb, unsigned int bitnr)
include/linux/sbitmap.h
338
return test_bit(SB_NR_TO_BIT(sb, bitnr), __sbitmap_word(sb, bitnr));
include/linux/sbitmap.h
366
void sbitmap_show(struct sbitmap *sb, struct seq_file *m);
include/linux/sbitmap.h
376
unsigned int sbitmap_weight(const struct sbitmap *sb);
include/linux/sbitmap.h
387
void sbitmap_bitmap_show(struct sbitmap *sb, struct seq_file *m);
include/linux/sbitmap.h
412
sbitmap_free(&sbq->sb);
include/linux/security.h
353
int security_quotactl(int cmds, int type, int id, const struct super_block *sb);
include/linux/security.h
366
int security_sb_alloc(struct super_block *sb);
include/linux/security.h
367
void security_sb_delete(struct super_block *sb);
include/linux/security.h
368
void security_sb_free(struct super_block *sb);
include/linux/security.h
371
int security_sb_mnt_opts_compat(struct super_block *sb, void *mnt_opts);
include/linux/security.h
372
int security_sb_remount(struct super_block *sb, void *mnt_opts);
include/linux/security.h
373
int security_sb_kern_mount(const struct super_block *sb);
include/linux/security.h
374
int security_sb_show_options(struct seq_file *m, struct super_block *sb);
include/linux/security.h
380
int security_sb_set_mnt_opts(struct super_block *sb,
include/linux/security.h
699
const struct super_block *sb)
include/linux/security.h
765
static inline int security_sb_alloc(struct super_block *sb)
include/linux/security.h
770
static inline void security_sb_delete(struct super_block *sb)
include/linux/security.h
773
static inline void security_sb_free(struct super_block *sb)
include/linux/security.h
782
static inline int security_sb_remount(struct super_block *sb,
include/linux/security.h
788
static inline int security_sb_mnt_opts_compat(struct super_block *sb,
include/linux/security.h
795
static inline int security_sb_kern_mount(struct super_block *sb)
include/linux/security.h
801
struct super_block *sb)
include/linux/security.h
829
static inline int security_sb_set_mnt_opts(struct super_block *sb,
include/linux/ssb/ssb_driver_chipcommon.h
470
#define SSB_CHIPCO_CLK_SB2MIPS_T6(sb) (2 * (sb))
include/linux/sunrpc/rpc_pipe_fs.h
84
extern struct dentry *rpc_d_lookup_sb(const struct super_block *sb,
include/linux/usb/xhci-sideband.h
101
xhci_sideband_interrupter_id(struct xhci_sideband *sb);
include/linux/usb/xhci-sideband.h
104
void xhci_sideband_notify_ep_ring_free(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
107
static inline void xhci_sideband_notify_ep_ring_free(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
72
xhci_sideband_unregister(struct xhci_sideband *sb);
include/linux/usb/xhci-sideband.h
74
xhci_sideband_add_endpoint(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
77
xhci_sideband_remove_endpoint(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
80
xhci_sideband_stop_endpoint(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
83
xhci_sideband_get_endpoint_buffer(struct xhci_sideband *sb,
include/linux/usb/xhci-sideband.h
86
xhci_sideband_get_event_buffer(struct xhci_sideband *sb);
include/linux/usb/xhci-sideband.h
96
xhci_sideband_create_interrupter(struct xhci_sideband *sb, int num_seg,
include/linux/usb/xhci-sideband.h
99
xhci_sideband_remove_interrupter(struct xhci_sideband *sb);
include/linux/writeback.h
184
void try_to_writeback_inodes_sb(struct super_block *sb, enum wb_reason reason);
include/linux/writeback.h
210
void cgroup_writeback_umount(struct super_block *sb);
include/linux/writeback.h
298
static inline void cgroup_writeback_umount(struct super_block *sb)
include/net/9p/client.h
277
int p9_client_statfs(struct p9_fid *fid, struct p9_rstatfs *sb);
include/trace/events/ext4.h
1016
TP_PROTO(struct super_block *sb, int wait),
include/trace/events/ext4.h
1018
TP_ARGS(sb, wait),
include/trace/events/ext4.h
1027
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1167
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1173
TP_ARGS(sb, inode, group, start, len),
include/trace/events/ext4.h
1184
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1200
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1206
TP_ARGS(sb, inode, group, start, len)
include/trace/events/ext4.h
1211
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1217
TP_ARGS(sb, inode, group, start, len)
include/trace/events/ext4.h
1345
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1347
TP_ARGS(sb, group),
include/trace/events/ext4.h
1356
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1367
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1369
TP_ARGS(sb, group)
include/trace/events/ext4.h
1374
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1376
TP_ARGS(sb, group)
include/trace/events/ext4.h
1381
TP_PROTO(struct super_block *sb, unsigned long group),
include/trace/events/ext4.h
1383
TP_ARGS(sb, group)
include/trace/events/ext4.h
1387
TP_PROTO(struct super_block *sb, unsigned long group, bool prefetch),
include/trace/events/ext4.h
1389
TP_ARGS(sb, group, prefetch),
include/trace/events/ext4.h
1399
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1770
TP_PROTO(struct super_block *sb, unsigned long ino),
include/trace/events/ext4.h
1772
TP_ARGS(sb, ino),
include/trace/events/ext4.h
1780
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1790
TP_PROTO(struct super_block *sb, int blocks, int rsv_blocks,
include/trace/events/ext4.h
1793
TP_ARGS(sb, blocks, rsv_blocks, revoke_creds, type, IP),
include/trace/events/ext4.h
1805
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1853
TP_PROTO(struct super_block *sb, int blocks, unsigned long IP),
include/trace/events/ext4.h
1855
TP_ARGS(sb, blocks, IP),
include/trace/events/ext4.h
1864
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
1875
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1880
TP_ARGS(sb, group, start, len),
include/trace/events/ext4.h
1891
__entry->dev_major = MAJOR(sb->s_dev);
include/trace/events/ext4.h
1892
__entry->dev_minor = MINOR(sb->s_dev);
include/trace/events/ext4.h
1905
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1910
TP_ARGS(sb, group, start, len)
include/trace/events/ext4.h
1915
TP_PROTO(struct super_block *sb,
include/trace/events/ext4.h
1920
TP_ARGS(sb, group, start, len)
include/trace/events/ext4.h
1962
TP_PROTO(struct super_block *sb, struct ext4_map_blocks *map, int ret),
include/trace/events/ext4.h
1964
TP_ARGS(sb, map, ret),
include/trace/events/ext4.h
1976
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2389
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2391
TP_ARGS(sb, nr_to_scan, cache_cnt),
include/trace/events/ext4.h
2400
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2411
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2413
TP_ARGS(sb, nr_to_scan, cache_cnt)
include/trace/events/ext4.h
2417
TP_PROTO(struct super_block *sb, int nr_to_scan, int cache_cnt),
include/trace/events/ext4.h
2419
TP_ARGS(sb, nr_to_scan, cache_cnt)
include/trace/events/ext4.h
2423
TP_PROTO(struct super_block *sb, int nr_shrunk, int cache_cnt),
include/trace/events/ext4.h
2425
TP_ARGS(sb, nr_shrunk, cache_cnt),
include/trace/events/ext4.h
2434
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2495
TP_PROTO(struct super_block *sb, int nr_shrunk, u64 scan_time,
include/trace/events/ext4.h
2498
TP_ARGS(sb, nr_shrunk, scan_time, nr_skipped, retried),
include/trace/events/ext4.h
2509
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2563
TP_PROTO(struct super_block *sb, u32 keydev, u32 agno, u64 bno, u64 len,
include/trace/events/ext4.h
2565
TP_ARGS(sb, keydev, agno, bno, len, owner),
include/trace/events/ext4.h
2575
__entry->dev = sb->s_bdev->bd_dev;
include/trace/events/ext4.h
2592
TP_PROTO(struct super_block *sb, u32 keydev, u32 agno, u64 bno, u64 len, \
include/trace/events/ext4.h
2594
TP_ARGS(sb, keydev, agno, bno, len, owner))
include/trace/events/ext4.h
2600
TP_PROTO(struct super_block *sb, struct ext4_fsmap *fsmap),
include/trace/events/ext4.h
2601
TP_ARGS(sb, fsmap),
include/trace/events/ext4.h
2611
__entry->dev = sb->s_bdev->bd_dev;
include/trace/events/ext4.h
2628
TP_PROTO(struct super_block *sb, struct ext4_fsmap *fsmap), \
include/trace/events/ext4.h
2629
TP_ARGS(sb, fsmap))
include/trace/events/ext4.h
2635
TP_PROTO(struct super_block *sb, unsigned long flags),
include/trace/events/ext4.h
2637
TP_ARGS(sb, flags),
include/trace/events/ext4.h
2645
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2655
TP_PROTO(struct super_block *sb, const char *function,
include/trace/events/ext4.h
2658
TP_ARGS(sb, function, line),
include/trace/events/ext4.h
2667
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2678
TP_PROTO(struct super_block *sb, ext4_group_t group,
include/trace/events/ext4.h
2681
TP_ARGS(sb, group, next, prefetch_ios),
include/trace/events/ext4.h
2691
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2703
TP_PROTO(struct super_block *sb, ext4_group_t group),
include/trace/events/ext4.h
2705
TP_ARGS(sb, group),
include/trace/events/ext4.h
2713
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2722
TP_PROTO(struct super_block *sb, int error, int off),
include/trace/events/ext4.h
2724
TP_ARGS(sb, error, off),
include/trace/events/ext4.h
2733
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2744
TP_PROTO(struct super_block *sb, int tag, int ino, int priv1, int priv2),
include/trace/events/ext4.h
2746
TP_ARGS(sb, tag, ino, priv1, priv2),
include/trace/events/ext4.h
2757
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2770
TP_PROTO(struct super_block *sb, tid_t commit_tid),
include/trace/events/ext4.h
2772
TP_ARGS(sb, commit_tid),
include/trace/events/ext4.h
2780
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2789
TP_PROTO(struct super_block *sb, int nblks, int reason,
include/trace/events/ext4.h
2792
TP_ARGS(sb, nblks, reason, commit_tid),
include/trace/events/ext4.h
2805
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2808
__entry->num_fc = EXT4_SB(sb)->s_fc_stats.fc_num_commits;
include/trace/events/ext4.h
2810
EXT4_SB(sb)->s_fc_stats.fc_ineligible_commits;
include/trace/events/ext4.h
2811
__entry->nblks_agg = EXT4_SB(sb)->s_fc_stats.fc_numblks;
include/trace/events/ext4.h
2826
TP_PROTO(struct super_block *sb),
include/trace/events/ext4.h
2828
TP_ARGS(sb),
include/trace/events/ext4.h
2841
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
2844
EXT4_SB(sb)->s_fc_stats.fc_ineligible_reason_count[i];
include/trace/events/ext4.h
2846
__entry->fc_commits = EXT4_SB(sb)->s_fc_stats.fc_num_commits;
include/trace/events/ext4.h
2848
EXT4_SB(sb)->s_fc_stats.fc_ineligible_commits;
include/trace/events/ext4.h
2849
__entry->fc_numblks = EXT4_SB(sb)->s_fc_stats.fc_numblks;
include/trace/events/ext4.h
2989
struct super_block *sb = journal->j_private;
include/trace/events/ext4.h
2991
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
3003
TP_PROTO(struct super_block *sb, ext4_fsblk_t fsblk,
include/trace/events/ext4.h
3006
TP_ARGS(sb, fsblk, flags),
include/trace/events/ext4.h
3015
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
690
TP_PROTO(struct super_block *sb, unsigned long long blk,
include/trace/events/ext4.h
693
TP_ARGS(sb, blk, count),
include/trace/events/ext4.h
703
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
786
TP_PROTO(struct super_block *sb, struct ext4_prealloc_space *pa),
include/trace/events/ext4.h
788
TP_ARGS(sb, pa),
include/trace/events/ext4.h
798
__entry->dev = sb->s_dev;
include/trace/events/ext4.h
832
TP_PROTO(struct super_block *sb, int needed),
include/trace/events/ext4.h
834
TP_ARGS(sb, needed),
include/trace/events/ext4.h
843
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
1233
TP_PROTO(struct super_block *sb, int type, struct bio *bio),
include/trace/events/f2fs.h
1235
TP_ARGS(sb, type, bio),
include/trace/events/f2fs.h
1248
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
1268
TP_PROTO(struct super_block *sb, int type, struct bio *bio),
include/trace/events/f2fs.h
1270
TP_ARGS(sb, type, bio),
include/trace/events/f2fs.h
1277
TP_PROTO(struct super_block *sb, int type, struct bio *bio),
include/trace/events/f2fs.h
1279
TP_ARGS(sb, type, bio),
include/trace/events/f2fs.h
1286
TP_PROTO(struct super_block *sb, int type, struct bio *bio),
include/trace/events/f2fs.h
1288
TP_ARGS(sb, type, bio),
include/trace/events/f2fs.h
1295
TP_PROTO(struct super_block *sb, int type, struct bio *bio),
include/trace/events/f2fs.h
1297
TP_ARGS(sb, type, bio),
include/trace/events/f2fs.h
1604
TP_PROTO(struct super_block *sb, int reason, u16 phase),
include/trace/events/f2fs.h
1606
TP_ARGS(sb, reason, phase),
include/trace/events/f2fs.h
1615
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
1920
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
1962
TP_PROTO(struct super_block *sb, int type, s64 count),
include/trace/events/f2fs.h
1964
TP_ARGS(sb, type, count),
include/trace/events/f2fs.h
1973
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
1986
TP_PROTO(struct super_block *sb, int type, s64 count),
include/trace/events/f2fs.h
1988
TP_ARGS(sb, type, count)
include/trace/events/f2fs.h
1993
TP_PROTO(struct super_block *sb, int type, s64 count),
include/trace/events/f2fs.h
1995
TP_ARGS(sb, type, count)
include/trace/events/f2fs.h
2011
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
2156
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
2258
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
2494
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
2547
__entry->dev = sbi->sb->s_dev;
include/trace/events/f2fs.h
307
TP_PROTO(struct super_block *sb, int wait),
include/trace/events/f2fs.h
309
TP_ARGS(sb, wait),
include/trace/events/f2fs.h
318
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
319
__entry->dirty = is_sbi_flag_set(F2FS_SB(sb), SBI_IS_DIRTY);
include/trace/events/f2fs.h
691
TP_PROTO(struct super_block *sb, unsigned int wait_ms,
include/trace/events/f2fs.h
694
TP_ARGS(sb, wait_ms, prefree, free),
include/trace/events/f2fs.h
704
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
719
TP_PROTO(struct super_block *sb, int gc_type, bool no_bg_gc,
include/trace/events/f2fs.h
726
TP_ARGS(sb, gc_type, no_bg_gc, nr_free_secs, dirty_nodes,
include/trace/events/f2fs.h
745
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
776
TP_PROTO(struct super_block *sb, int ret, int seg_freed,
include/trace/events/f2fs.h
782
TP_ARGS(sb, ret, seg_freed, sec_freed, dirty_nodes, dirty_dents,
include/trace/events/f2fs.h
800
__entry->dev = sb->s_dev;
include/trace/events/f2fs.h
831
TP_PROTO(struct super_block *sb, int type, int gc_type,
include/trace/events/f2fs.h
835
TP_ARGS(sb, type, gc_type, p, pre_victim, prefree, free),
include/trace/events/f2fs.h
852
__entry->dev = sb->s_dev;
include/trace/events/nilfs2.h
67
TP_PROTO(struct super_block *sb,
include/trace/events/nilfs2.h
73
TP_ARGS(sb, ti, count, flags, state),
include/trace/events/nilfs2.h
76
__field(void *, sb)
include/trace/events/nilfs2.h
84
__entry->sb = sb;
include/trace/events/nilfs2.h
92
__entry->sb,
include/trace/events/writeback.h
400
__entry->sb_dev = work->sb ? work->sb->s_dev : 0;
include/uapi/linux/bfs_fs.h
78
#define BFS_UNCLEAN(bfs_sb, sb) \
include/uapi/linux/bfs_fs.h
79
((le32_to_cpu(bfs_sb->s_from) != -1) && (le32_to_cpu(bfs_sb->s_to) != -1) && !(sb->s_flags & SB_RDONLY))
include/uapi/linux/raid/md_p.h
211
static inline __u64 md_event(mdp_super_t *sb) {
include/uapi/linux/raid/md_p.h
212
__u64 ev = sb->events_hi;
include/uapi/linux/raid/md_p.h
213
return (ev<<32)| sb->events_lo;
ipc/mqueue.c
289
static struct inode *mqueue_get_inode(struct super_block *sb,
ipc/mqueue.c
296
inode = new_inode(sb);
ipc/mqueue.c
403
static int mqueue_fill_super(struct super_block *sb, struct fs_context *fc)
ipc/mqueue.c
406
struct ipc_namespace *ns = sb->s_fs_info;
ipc/mqueue.c
408
sb->s_iflags |= SB_I_NOEXEC | SB_I_NODEV;
ipc/mqueue.c
409
sb->s_blocksize = PAGE_SIZE;
ipc/mqueue.c
410
sb->s_blocksize_bits = PAGE_SHIFT;
ipc/mqueue.c
411
sb->s_magic = MQUEUE_MAGIC;
ipc/mqueue.c
412
sb->s_op = &mqueue_super_ops;
ipc/mqueue.c
413
sb->s_d_flags = DCACHE_DONTCACHE;
ipc/mqueue.c
415
inode = mqueue_get_inode(sb, ns, S_IFDIR | S_ISVTX | S_IRWXUGO, NULL);
ipc/mqueue.c
419
sb->s_root = d_make_root(inode);
ipc/mqueue.c
420
if (!sb->s_root)
ipc/mqueue.c
497
static struct inode *mqueue_alloc_inode(struct super_block *sb)
ipc/mqueue.c
501
ei = alloc_inode_sb(sb, mqueue_inode_cachep, GFP_KERNEL);
kernel/bpf/inode.c
1000
struct bpf_mount_opts *opts = sb->s_fs_info;
kernel/bpf/inode.c
1008
ret = simple_fill_super(sb, BPF_FS_MAGIC, bpf_rfiles);
kernel/bpf/inode.c
1012
sb->s_op = &bpf_super_ops;
kernel/bpf/inode.c
1014
inode = sb->s_root->d_inode;
kernel/bpf/inode.c
1019
populate_bpffs(sb->s_root);
kernel/bpf/inode.c
102
struct inode *bpf_get_inode(struct super_block *sb,
kernel/bpf/inode.c
1066
static void bpf_kill_super(struct super_block *sb)
kernel/bpf/inode.c
1068
struct bpf_mount_opts *opts = sb->s_fs_info;
kernel/bpf/inode.c
1070
kill_anon_super(sb);
kernel/bpf/inode.c
117
inode = new_inode(sb);
kernel/bpf/inode.c
997
static int bpf_fill_super(struct super_block *sb, struct fs_context *fc)
kernel/bpf/token.c
120
struct super_block *sb;
kernel/bpf/token.c
128
sb = path.dentry->d_sb;
kernel/bpf/token.c
130
if (path.dentry != sb->s_root)
kernel/bpf/token.c
132
if (sb->s_op != &bpf_super_ops)
kernel/bpf/token.c
138
userns = sb->s_user_ns;
kernel/bpf/token.c
153
mnt_opts = sb->s_fs_info;
kernel/bpf/token.c
161
inode = bpf_get_inode(sb, NULL, mode);
kernel/cgroup/cgroup.c
2276
struct super_block *sb = fc->root->d_sb;
kernel/cgroup/cgroup.c
2287
nsdentry = kernfs_node_dentry(cgrp->kn, sb);
kernel/cgroup/cgroup.c
2290
deactivate_locked_super(sb);
kernel/cgroup/cgroup.c
2375
static void cgroup_kill_sb(struct super_block *sb)
kernel/cgroup/cgroup.c
2377
struct kernfs_root *kf_root = kernfs_root_from_sb(sb);
kernel/cgroup/cgroup.c
2390
kernfs_kill_sb(sb);
kernel/cgroup/cgroup.c
5298
static int cgroup_may_write(const struct cgroup *cgrp, struct super_block *sb)
kernel/cgroup/cgroup.c
5305
inode = kernfs_get_inode(sb, cgrp->procs_file.kn);
kernel/cgroup/cgroup.c
5316
struct super_block *sb,
kernel/cgroup/cgroup.c
5329
ret = cgroup_may_write(com_cgrp, sb);
kernel/cgroup/cgroup.c
5347
struct super_block *sb, bool threadgroup,
kernel/cgroup/cgroup.c
5352
ret = cgroup_procs_write_permission(src_cgrp, dst_cgrp, sb, ns);
kernel/cgroup/cgroup.c
6775
struct super_block *sb;
kernel/cgroup/cgroup.c
6801
sb = fd_file(f)->f_path.dentry->d_sb;
kernel/cgroup/cgroup.c
6820
ret = cgroup_may_write(dst_cgrp, sb);
kernel/cgroup/cgroup.c
6838
ret = cgroup_attach_permissions(cset->dfl_cgrp, dst_cgrp, sb,
lib/codetag.c
140
DECLARE_SEQ_BUF(sb, KSYM_NAME_LEN);
lib/codetag.c
144
seq_buf_printf(&sb, "%s%s", prefix, name);
lib/codetag.c
145
if (seq_buf_has_overflowed(&sb))
lib/codetag.c
148
buf = seq_buf_str(&sb);
lib/crypto/powerpc/curve25519.h
64
unsigned char *sb = PTR_ALIGN((void *)sbuf, 8);
lib/crypto/powerpc/curve25519.h
66
memcpy(sb, s, 32);
lib/crypto/powerpc/curve25519.h
67
x25519_fe51_frombytes(h, sb);
lib/sbitmap.c
101
int sbitmap_init_node(struct sbitmap *sb, unsigned int depth, int shift,
lib/sbitmap.c
115
sb->shift = shift;
lib/sbitmap.c
116
sb->depth = depth;
lib/sbitmap.c
117
sb->map_nr = DIV_ROUND_UP(sb->depth, bits_per_word);
lib/sbitmap.c
118
sb->round_robin = round_robin;
lib/sbitmap.c
12
static int init_alloc_hint(struct sbitmap *sb, gfp_t flags)
lib/sbitmap.c
121
sb->map = NULL;
lib/sbitmap.c
126
if (init_alloc_hint(sb, flags))
lib/sbitmap.c
129
sb->alloc_hint = NULL;
lib/sbitmap.c
132
sb->map = kvzalloc_node(sb->map_nr * sizeof(*sb->map), flags, node);
lib/sbitmap.c
133
if (!sb->map) {
lib/sbitmap.c
134
free_percpu(sb->alloc_hint);
lib/sbitmap.c
138
for (i = 0; i < sb->map_nr; i++)
lib/sbitmap.c
139
raw_spin_lock_init(&sb->map[i].swap_lock);
lib/sbitmap.c
14
unsigned depth = sb->depth;
lib/sbitmap.c
145
void sbitmap_resize(struct sbitmap *sb, unsigned int depth)
lib/sbitmap.c
147
unsigned int bits_per_word = 1U << sb->shift;
lib/sbitmap.c
150
for (i = 0; i < sb->map_nr; i++)
lib/sbitmap.c
151
sbitmap_deferred_clear(&sb->map[i], 0, 0, 0);
lib/sbitmap.c
153
sb->depth = depth;
lib/sbitmap.c
154
sb->map_nr = DIV_ROUND_UP(sb->depth, bits_per_word);
lib/sbitmap.c
16
sb->alloc_hint = alloc_percpu_gfp(unsigned int, flags);
lib/sbitmap.c
17
if (!sb->alloc_hint)
lib/sbitmap.c
20
if (depth && !sb->round_robin) {
lib/sbitmap.c
211
static unsigned int __map_depth_with_shallow(const struct sbitmap *sb,
lib/sbitmap.c
218
word_depth = __map_depth(sb, index);
lib/sbitmap.c
219
if (shallow_depth >= sb->depth)
lib/sbitmap.c
223
reminder = do_div(shallow_word_depth, sb->depth);
lib/sbitmap.c
231
static int sbitmap_find_bit(struct sbitmap *sb,
lib/sbitmap.c
24
*per_cpu_ptr(sb->alloc_hint, i) = get_random_u32_below(depth);
lib/sbitmap.c
240
for (i = 0; i < sb->map_nr; i++) {
lib/sbitmap.c
241
unsigned int depth = __map_depth_with_shallow(sb, index,
lib/sbitmap.c
245
nr = sbitmap_find_bit_in_word(&sb->map[index], depth,
lib/sbitmap.c
248
nr += index << sb->shift;
lib/sbitmap.c
254
if (++index >= sb->map_nr)
lib/sbitmap.c
261
static int __sbitmap_get(struct sbitmap *sb, unsigned int alloc_hint)
lib/sbitmap.c
265
index = SB_NR_TO_INDEX(sb, alloc_hint);
lib/sbitmap.c
272
if (sb->round_robin)
lib/sbitmap.c
273
alloc_hint = SB_NR_TO_BIT(sb, alloc_hint);
lib/sbitmap.c
277
return sbitmap_find_bit(sb, UINT_MAX, index, alloc_hint,
lib/sbitmap.c
278
!sb->round_robin);
lib/sbitmap.c
281
int sbitmap_get(struct sbitmap *sb)
lib/sbitmap.c
286
if (WARN_ON_ONCE(unlikely(!sb->alloc_hint)))
lib/sbitmap.c
289
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
29
static inline unsigned update_alloc_hint_before_get(struct sbitmap *sb,
lib/sbitmap.c
290
hint = update_alloc_hint_before_get(sb, depth);
lib/sbitmap.c
291
nr = __sbitmap_get(sb, hint);
lib/sbitmap.c
292
update_alloc_hint_after_get(sb, depth, hint, nr);
lib/sbitmap.c
298
static int __sbitmap_get_shallow(struct sbitmap *sb,
lib/sbitmap.c
304
index = SB_NR_TO_INDEX(sb, alloc_hint);
lib/sbitmap.c
305
alloc_hint = SB_NR_TO_BIT(sb, alloc_hint);
lib/sbitmap.c
307
return sbitmap_find_bit(sb, shallow_depth, index, alloc_hint, true);
lib/sbitmap.c
325
static int sbitmap_get_shallow(struct sbitmap *sb, unsigned long shallow_depth)
lib/sbitmap.c
330
if (WARN_ON_ONCE(unlikely(!sb->alloc_hint)))
lib/sbitmap.c
333
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
334
hint = update_alloc_hint_before_get(sb, depth);
lib/sbitmap.c
335
nr = __sbitmap_get_shallow(sb, hint, shallow_depth);
lib/sbitmap.c
336
update_alloc_hint_after_get(sb, depth, hint, nr);
lib/sbitmap.c
34
hint = this_cpu_read(*sb->alloc_hint);
lib/sbitmap.c
341
bool sbitmap_any_bit_set(const struct sbitmap *sb)
lib/sbitmap.c
345
for (i = 0; i < sb->map_nr; i++) {
lib/sbitmap.c
346
if (sb->map[i].word & ~sb->map[i].cleared)
lib/sbitmap.c
353
static unsigned int __sbitmap_weight(const struct sbitmap *sb, bool set)
lib/sbitmap.c
357
for (i = 0; i < sb->map_nr; i++) {
lib/sbitmap.c
358
const struct sbitmap_word *word = &sb->map[i];
lib/sbitmap.c
359
unsigned int word_depth = __map_depth(sb, i);
lib/sbitmap.c
369
static unsigned int sbitmap_cleared(const struct sbitmap *sb)
lib/sbitmap.c
37
this_cpu_write(*sb->alloc_hint, hint);
lib/sbitmap.c
371
return __sbitmap_weight(sb, false);
lib/sbitmap.c
374
unsigned int sbitmap_weight(const struct sbitmap *sb)
lib/sbitmap.c
376
return __sbitmap_weight(sb, true) - sbitmap_cleared(sb);
lib/sbitmap.c
380
void sbitmap_show(struct sbitmap *sb, struct seq_file *m)
lib/sbitmap.c
382
seq_printf(m, "depth=%u\n", sb->depth);
lib/sbitmap.c
383
seq_printf(m, "busy=%u\n", sbitmap_weight(sb));
lib/sbitmap.c
384
seq_printf(m, "cleared=%u\n", sbitmap_cleared(sb));
lib/sbitmap.c
385
seq_printf(m, "bits_per_word=%u\n", 1U << sb->shift);
lib/sbitmap.c
386
seq_printf(m, "map_nr=%u\n", sb->map_nr);
lib/sbitmap.c
402
void sbitmap_bitmap_show(struct sbitmap *sb, struct seq_file *m)
lib/sbitmap.c
409
for (i = 0; i < sb->map_nr; i++) {
lib/sbitmap.c
410
unsigned long word = READ_ONCE(sb->map[i].word);
lib/sbitmap.c
411
unsigned long cleared = READ_ONCE(sb->map[i].cleared);
lib/sbitmap.c
412
unsigned int word_bits = __map_depth(sb, i);
lib/sbitmap.c
43
static inline void update_alloc_hint_after_get(struct sbitmap *sb,
lib/sbitmap.c
454
ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node,
lib/sbitmap.c
468
sbitmap_free(&sbq->sb);
lib/sbitmap.c
493
unsigned int depth = (sbq->sb.depth + users - 1) / users;
lib/sbitmap.c
50
this_cpu_write(*sb->alloc_hint, 0);
lib/sbitmap.c
505
sbitmap_resize(&sbq->sb, depth);
lib/sbitmap.c
51
} else if (nr == hint || unlikely(sb->round_robin)) {
lib/sbitmap.c
511
return sbitmap_get(&sbq->sb);
lib/sbitmap.c
518
struct sbitmap *sb = &sbq->sb;
lib/sbitmap.c
523
if (unlikely(sb->round_robin))
lib/sbitmap.c
526
depth = READ_ONCE(sb->depth);
lib/sbitmap.c
527
hint = update_alloc_hint_before_get(sb, depth);
lib/sbitmap.c
529
index = SB_NR_TO_INDEX(sb, hint);
lib/sbitmap.c
531
for (i = 0; i < sb->map_nr; i++) {
lib/sbitmap.c
532
struct sbitmap_word *map = &sb->map[index];
lib/sbitmap.c
534
unsigned int map_depth = __map_depth(sb, index);
lib/sbitmap.c
552
*offset = nr + (index << sb->shift);
lib/sbitmap.c
553
update_alloc_hint_after_get(sb, depth, hint,
lib/sbitmap.c
56
this_cpu_write(*sb->alloc_hint, hint);
lib/sbitmap.c
560
if (++index >= sb->map_nr)
lib/sbitmap.c
572
return sbitmap_get_shallow(&sbq->sb, shallow_depth);
lib/sbitmap.c
580
sbitmap_queue_update_wake_batch(sbq, sbq->sb.depth);
lib/sbitmap.c
636
static inline void sbitmap_update_cpu_hint(struct sbitmap *sb, int cpu, int tag)
lib/sbitmap.c
638
if (likely(!sb->round_robin && tag < sb->depth))
lib/sbitmap.c
639
data_race(*per_cpu_ptr(sb->alloc_hint, cpu) = tag);
lib/sbitmap.c
645
struct sbitmap *sb = &sbq->sb;
lib/sbitmap.c
656
this_addr = &sb->map[SB_NR_TO_INDEX(sb, tag)].word;
lib/sbitmap.c
664
mask |= (1UL << SB_NR_TO_BIT(sb, tag));
lib/sbitmap.c
672
sbitmap_update_cpu_hint(&sbq->sb, raw_smp_processor_id(),
lib/sbitmap.c
690
sbitmap_deferred_clear_bit(&sbq->sb, nr);
lib/sbitmap.c
700
sbitmap_update_cpu_hint(&sbq->sb, cpu, nr);
lib/sbitmap.c
730
sbitmap_show(&sbq->sb, m);
lib/sbitmap.c
738
seq_printf(m, "%u", *per_cpu_ptr(sbq->sb.alloc_hint, i));
lib/sbitmap.c
754
seq_printf(m, "round_robin=%d\n", sbq->sb.round_robin);
lib/test_bpf.c
848
int di, si, bt, db, sb;
lib/test_bpf.c
875
for (sb = -(block1 / 2);
lib/test_bpf.c
876
sb < (block1 + 1) / 2; sb++) {
lib/test_bpf.c
880
src = value(si, sb, sgn[k][1]);
lib/test_bpf.c
893
for (sb = -(block2 / 2);
lib/test_bpf.c
894
sb < (block2 + 1) / 2; sb++) {
lib/test_bpf.c
898
src = value(bt % sbits, sb, sgn[k][1]);
mm/backing-dev.c
1202
struct super_block *sb;
mm/backing-dev.c
1207
sb = inode->i_sb;
mm/backing-dev.c
1209
if (sb_is_blkdev_sb(sb))
mm/backing-dev.c
1212
return sb->s_bdi;
mm/shmem.c
164
static inline struct shmem_sb_info *SHMEM_SB(struct super_block *sb)
mm/shmem.c
166
return sb->s_fs_info;
mm/shmem.c
2819
struct super_block *sb;
mm/shmem.c
2825
sb = file_inode(file)->i_sb;
mm/shmem.c
2833
sb = shm_mnt->mnt_sb;
mm/shmem.c
2843
if (SHMEM_SB(sb)->huge != SHMEM_HUGE_NEVER)
mm/shmem.c
2852
if (SHMEM_SB(sb)->huge == SHMEM_HUGE_NEVER && !order)
mm/shmem.c
298
static int shmem_enable_quotas(struct super_block *sb,
mm/shmem.c
2993
struct super_block *sb = inode->i_sb;
mm/shmem.c
2997
if (!sb->s_encoding)
mm/shmem.c
303
sb_dqopt(sb)->flags |= DQUOT_QUOTA_SYS_FILE | DQUOT_NOLIST_DIRTY;
mm/shmem.c
3065
struct super_block *sb,
mm/shmem.c
307
err = dquot_load_quota_sb(sb, type, QFMT_SHMEM,
mm/shmem.c
3071
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem.c
3075
err = shmem_reserve_inode(sb, &ino);
mm/shmem.c
3079
inode = new_inode(sb);
mm/shmem.c
3081
shmem_free_inode(sb, 0);
mm/shmem.c
3148
struct super_block *sb, struct inode *dir,
mm/shmem.c
3154
inode = __shmem_get_inode(idmap, sb, dir, mode, dev, flags);
mm/shmem.c
3176
struct super_block *sb, struct inode *dir,
mm/shmem.c
3179
return __shmem_get_inode(idmap, sb, dir, mode, dev, flags);
mm/shmem.c
319
dquot_quota_off(sb, type);
mm/shmem.c
323
static void shmem_disable_quotas(struct super_block *sb)
mm/shmem.c
328
dquot_quota_off(sb, type);
mm/shmem.c
347
static int shmem_reserve_inode(struct super_block *sb, ino_t *inop)
mm/shmem.c
349
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem.c
352
if (!(sb->s_flags & SB_KERNMOUNT)) {
mm/shmem.c
373
__func__, MINOR(sb->s_dev));
mm/shmem.c
413
static void shmem_free_inode(struct super_block *sb, size_t freed_ispace)
mm/shmem.c
415
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem.c
4437
static struct dentry *shmem_fh_to_dentry(struct super_block *sb,
mm/shmem.c
4450
inode = ilookup5(sb, (unsigned long)(inum + fid->raw[0]),
mm/shmem.c
4980
static void shmem_put_super(struct super_block *sb)
mm/shmem.c
4982
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem.c
4985
if (sb->s_encoding)
mm/shmem.c
4986
utf8_unload(sb->s_encoding);
mm/shmem.c
4990
shmem_disable_quotas(sb);
mm/shmem.c
4996
sb->s_fs_info = NULL;
mm/shmem.c
5006
static int shmem_fill_super(struct super_block *sb, struct fs_context *fc)
mm/shmem.c
5019
sb->s_fs_info = sbinfo;
mm/shmem.c
5027
if (!(sb->s_flags & SB_KERNMOUNT)) {
mm/shmem.c
5036
sb->s_flags |= SB_NOUSER;
mm/shmem.c
5038
sb->s_export_op = &shmem_export_ops;
mm/shmem.c
5039
sb->s_flags |= SB_NOSEC;
mm/shmem.c
5049
sb->s_encoding = ctx->encoding;
mm/shmem.c
5050
set_default_d_op(sb, &shmem_ci_dentry_ops);
mm/shmem.c
5052
sb->s_encoding_flags = SB_ENC_STRICT_MODE_FL;
mm/shmem.c
5057
sb->s_flags |= SB_NOUSER;
mm/shmem.c
5059
sb->s_d_flags |= DCACHE_DONTCACHE;
mm/shmem.c
5063
if (sb->s_flags & SB_KERNMOUNT) {
mm/shmem.c
5087
sb->s_maxbytes = MAX_LFS_FILESIZE;
mm/shmem.c
5088
sb->s_blocksize = PAGE_SIZE;
mm/shmem.c
5089
sb->s_blocksize_bits = PAGE_SHIFT;
mm/shmem.c
5090
sb->s_magic = TMPFS_MAGIC;
mm/shmem.c
5091
sb->s_op = &shmem_ops;
mm/shmem.c
5092
sb->s_time_gran = 1;
mm/shmem.c
5094
sb->s_xattr = shmem_xattr_handlers;
mm/shmem.c
5097
sb->s_flags |= SB_POSIXACL;
mm/shmem.c
5101
super_set_uuid(sb, uuid.b, sizeof(uuid));
mm/shmem.c
5105
sb->dq_op = &shmem_quota_operations;
mm/shmem.c
5106
sb->s_qcop = &dquot_quotactl_sysfile_ops;
mm/shmem.c
5107
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP;
mm/shmem.c
5113
if (shmem_enable_quotas(sb, ctx->quota_types))
mm/shmem.c
5118
inode = shmem_get_inode(&nop_mnt_idmap, sb, NULL,
mm/shmem.c
5127
sb->s_root = d_make_root(inode);
mm/shmem.c
5128
if (!sb->s_root)
mm/shmem.c
5133
shmem_put_super(sb);
mm/shmem.c
5164
static struct inode *shmem_alloc_inode(struct super_block *sb)
mm/shmem.c
5167
info = alloc_inode_sb(sb, shmem_inode_cachep, GFP_KERNEL);
mm/shmem.c
5820
struct super_block *sb, struct inode *dir,
mm/shmem.c
5823
struct inode *inode = ramfs_get_inode(sb, dir, mode, dev);
mm/shmem.c
834
static long shmem_unused_huge_scan(struct super_block *sb,
mm/shmem.c
837
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem.c
845
static long shmem_unused_huge_count(struct super_block *sb,
mm/shmem.c
848
struct shmem_sb_info *sbinfo = SHMEM_SB(sb);
mm/shmem_quota.c
114
static int shmem_get_next_id(struct super_block *sb, struct kqid *qid)
mm/shmem_quota.c
116
struct mem_dqinfo *info = sb_dqinfo(sb, qid->type);
mm/shmem_quota.c
119
struct quota_info *dqopt = sb_dqopt(sb);
mm/shmem_quota.c
123
if (!sb_has_quota_active(sb, qid->type))
mm/shmem_quota.c
325
static int shmem_dquot_write_info(struct super_block *sb, int type)
mm/shmem_quota.c
55
static int shmem_check_quota_file(struct super_block *sb, int type)
mm/shmem_quota.c
65
static int shmem_read_file_info(struct super_block *sb, int type)
mm/shmem_quota.c
67
struct quota_info *dqopt = sb_dqopt(sb);
mm/shmem_quota.c
84
static int shmem_write_file_info(struct super_block *sb, int type)
mm/shmem_quota.c
93
static int shmem_free_file_info(struct super_block *sb, int type)
mm/shmem_quota.c
95
struct mem_dqinfo *info = &sb_dqopt(sb)->info[type];
net/9p/client.c
1819
int p9_client_statfs(struct p9_fid *fid, struct p9_rstatfs *sb)
net/9p/client.c
1835
err = p9pdu_readf(&req->rc, clnt->proto_version, "ddqqqqqqd", &sb->type,
net/9p/client.c
1836
&sb->bsize, &sb->blocks, &sb->bfree, &sb->bavail,
net/9p/client.c
1837
&sb->files, &sb->ffree, &sb->fsid, &sb->namelen);
net/9p/client.c
1846
fid->fid, sb->type, sb->bsize, sb->blocks, sb->bfree,
net/9p/client.c
1847
sb->bavail, sb->files, sb->ffree, sb->fsid, sb->namelen);
net/ceph/debugfs.c
87
char sb[64];
net/ceph/debugfs.c
92
ceph_osdmap_state_str(sb, sizeof(sb), state),
net/core/gen_stats.c
226
struct gnet_stats_basic sb;
net/core/gen_stats.c
229
memset(&sb, 0, sizeof(sb));
net/core/gen_stats.c
230
sb.bytes = bstats_bytes;
net/core/gen_stats.c
231
sb.packets = bstats_packets;
net/core/gen_stats.c
232
res = gnet_stats_copy(d, type, &sb, sizeof(sb), TCA_STATS_PAD);
net/core/gen_stats.c
233
if (res < 0 || sb.packets == bstats_packets)
net/mac80211/tests/s1g_tim.c
63
for (u8 sb = 0; sb < 8; sb++) {
net/mac80211/tests/s1g_tim.c
64
if (!(blkmap & BIT(sb)))
net/mac80211/tests/s1g_tim.c
72
oct, sb, bits, sub);
net/mac80211/tx.c
5493
struct s1g_short_beacon_data *sb,
net/mac80211/tx.c
5501
skb = dev_alloc_skb(local->tx_headroom + sb->short_head_len +
net/mac80211/tx.c
5502
sb->short_tail_len + 256 +
net/mac80211/tx.c
5508
skb_put_data(skb, sb->short_head, sb->short_head_len);
net/mac80211/tx.c
5512
if (sb->short_tail)
net/mac80211/tx.c
5513
skb_put_data(skb, sb->short_tail, sb->short_tail_len);
net/netfilter/ipvs/ip_vs_sync.c
1640
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
1642
sb = sb_dequeue(ipvs, ms);
net/netfilter/ipvs/ip_vs_sync.c
1643
if (sb)
net/netfilter/ipvs/ip_vs_sync.c
1644
return sb;
net/netfilter/ipvs/ip_vs_sync.c
1655
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
1662
sb = next_sync_buff(ipvs, ms);
net/netfilter/ipvs/ip_vs_sync.c
1665
if (!sb) {
net/netfilter/ipvs/ip_vs_sync.c
1669
while (ip_vs_send_sync_msg(tinfo->sock, sb->mesg) < 0) {
net/netfilter/ipvs/ip_vs_sync.c
1679
ip_vs_sync_buff_release(sb);
net/netfilter/ipvs/ip_vs_sync.c
1684
if (sb)
net/netfilter/ipvs/ip_vs_sync.c
1685
ip_vs_sync_buff_release(sb);
net/netfilter/ipvs/ip_vs_sync.c
1688
while ((sb = sb_dequeue(ipvs, ms)))
net/netfilter/ipvs/ip_vs_sync.c
1689
ip_vs_sync_buff_release(sb);
net/netfilter/ipvs/ip_vs_sync.c
1693
sb = get_curr_sync_buff(ipvs, ms, 0);
net/netfilter/ipvs/ip_vs_sync.c
1694
if (sb)
net/netfilter/ipvs/ip_vs_sync.c
1695
ip_vs_sync_buff_release(sb);
net/netfilter/ipvs/ip_vs_sync.c
305
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
309
sb = NULL;
net/netfilter/ipvs/ip_vs_sync.c
312
sb = list_entry(ms->sync_queue.next, struct ip_vs_sync_buff,
net/netfilter/ipvs/ip_vs_sync.c
314
list_del(&sb->list);
net/netfilter/ipvs/ip_vs_sync.c
321
return sb;
net/netfilter/ipvs/ip_vs_sync.c
330
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
332
if (!(sb=kmalloc_obj(struct ip_vs_sync_buff, GFP_ATOMIC)))
net/netfilter/ipvs/ip_vs_sync.c
337
sb->mesg = kmalloc(len, GFP_ATOMIC);
net/netfilter/ipvs/ip_vs_sync.c
338
if (!sb->mesg) {
net/netfilter/ipvs/ip_vs_sync.c
339
kfree(sb);
net/netfilter/ipvs/ip_vs_sync.c
342
sb->mesg->reserved = 0; /* old nr_conns i.e. must be zero now */
net/netfilter/ipvs/ip_vs_sync.c
343
sb->mesg->version = SYNC_PROTO_VER;
net/netfilter/ipvs/ip_vs_sync.c
344
sb->mesg->syncid = ipvs->mcfg.syncid;
net/netfilter/ipvs/ip_vs_sync.c
345
sb->mesg->size = htons(sizeof(struct ip_vs_sync_mesg));
net/netfilter/ipvs/ip_vs_sync.c
346
sb->mesg->nr_conns = 0;
net/netfilter/ipvs/ip_vs_sync.c
347
sb->mesg->spare = 0;
net/netfilter/ipvs/ip_vs_sync.c
348
sb->head = (unsigned char *)sb->mesg + sizeof(struct ip_vs_sync_mesg);
net/netfilter/ipvs/ip_vs_sync.c
349
sb->end = (unsigned char *)sb->mesg + len;
net/netfilter/ipvs/ip_vs_sync.c
351
sb->firstuse = jiffies;
net/netfilter/ipvs/ip_vs_sync.c
352
return sb;
net/netfilter/ipvs/ip_vs_sync.c
355
static inline void ip_vs_sync_buff_release(struct ip_vs_sync_buff *sb)
net/netfilter/ipvs/ip_vs_sync.c
357
kfree(sb->mesg);
net/netfilter/ipvs/ip_vs_sync.c
358
kfree(sb);
net/netfilter/ipvs/ip_vs_sync.c
364
struct ip_vs_sync_buff *sb = ms->sync_buff;
net/netfilter/ipvs/ip_vs_sync.c
373
list_add_tail(&sb->list, &ms->sync_queue);
net/netfilter/ipvs/ip_vs_sync.c
380
ip_vs_sync_buff_release(sb);
net/netfilter/ipvs/ip_vs_sync.c
392
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
395
sb = ms->sync_buff;
net/netfilter/ipvs/ip_vs_sync.c
396
if (sb && time_after_eq(jiffies - sb->firstuse, time)) {
net/netfilter/ipvs/ip_vs_sync.c
400
sb = NULL;
net/netfilter/ipvs/ip_vs_sync.c
402
return sb;
net/netfilter/ipvs/ip_vs_sync.c
417
struct ip_vs_sync_buff *sb;
net/netfilter/ipvs/ip_vs_sync.c
420
if (!(sb=kmalloc_obj(struct ip_vs_sync_buff, GFP_ATOMIC)))
net/netfilter/ipvs/ip_vs_sync.c
425
sb->mesg = kmalloc(len, GFP_ATOMIC);
net/netfilter/ipvs/ip_vs_sync.c
426
if (!sb->mesg) {
net/netfilter/ipvs/ip_vs_sync.c
427
kfree(sb);
net/netfilter/ipvs/ip_vs_sync.c
430
mesg = (struct ip_vs_sync_mesg_v0 *)sb->mesg;
net/netfilter/ipvs/ip_vs_sync.c
434
sb->head = (unsigned char *)mesg + sizeof(struct ip_vs_sync_mesg_v0);
net/netfilter/ipvs/ip_vs_sync.c
435
sb->end = (unsigned char *)mesg + len;
net/netfilter/ipvs/ip_vs_sync.c
436
sb->firstuse = jiffies;
net/netfilter/ipvs/ip_vs_sync.c
437
return sb;
net/socket.c
318
static struct inode *sock_alloc_inode(struct super_block *sb)
net/socket.c
322
ei = alloc_inode_sb(sb, sock_inode_cachep, GFP_KERNEL);
net/sunrpc/clnt.c
115
static int rpc_setup_pipedir_sb(struct super_block *sb,
net/sunrpc/clnt.c
124
dir = rpc_d_lookup_sb(sb, dir_name);
net/sunrpc/clnt.c
180
struct super_block *sb)
net/sunrpc/clnt.c
184
return rpc_setup_pipedir_sb(sb, clnt);
net/sunrpc/clnt.c
196
struct super_block *sb)
net/sunrpc/clnt.c
202
error = __rpc_clnt_handle_event(clnt, event, sb);
net/sunrpc/clnt.c
228
struct super_block *sb = ptr;
net/sunrpc/clnt.c
232
while ((clnt = rpc_get_client_for_event(sb->s_fs_info, event))) {
net/sunrpc/clnt.c
233
error = __rpc_pipefs_event(clnt, event, sb);
net/sunrpc/rpc_pipe.c
1005
struct dentry *rpc_d_lookup_sb(const struct super_block *sb,
net/sunrpc/rpc_pipe.c
1008
return try_lookup_noperm(&QSTR(dir_name), sb->s_root);
net/sunrpc/rpc_pipe.c
1117
rpc_fill_super(struct super_block *sb, struct fs_context *fc)
net/sunrpc/rpc_pipe.c
1121
struct net *net = sb->s_fs_info;
net/sunrpc/rpc_pipe.c
1125
sb->s_blocksize = PAGE_SIZE;
net/sunrpc/rpc_pipe.c
1126
sb->s_blocksize_bits = PAGE_SHIFT;
net/sunrpc/rpc_pipe.c
1127
sb->s_magic = RPCAUTH_GSSMAGIC;
net/sunrpc/rpc_pipe.c
1128
sb->s_op = &s_ops;
net/sunrpc/rpc_pipe.c
1129
sb->s_d_flags = DCACHE_DONTCACHE;
net/sunrpc/rpc_pipe.c
1130
sb->s_time_gran = 1;
net/sunrpc/rpc_pipe.c
1132
inode = rpc_get_inode(sb, S_IFDIR | 0555);
net/sunrpc/rpc_pipe.c
1133
sb->s_root = root = d_make_root(inode);
net/sunrpc/rpc_pipe.c
1146
sn->pipefs_sb = sb;
net/sunrpc/rpc_pipe.c
1149
sb);
net/sunrpc/rpc_pipe.c
1188
static void rpc_kill_sb(struct super_block *sb)
net/sunrpc/rpc_pipe.c
1190
struct net *net = sb->s_fs_info;
net/sunrpc/rpc_pipe.c
1194
if (sn->pipefs_sb != sb) {
net/sunrpc/rpc_pipe.c
1203
sb);
net/sunrpc/rpc_pipe.c
1206
kill_anon_super(sb);
net/sunrpc/rpc_pipe.c
198
rpc_alloc_inode(struct super_block *sb)
net/sunrpc/rpc_pipe.c
201
rpci = alloc_inode_sb(sb, rpc_inode_cachep, GFP_KERNEL);
net/sunrpc/rpc_pipe.c
468
rpc_get_inode(struct super_block *sb, umode_t mode)
net/sunrpc/rpc_pipe.c
470
struct inode *inode = new_inode(sb);
net/wireless/nl80211.c
6639
struct cfg80211_s1g_short_beacon *sb)
net/wireless/nl80211.c
6656
sb->short_head = nla_data(tb[NL80211_S1G_SHORT_BEACON_ATTR_HEAD]);
net/wireless/nl80211.c
6657
sb->short_head_len = nla_len(tb[NL80211_S1G_SHORT_BEACON_ATTR_HEAD]);
net/wireless/nl80211.c
6658
sb->short_tail_len = 0;
net/wireless/nl80211.c
6661
sb->short_tail =
net/wireless/nl80211.c
6663
sb->short_tail_len =
net/wireless/nl80211.c
6667
sb->update = true;
scripts/elf-parse.c
25
struct stat sb;
scripts/elf-parse.c
33
if (fstat(fd, &sb) < 0) {
scripts/elf-parse.c
37
if (!S_ISREG(sb.st_mode)) {
scripts/elf-parse.c
42
addr = mmap(0, sb.st_size, PROT_READ|PROT_WRITE, MAP_SHARED, fd, 0);
scripts/elf-parse.c
48
*size = sb.st_size;
scripts/kallsyms.c
318
const struct sym_entry *sb = *(const struct sym_entry **)b;
scripts/kallsyms.c
320
ret = strcmp(sym_name(sa), sym_name(sb));
scripts/kallsyms.c
322
if (sa->addr > sb->addr)
scripts/kallsyms.c
324
else if (sa->addr < sb->addr)
scripts/kallsyms.c
328
return (int)(sa->seq - sb->seq);
scripts/kallsyms.c
649
const struct sym_entry *sb = *(const struct sym_entry **)b;
scripts/kallsyms.c
653
if (sa->addr > sb->addr)
scripts/kallsyms.c
655
if (sa->addr < sb->addr)
scripts/kallsyms.c
660
wb = (sb->sym[0] == 'w') || (sb->sym[0] == 'W');
scripts/kallsyms.c
666
wb = may_be_linker_script_provide_symbol(sb);
scripts/kallsyms.c
672
wb = strspn(sym_name(sb), "_");
scripts/kallsyms.c
677
return sa->seq - sb->seq;
scripts/recordmcount.c
184
sb.st_size = 0;
scripts/recordmcount.c
191
if (fstat(fd_map, &sb) < 0) {
scripts/recordmcount.c
195
if (!S_ISREG(sb.st_mode)) {
scripts/recordmcount.c
199
file_map = mmap(0, sb.st_size, PROT_READ|PROT_WRITE, MAP_PRIVATE,
scripts/recordmcount.c
203
file_map = umalloc(sb.st_size);
scripts/recordmcount.c
208
if (read(fd_map, file_map, sb.st_size) != sb.st_size) {
scripts/recordmcount.c
220
file_end = file_map + sb.st_size;
scripts/recordmcount.c
344
fd_map = open(tmp_file, O_WRONLY | O_TRUNC | O_CREAT, sb.st_mode);
scripts/recordmcount.c
349
n = write(fd_map, file_map, sb.st_size);
scripts/recordmcount.c
350
if (n != sb.st_size) {
scripts/recordmcount.c
58
static struct stat sb; /* Remember .st_size, etc. */
scripts/recordmcount.c
81
munmap(file_map, sb.st_size);
scripts/recordmcount.c
99
file_ptr = file_map + (sb.st_size - offset);
scripts/recordmcount.h
282
shstr->sh_offset = _w(sb.st_size);
scripts/recordmcount.h
283
t += sb.st_size;
scripts/recordmcount.h
290
if (ulseek(sb.st_size, SEEK_SET) < 0)
security/apparmor/apparmorfs.c
237
static int apparmorfs_fill_super(struct super_block *sb, struct fs_context *fc)
security/apparmor/apparmorfs.c
242
error = simple_fill_super(sb, AAFS_MAGIC, files);
security/apparmor/apparmorfs.c
245
sb->s_op = &aafs_super_ops;
security/inode.c
42
static int securityfs_fill_super(struct super_block *sb, struct fs_context *fc)
security/inode.c
47
error = simple_fill_super(sb, SECURITYFS_MAGIC, files);
security/inode.c
51
sb->s_op = &securityfs_super_operations;
security/ipe/ipe.c
29
struct ipe_superblock *ipe_sb(const struct super_block *sb)
security/ipe/ipe.c
31
return sb->s_security + ipe_blobs.lbs_superblock;
security/ipe/ipe.h
15
struct ipe_superblock *ipe_sb(const struct super_block *sb);
security/landlock/fs.c
1270
static void hook_sb_delete(struct super_block *const sb)
security/landlock/fs.c
1277
spin_lock(&sb->s_inode_list_lock);
security/landlock/fs.c
1278
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
security/landlock/fs.c
1353
spin_unlock(&sb->s_inode_list_lock);
security/landlock/fs.c
1360
spin_lock(&sb->s_inode_list_lock);
security/landlock/fs.c
1364
spin_unlock(&sb->s_inode_list_lock);
security/landlock/fs.c
1370
wait_var_event(&landlock_superblock(sb)->inode_refs,
security/landlock/fs.c
1371
!atomic_long_read(&landlock_superblock(sb)->inode_refs));
security/landlock/fs.c
1469
static int hook_sb_remount(struct super_block *const sb, void *const mnt_opts)
security/landlock/fs.c
1479
log_fs_change_topology_dentry(subject, handle_layer, sb->s_root);
security/landlock/fs.c
59
struct super_block *sb;
security/landlock/fs.c
75
sb = inode->i_sb;
security/landlock/fs.c
76
atomic_long_inc(&landlock_superblock(sb)->inode_refs);
security/landlock/fs.c
90
if (atomic_long_dec_and_test(&landlock_superblock(sb)->inode_refs))
security/landlock/fs.c
91
wake_up_var(&landlock_superblock(sb)->inode_refs);
security/security.c
1010
int security_sb_kern_mount(const struct super_block *sb)
security/security.c
1012
return call_int_hook(sb_kern_mount, sb);
security/security.c
1024
int security_sb_show_options(struct seq_file *m, struct super_block *sb)
security/security.c
1026
return call_int_hook(sb_show_options, m, sb);
security/security.c
1106
int security_sb_set_mnt_opts(struct super_block *sb,
security/security.c
1115
rc = scall->hl->hook.sb_set_mnt_opts(sb, mnt_opts, kern_flags,
security/security.c
353
static int lsm_superblock_alloc(struct super_block *sb)
security/security.c
355
return lsm_blob_alloc(&sb->s_security, blob_sizes.lbs_superblock,
security/security.c
649
int security_quotactl(int cmds, int type, int id, const struct super_block *sb)
security/security.c
651
return call_int_hook(quotactl, cmds, type, id, sb);
security/security.c
900
int security_sb_alloc(struct super_block *sb)
security/security.c
902
int rc = lsm_superblock_alloc(sb);
security/security.c
906
rc = call_int_hook(sb_alloc_security, sb);
security/security.c
908
security_sb_free(sb);
security/security.c
919
void security_sb_delete(struct super_block *sb)
security/security.c
921
call_void_hook(sb_delete, sb);
security/security.c
931
void security_sb_free(struct super_block *sb)
security/security.c
933
call_void_hook(sb_free_security, sb);
security/security.c
934
kfree(sb->s_security);
security/security.c
935
sb->s_security = NULL;
security/security.c
978
int security_sb_mnt_opts_compat(struct super_block *sb,
security/security.c
981
return call_int_hook(sb_mnt_opts_compat, sb, mnt_opts);
security/security.c
995
int security_sb_remount(struct super_block *sb,
security/security.c
998
return call_int_hook(sb_remount, sb, mnt_opts);
security/selinux/hooks.c
1096
static int selinux_sb_show_options(struct seq_file *m, struct super_block *sb)
security/selinux/hooks.c
1098
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
1129
struct dentry *root = sb->s_root;
security/selinux/hooks.c
1336
struct super_block *sb = dentry->d_sb;
security/selinux/hooks.c
1356
rc = security_genfs_sid(sb->s_type->name,
security/selinux/hooks.c
1966
const struct super_block *sb,
security/selinux/hooks.c
1973
sbsec = selinux_superblock(sb);
security/selinux/hooks.c
2168
static int selinux_quotactl(int cmds, int type, int id, const struct super_block *sb)
security/selinux/hooks.c
2173
if (!sb)
security/selinux/hooks.c
2185
rc = superblock_has_perm(cred, sb, FILESYSTEM__QUOTAMOD, NULL);
security/selinux/hooks.c
2194
rc = superblock_has_perm(cred, sb, FILESYSTEM__QUOTAGET, NULL);
security/selinux/hooks.c
2583
static int selinux_sb_alloc_security(struct super_block *sb)
security/selinux/hooks.c
2585
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
2674
static int selinux_sb_mnt_opts_compat(struct super_block *sb, void *mnt_opts)
security/selinux/hooks.c
2677
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
2706
root_isec = backing_inode_security(sb->s_root);
security/selinux/hooks.c
2719
static int selinux_sb_remount(struct super_block *sb, void *mnt_opts)
security/selinux/hooks.c
2722
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
2742
root_isec = backing_inode_security(sb->s_root);
security/selinux/hooks.c
2756
"during remount (dev %s, type=%s)\n", sb->s_id,
security/selinux/hooks.c
2757
sb->s_type->name);
security/selinux/hooks.c
2761
static int selinux_sb_kern_mount(const struct super_block *sb)
security/selinux/hooks.c
2767
ad.u.dentry = sb->s_root;
security/selinux/hooks.c
2768
return superblock_has_perm(cred, sb, FILESYSTEM__MOUNT, &ad);
security/selinux/hooks.c
471
static int selinux_is_genfs_special_handling(struct super_block *sb)
security/selinux/hooks.c
474
return !strcmp(sb->s_type->name, "sysfs") ||
security/selinux/hooks.c
475
!strcmp(sb->s_type->name, "pstore") ||
security/selinux/hooks.c
476
!strcmp(sb->s_type->name, "debugfs") ||
security/selinux/hooks.c
477
!strcmp(sb->s_type->name, "tracefs") ||
security/selinux/hooks.c
478
!strcmp(sb->s_type->name, "rootfs") ||
security/selinux/hooks.c
480
(!strcmp(sb->s_type->name, "cgroup") ||
security/selinux/hooks.c
481
!strcmp(sb->s_type->name, "cgroup2"))) ||
security/selinux/hooks.c
483
!strcmp(sb->s_type->name, "functionfs"));
security/selinux/hooks.c
486
static int selinux_is_sblabel_mnt(struct super_block *sb)
security/selinux/hooks.c
488
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
504
return selinux_is_genfs_special_handling(sb);
security/selinux/hooks.c
514
static int sb_check_xattr_support(struct super_block *sb)
security/selinux/hooks.c
516
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
517
struct dentry *root = sb->s_root;
security/selinux/hooks.c
531
sb->s_id, sb->s_type->name);
security/selinux/hooks.c
539
sb->s_id, sb->s_type->name);
security/selinux/hooks.c
543
sb->s_id, sb->s_type->name, -rc);
security/selinux/hooks.c
551
rc = security_genfs_sid(sb->s_type->name, "/",
security/selinux/hooks.c
557
sb->s_id, sb->s_type->name);
security/selinux/hooks.c
563
static int sb_finish_set_opts(struct super_block *sb)
security/selinux/hooks.c
565
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
566
struct dentry *root = sb->s_root;
security/selinux/hooks.c
571
rc = sb_check_xattr_support(sb);
security/selinux/hooks.c
583
if (selinux_is_sblabel_mnt(sb))
security/selinux/hooks.c
639
static int selinux_set_mnt_opts(struct super_block *sb,
security/selinux/hooks.c
645
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/hooks.c
646
struct dentry *root = sb->s_root;
security/selinux/hooks.c
690
if ((sbsec->flags & SE_SBINITIALIZED) && (sb->s_type->fs_flags & FS_BINARY_MOUNTDATA)
security/selinux/hooks.c
7138
struct super_block *sb;
security/selinux/hooks.c
7147
sb = path.dentry->d_sb;
security/selinux/hooks.c
7148
sbsec = selinux_superblock(sb);
security/selinux/hooks.c
742
if (strcmp(sb->s_type->name, "proc") == 0)
security/selinux/hooks.c
745
if (!strcmp(sb->s_type->name, "debugfs") ||
security/selinux/hooks.c
746
!strcmp(sb->s_type->name, "tracefs") ||
security/selinux/hooks.c
747
!strcmp(sb->s_type->name, "binder") ||
security/selinux/hooks.c
748
!strcmp(sb->s_type->name, "bpf") ||
security/selinux/hooks.c
749
!strcmp(sb->s_type->name, "pstore") ||
security/selinux/hooks.c
750
!strcmp(sb->s_type->name, "securityfs") ||
security/selinux/hooks.c
752
!strcmp(sb->s_type->name, "functionfs")))
security/selinux/hooks.c
755
if (!strcmp(sb->s_type->name, "sysfs") ||
security/selinux/hooks.c
756
!strcmp(sb->s_type->name, "cgroup") ||
security/selinux/hooks.c
757
!strcmp(sb->s_type->name, "cgroup2"))
security/selinux/hooks.c
765
rc = security_fs_use(sb);
security/selinux/hooks.c
768
__func__, sb->s_type->name, rc);
security/selinux/hooks.c
7767
static void delayed_superblock_init(struct super_block *sb, void *unused)
security/selinux/hooks.c
7769
selinux_set_mnt_opts(sb, NULL, 0, NULL);
security/selinux/hooks.c
778
if (sb->s_user_ns != &init_user_ns &&
security/selinux/hooks.c
779
strcmp(sb->s_type->name, "tmpfs") &&
security/selinux/hooks.c
780
strcmp(sb->s_type->name, "ramfs") &&
security/selinux/hooks.c
781
strcmp(sb->s_type->name, "devpts") &&
security/selinux/hooks.c
782
strcmp(sb->s_type->name, "overlay")) {
security/selinux/hooks.c
879
rc = sb_finish_set_opts(sb);
security/selinux/hooks.c
886
"security settings for (dev %s, type %s)\n", sb->s_id,
security/selinux/hooks.c
887
sb->s_type->name);
security/selinux/include/security.h
356
int security_fs_use(struct super_block *sb);
security/selinux/selinuxfs.c
100
struct selinux_fs_info *fsi = sb->s_fs_info;
security/selinux/selinuxfs.c
109
kfree(sb->s_fs_info);
security/selinux/selinuxfs.c
110
sb->s_fs_info = NULL;
security/selinux/selinuxfs.c
1202
static struct inode *sel_make_inode(struct super_block *sb, umode_t mode)
security/selinux/selinuxfs.c
1204
struct inode *ret = new_inode(sb);
security/selinux/selinuxfs.c
1606
struct super_block *sb = dir->d_sb;
security/selinux/selinuxfs.c
1607
struct selinux_fs_info *fsi = sb->s_fs_info;
security/selinux/selinuxfs.c
1637
struct super_block *sb = dir->d_sb;
security/selinux/selinuxfs.c
1638
struct selinux_fs_info *fsi = sb->s_fs_info;
security/selinux/selinuxfs.c
1811
struct super_block *sb = dir->d_sb;
security/selinux/selinuxfs.c
1812
struct selinux_fs_info *fsi = sb->s_fs_info;
security/selinux/selinuxfs.c
1876
struct super_block *sb = dir->d_sb;
security/selinux/selinuxfs.c
1889
inode = sel_make_inode(sb, S_IFREG | 0444);
security/selinux/selinuxfs.c
1931
static struct dentry *sel_make_swapover_dir(struct super_block *sb,
security/selinux/selinuxfs.c
1934
struct dentry *dentry = d_alloc_name(sb->s_root, ".swapover");
security/selinux/selinuxfs.c
1940
inode = sel_make_inode(sb, S_IFDIR);
security/selinux/selinuxfs.c
1950
inode_lock(sb->s_root->d_inode);
security/selinux/selinuxfs.c
1952
inc_nlink(sb->s_root->d_inode);
security/selinux/selinuxfs.c
1953
inode_unlock(sb->s_root->d_inode);
security/selinux/selinuxfs.c
1960
static int sel_fill_super(struct super_block *sb, struct fs_context *fc)
security/selinux/selinuxfs.c
1991
ret = selinux_fs_info_create(sb);
security/selinux/selinuxfs.c
1995
ret = simple_fill_super(sb, SELINUX_MAGIC, selinux_files);
security/selinux/selinuxfs.c
1999
fsi = sb->s_fs_info;
security/selinux/selinuxfs.c
2000
fsi->bool_dir = sel_make_dir(sb->s_root, BOOL_DIR_NAME, &fsi->last_ino);
security/selinux/selinuxfs.c
2008
inode = sel_make_inode(sb, S_IFCHR | S_IRUGO | S_IWUGO);
security/selinux/selinuxfs.c
2019
ret = sel_attach_file(sb->s_root, NULL_FILE_NAME, inode);
security/selinux/selinuxfs.c
2023
dentry = sel_make_dir(sb->s_root, "avc", &fsi->last_ino);
security/selinux/selinuxfs.c
2033
dentry = sel_make_dir(sb->s_root, "ss", &fsi->last_ino);
security/selinux/selinuxfs.c
2043
dentry = sel_make_dir(sb->s_root, "initial_contexts", &fsi->last_ino);
security/selinux/selinuxfs.c
2053
fsi->class_dir = sel_make_dir(sb->s_root, CLASS_DIR_NAME, &fsi->last_ino);
security/selinux/selinuxfs.c
2060
dentry = sel_make_dir(sb->s_root, "policy_capabilities", &fsi->last_ino);
security/selinux/selinuxfs.c
2095
static void sel_kill_sb(struct super_block *sb)
security/selinux/selinuxfs.c
2097
selinux_fs_info_free(sb);
security/selinux/selinuxfs.c
2098
kill_anon_super(sb);
security/selinux/selinuxfs.c
339
static struct dentry *sel_make_swapover_dir(struct super_block *sb,
security/selinux/selinuxfs.c
515
tmp_parent = sel_make_swapover_dir(fsi->sb, &tmp_ino);
security/selinux/selinuxfs.c
544
rd.new_parent = fsi->sb->s_root;
security/selinux/selinuxfs.c
81
struct super_block *sb;
security/selinux/selinuxfs.c
84
static int selinux_fs_info_create(struct super_block *sb)
security/selinux/selinuxfs.c
93
fsi->sb = sb;
security/selinux/selinuxfs.c
94
sb->s_fs_info = fsi;
security/selinux/selinuxfs.c
98
static void selinux_fs_info_free(struct super_block *sb)
security/selinux/ss/services.c
2987
int security_fs_use(struct super_block *sb)
security/selinux/ss/services.c
2994
struct superblock_security_struct *sbsec = selinux_superblock(sb);
security/selinux/ss/services.c
2995
const char *fstype = sb->s_type->name;
security/smack/smack_lsm.c
531
static int smack_sb_alloc_security(struct super_block *sb)
security/smack/smack_lsm.c
533
struct superblock_smack *sbsp = smack_superblock(sb);
security/smack/smack_lsm.c
780
static int smack_set_mnt_opts(struct super_block *sb,
security/smack/smack_lsm.c
785
struct dentry *root = sb->s_root;
security/smack/smack_lsm.c
787
struct superblock_smack *sp = smack_superblock(sb);
security/smack/smack_lsm.c
813
if (sb->s_user_ns != &init_user_ns &&
security/smack/smack_lsm.c
814
sb->s_magic != SYSFS_MAGIC && sb->s_magic != TMPFS_MAGIC &&
security/smack/smack_lsm.c
815
sb->s_magic != RAMFS_MAGIC) {
security/smack/smackfs.c
2879
static int smk_fill_super(struct super_block *sb, struct fs_context *fc)
security/smack/smackfs.c
2940
rc = simple_fill_super(sb, SMACK_MAGIC, smack_files);
security/tomoyo/realpath.c
156
struct super_block *sb = dentry->d_sb;
security/tomoyo/realpath.c
162
if (sb->s_magic == PROC_SUPER_MAGIC && *pos == '/') {
security/tomoyo/realpath.c
165
struct pid_namespace *proc_pidns = proc_pid_ns(sb);
security/tomoyo/realpath.c
177
if (!MAJOR(sb->s_dev))
security/tomoyo/realpath.c
180
struct inode *inode = d_backing_inode(sb->s_root);
security/tomoyo/realpath.c
193
const dev_t dev = sb->s_dev;
security/tomoyo/realpath.c
208
const char *name = sb->s_type->name;
security/tomoyo/realpath.c
243
struct super_block *sb = dentry->d_sb;
security/tomoyo/realpath.c
261
inode = d_backing_inode(sb->s_root);
security/tomoyo/realpath.c
266
!(sb->s_type->fs_flags & FS_REQUIRES_DEV)))
sound/hda/codecs/side-codecs/tas2781_hda_i2c.c
68
struct acpi_resource_i2c_serialbus *sb;
sound/hda/codecs/side-codecs/tas2781_hda_i2c.c
70
if (i2c_acpi_get_i2c_resource(ares, &sb)) {
sound/hda/codecs/side-codecs/tas2781_hda_i2c.c
72
sb->slave_address != tas_priv->global_addr) {
sound/hda/codecs/side-codecs/tas2781_hda_i2c.c
74
(unsigned int)sb->slave_address;
sound/isa/cmi8330.c
156
struct snd_sb *sb;
sound/isa/cmi8330.c
310
err = cmi8330_add_sb_mixers(acard->sb);
sound/isa/cmi8330.c
447
chip->streams[CMI_SB_STREAM].private_data = chip->sb;
sound/isa/cmi8330.c
474
snd_sbmixer_suspend(acard->sb);
sound/isa/cmi8330.c
482
snd_sbdsp_reset(acard->sb);
sound/isa/cmi8330.c
483
snd_sbmixer_suspend(acard->sb);
sound/isa/cmi8330.c
544
SB_HW_AUTO, &acard->sb);
sound/isa/cmi8330.c
549
if (acard->sb->hardware != SB_HW_16) {
sound/isa/sb/sb_mixer.c
107
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
115
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
116
left = (snd_sbmixer_read(sb, left_reg) >> left_shift) & mask;
sound/isa/sb/sb_mixer.c
117
right = (snd_sbmixer_read(sb, right_reg) >> right_shift) & mask;
sound/isa/sb/sb_mixer.c
125
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
136
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
138
oleft = snd_sbmixer_read(sb, left_reg);
sound/isa/sb/sb_mixer.c
142
snd_sbmixer_write(sb, left_reg, left);
sound/isa/sb/sb_mixer.c
144
oleft = snd_sbmixer_read(sb, left_reg);
sound/isa/sb/sb_mixer.c
145
oright = snd_sbmixer_read(sb, right_reg);
sound/isa/sb/sb_mixer.c
150
snd_sbmixer_write(sb, left_reg, left);
sound/isa/sb/sb_mixer.c
151
snd_sbmixer_write(sb, right_reg, right);
sound/isa/sb/sb_mixer.c
172
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
175
scoped_guard(spinlock_irqsave, &sb->mixer_lock) {
sound/isa/sb/sb_mixer.c
176
oval = snd_sbmixer_read(sb, SB_DT019X_CAPTURE_SW);
sound/isa/sb/sb_mixer.c
207
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
232
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
233
oval = snd_sbmixer_read(sb, SB_DT019X_CAPTURE_SW);
sound/isa/sb/sb_mixer.c
236
snd_sbmixer_write(sb, SB_DT019X_CAPTURE_SW, nval);
sound/isa/sb/sb_mixer.c
257
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
260
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
261
oval = snd_sbmixer_read(sb, SB_ALS4000_MONO_IO_CTRL);
sound/isa/sb/sb_mixer.c
273
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
279
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
280
oval = snd_sbmixer_read(sb, SB_ALS4000_MONO_IO_CTRL);
sound/isa/sb/sb_mixer.c
286
snd_sbmixer_write(sb, SB_ALS4000_MONO_IO_CTRL, nval);
sound/isa/sb/sb_mixer.c
306
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
309
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
310
oval = snd_sbmixer_read(sb, SB_DSP_CAPTURE_SOURCE);
sound/isa/sb/sb_mixer.c
327
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
344
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
345
oval = snd_sbmixer_read(sb, SB_DSP_CAPTURE_SOURCE);
sound/isa/sb/sb_mixer.c
349
snd_sbmixer_write(sb, SB_DSP_CAPTURE_SOURCE, nval);
sound/isa/sb/sb_mixer.c
368
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
375
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
376
val1 = snd_sbmixer_read(sb, reg1);
sound/isa/sb/sb_mixer.c
377
val2 = snd_sbmixer_read(sb, reg2);
sound/isa/sb/sb_mixer.c
387
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
395
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
396
oval1 = snd_sbmixer_read(sb, reg1);
sound/isa/sb/sb_mixer.c
397
oval2 = snd_sbmixer_read(sb, reg2);
sound/isa/sb/sb_mixer.c
406
snd_sbmixer_write(sb, reg1, val1);
sound/isa/sb/sb_mixer.c
407
snd_sbmixer_write(sb, reg2, val2);
sound/isa/sb/sb_mixer.c
59
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
65
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
66
val = (snd_sbmixer_read(sb, reg) >> shift) & mask;
sound/isa/sb/sb_mixer.c
73
struct snd_sb *sb = snd_kcontrol_chip(kcontrol);
sound/isa/sb/sb_mixer.c
81
guard(spinlock_irqsave)(&sb->mixer_lock);
sound/isa/sb/sb_mixer.c
82
oval = snd_sbmixer_read(sb, reg);
sound/isa/sb/sb_mixer.c
86
snd_sbmixer_write(sb, reg, val);
sound/pci/als4000.c
210
static inline void snd_als4k_gcr_write(struct snd_sb *sb,
sound/pci/als4000.c
214
snd_als4k_gcr_write_addr(sb->alt_port, reg, val);
sound/pci/als4000.c
225
static inline u32 snd_als4k_gcr_read(struct snd_sb *sb, enum als4k_gcr_t reg)
sound/pci/als4000.c
227
return snd_als4k_gcr_read_addr(sb->alt_port, reg);
sound/pci/au88x0/au88x0.h
197
static void vortex_adbdma_setstartbuffer(vortex_t * vortex, int adbdma, int sb);
sound/pci/au88x0/au88x0.h
203
static void vortex_wtdma_setstartbuffer(vortex_t * vortex, int wtdma, int sb);
sound/pci/au88x0/au88x0_core.c
1078
static void vortex_adbdma_setstartbuffer(vortex_t * vortex, int adbdma, int sb)
sound/pci/au88x0/au88x0_core.c
1083
sb << ((0xf - (adbdma & 0xf)) * 2));
sound/pci/au88x0/au88x0_core.c
1084
dma->period_real = dma->period_virt = sb;
sound/pci/au88x0/au88x0_core.c
1357
static void vortex_wtdma_setstartbuffer(vortex_t * vortex, int wtdma, int sb)
sound/pci/au88x0/au88x0_core.c
1362
sb << ((0xf - (wtdma & 0xf)) * 2));
sound/pci/au88x0/au88x0_core.c
1363
dma->period_real = dma->period_virt = sb;
sound/pci/cs5530.c
164
dma16, SB_HW_CS5530, &chip->sb);
sound/pci/cs5530.c
170
err = snd_sb16dsp_pcm(chip->sb, 0);
sound/pci/cs5530.c
176
err = snd_sbmixer_new(chip->sb);
sound/pci/cs5530.c
56
struct snd_sb *sb;
sound/usb/qcom/qc_audio_offload.c
1121
ret = xhci_sideband_add_endpoint(uadev[card_num].sb, ep);
sound/usb/qcom/qc_audio_offload.c
1129
sgt = xhci_sideband_get_endpoint_buffer(uadev[card_num].sb, ep);
sound/usb/qcom/qc_audio_offload.c
1158
xhci_sideband_remove_endpoint(uadev[card_num].sb, ep);
sound/usb/qcom/qc_audio_offload.c
1191
ret = xhci_sideband_create_interrupter(uadev[card_num].sb, 1, false,
sound/usb/qcom/qc_audio_offload.c
1198
sgt = xhci_sideband_get_event_buffer(uadev[card_num].sb);
sound/usb/qcom/qc_audio_offload.c
1226
xhci_sideband_remove_interrupter(uadev[card_num].sb);
sound/usb/qcom/qc_audio_offload.c
133
struct xhci_sideband *sb;
sound/usb/qcom/qc_audio_offload.c
1418
resp->interrupter_num = xhci_sideband_interrupter_id(uadev[card_num].sb);
sound/usb/qcom/qc_audio_offload.c
1491
xhci_sideband_remove_interrupter(uadev[card_num].sb);
sound/usb/qcom/qc_audio_offload.c
1498
xhci_sideband_remove_endpoint(uadev[card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1504
xhci_sideband_remove_endpoint(uadev[card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1626
xhci_sideband_stop_endpoint(uadev[pcm_card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1628
xhci_sideband_remove_endpoint(uadev[pcm_card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1639
xhci_sideband_stop_endpoint(uadev[pcm_card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1641
xhci_sideband_remove_endpoint(uadev[pcm_card_num].sb,
sound/usb/qcom/qc_audio_offload.c
1755
struct xhci_sideband *sb;
sound/usb/qcom/qc_audio_offload.c
1773
sb = xhci_sideband_register(intf, XHCI_SIDEBAND_VENDOR,
sound/usb/qcom/qc_audio_offload.c
1775
if (!sb)
sound/usb/qcom/qc_audio_offload.c
1778
sb = uadev[chip->card->number].sb;
sound/usb/qcom/qc_audio_offload.c
1782
uadev[chip->card->number].sb = sb;
sound/usb/qcom/qc_audio_offload.c
1813
xhci_sideband_unregister(sb);
sound/usb/qcom/qc_audio_offload.c
1814
uadev[chip->card->number].sb = NULL;
sound/usb/qcom/qc_audio_offload.c
1875
xhci_sideband_unregister(dev->sb);
sound/usb/qcom/qc_audio_offload.c
701
xhci_sideband_remove_interrupter(uadev[dev->chip->card->number].sb);
sound/usb/qcom/qc_audio_offload.c
785
xhci_sideband_remove_endpoint(uadev[pcm_card_num].sb, ep);
sound/usb/qcom/qc_audio_offload.c
790
xhci_sideband_remove_endpoint(uadev[pcm_card_num].sb, ep);
tools/arch/s390/include/uapi/asm/kvm.h
312
__u64 sb; /* Guest addr of summary bit */
tools/bpf/bpf_dbg.c
946
struct stat sb;
tools/bpf/bpf_dbg.c
955
ret = fstat(pcap_fd, &sb);
tools/bpf/bpf_dbg.c
961
if (!S_ISREG(sb.st_mode)) {
tools/bpf/bpf_dbg.c
966
pcap_map_size = sb.st_size;
tools/bpf/bpftool/cgroup.c
404
static int do_show_tree_fn(const char *fpath, const struct stat *sb,
tools/bpf/bpftool/common.c
551
static int do_build_table_cb(const char *fpath, const struct stat *sb,
tools/hv/hv_vss_daemon.c
118
struct stat sb;
tools/hv/hv_vss_daemon.c
142
if (stat(ent->mnt_fsname, &sb)) {
tools/hv/hv_vss_daemon.c
147
major(sb.st_rdev), minor(sb.st_rdev));
tools/lib/subcmd/parse-options.c
789
int sa = tolower(a->short_name), sb = tolower(b->short_name), ret;
tools/lib/subcmd/parse-options.c
793
if (sb == 0)
tools/lib/subcmd/parse-options.c
794
sb = 'z' + 1;
tools/lib/subcmd/parse-options.c
796
ret = sa - sb;
tools/perf/arch/x86/tests/amd-ibs-period.c
107
struct strbuf sb;
tools/perf/arch/x86/tests/amd-ibs-period.c
110
strbuf_init(&sb, 0);
tools/perf/arch/x86/tests/amd-ibs-period.c
111
strbuf_add(&sb, taskset, strlen(taskset));
tools/perf/arch/x86/tests/amd-ibs-period.c
112
strbuf_add(&sb, perf, strlen(perf));
tools/perf/arch/x86/tests/amd-ibs-period.c
113
strbuf_add(&sb, bench, strlen(bench));
tools/perf/arch/x86/tests/amd-ibs-period.c
114
cmd = strbuf_detach(&sb, NULL);
tools/perf/arch/x86/util/evsel.c
184
struct strbuf sb = STRBUF_INIT;
tools/perf/arch/x86/util/evsel.c
196
evlist__format_evsels(evsel->evlist, &sb, 2048);
tools/perf/arch/x86/util/evsel.c
198
"in '%s'.", sb.buf);
tools/perf/arch/x86/util/evsel.c
199
strbuf_release(&sb);
tools/perf/arch/x86/util/evsel.c
213
evlist__format_evsels(evsel->evlist, &sb, 2048);
tools/perf/arch/x86/util/evsel.c
216
evsel__name(evsel), sb.buf);
tools/perf/arch/x86/util/evsel.c
217
strbuf_release(&sb);
tools/perf/bench/inject-buildid.c
79
static int add_dso(const char *fpath, const struct stat *sb __maybe_unused,
tools/perf/builtin-record.c
2868
struct strbuf sb = STRBUF_INIT;
tools/perf/builtin-record.c
2870
evlist__format_evsels(rec->evlist, &sb, 2048);
tools/perf/builtin-record.c
2873
sb.buf, argv[0], emsg);
tools/perf/builtin-record.c
2874
strbuf_release(&sb);
tools/perf/tests/parse-events.c
27
struct strbuf sb = STRBUF_INIT;
tools/perf/tests/parse-events.c
32
evlist__format_evsels(evlist, &sb, 2048);
tools/perf/tests/parse-events.c
33
pr_debug("FAILED %s:%d: %s\nFor evlist: %s\n", __FILE__, line, test, sb.buf);
tools/perf/tests/parse-events.c
34
strbuf_release(&sb);
tools/perf/util/annotate-data.c
1114
struct strbuf sb;
tools/perf/util/annotate-data.c
1116
strbuf_init(&sb, 32);
tools/perf/util/annotate-data.c
1117
die_get_typename_from_type(&state->regs[reg].type, &sb);
tools/perf/util/annotate-data.c
1118
pr_debug_dtp("(%s)", sb.buf);
tools/perf/util/annotate-data.c
1119
strbuf_release(&sb);
tools/perf/util/annotate-data.c
1152
struct strbuf sb;
tools/perf/util/annotate-data.c
1154
strbuf_init(&sb, 32);
tools/perf/util/annotate-data.c
1155
die_get_typename_from_type(&state->regs[reg].type, &sb);
tools/perf/util/annotate-data.c
1156
pr_debug_dtp("(ptr->%s)", sb.buf);
tools/perf/util/annotate-data.c
1157
strbuf_release(&sb);
tools/perf/util/annotate-data.c
231
struct strbuf sb;
tools/perf/util/annotate-data.c
241
strbuf_init(&sb, 32);
tools/perf/util/annotate-data.c
242
die_get_typename(die, &sb);
tools/perf/util/annotate-data.c
268
member->type_name = strbuf_detach(&sb, NULL);
tools/perf/util/annotate-data.c
361
struct strbuf sb;
tools/perf/util/annotate-data.c
365
strbuf_init(&sb, 32);
tools/perf/util/annotate-data.c
366
if (die_get_typename_from_type(type_die, &sb) < 0)
tools/perf/util/annotate-data.c
367
strbuf_add(&sb, "(unknown type)", 14);
tools/perf/util/annotate-data.c
368
type_name = strbuf_detach(&sb, NULL);
tools/perf/util/annotate-data.c
44
struct strbuf sb;
tools/perf/util/annotate-data.c
79
strbuf_init(&sb, 32);
tools/perf/util/annotate-data.c
80
die_get_typename_from_type(die, &sb);
tools/perf/util/annotate-data.c
81
str = strbuf_detach(&sb, NULL);
tools/perf/util/annotate.c
1815
struct strbuf sb;
tools/perf/util/annotate.c
1820
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/annotate.c
1822
if (header && strbuf_addf(&sb, "# Branch counter abbr list:\n"))
tools/perf/util/annotate.c
1828
if (header && strbuf_addf(&sb, "#"))
tools/perf/util/annotate.c
1831
if (strbuf_addf(&sb, " %s = %s\n", pos->name, pos->abbr_name))
tools/perf/util/annotate.c
1835
if (header && strbuf_addf(&sb, "#"))
tools/perf/util/annotate.c
1837
if (strbuf_addf(&sb, " '-' No event occurs\n"))
tools/perf/util/annotate.c
1840
if (header && strbuf_addf(&sb, "#"))
tools/perf/util/annotate.c
1842
if (strbuf_addf(&sb, " '+' Event occurrences may be lost due to branch counter saturated\n"))
tools/perf/util/annotate.c
1845
*str = strbuf_detach(&sb, NULL);
tools/perf/util/annotate.c
1849
strbuf_release(&sb);
tools/perf/util/annotate.c
1863
struct strbuf sb;
tools/perf/util/annotate.c
1865
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/annotate.c
1881
if (strbuf_addstr(&sb, pos->abbr_name))
tools/perf/util/annotate.c
1885
if (strbuf_addstr(&sb, "=-"))
tools/perf/util/annotate.c
1888
if (strbuf_addf(&sb, "=%d", avg))
tools/perf/util/annotate.c
1892
if (strbuf_addch(&sb, '+'))
tools/perf/util/annotate.c
1895
if (strbuf_addch(&sb, ' '))
tools/perf/util/annotate.c
1899
if ((i < br_cntr_nr - 1) && strbuf_addch(&sb, ','))
tools/perf/util/annotate.c
1904
if (strbuf_addch(&sb, '|'))
tools/perf/util/annotate.c
1908
if (strbuf_addch(&sb, '-'))
tools/perf/util/annotate.c
1926
if (strbuf_addstr(&sb, pos->abbr_name))
tools/perf/util/annotate.c
1931
if (strbuf_addch(&sb, '+'))
tools/perf/util/annotate.c
1939
if (strbuf_addch(&sb, ' '))
tools/perf/util/annotate.c
1944
if (!verbose && strbuf_addch(&sb, br_cntr_nr ? '|' : ' '))
tools/perf/util/annotate.c
1947
*str = strbuf_detach(&sb, NULL);
tools/perf/util/annotate.c
1951
strbuf_release(&sb);
tools/perf/util/bpf-trace-summary.c
119
const struct syscall_data * const *sb = b;
tools/perf/util/bpf-trace-summary.c
121
return (*sa)->total_time > (*sb)->total_time ? -1 : 1;
tools/perf/util/cgroup.c
217
static int add_cgroup_name(const char *fpath, const struct stat *sb __maybe_unused,
tools/perf/util/env.c
396
struct strbuf sb;
tools/perf/util/env.c
407
if (strbuf_init(&sb, 128 * pmu_num) < 0)
tools/perf/util/env.c
411
if (strbuf_addf(&sb, "%u:%s", pmu->type, pmu->name) < 0)
tools/perf/util/env.c
414
if (strbuf_add(&sb, "", 1) < 0)
tools/perf/util/env.c
418
env->pmu_mappings = strbuf_detach(&sb, NULL);
tools/perf/util/env.c
423
strbuf_release(&sb);
tools/perf/util/evlist.c
2522
void evlist__format_evsels(struct evlist *evlist, struct strbuf *sb, size_t max_length)
tools/perf/util/evlist.c
2534
strbuf_addch(sb, '}');
tools/perf/util/evlist.c
2537
strbuf_addch(sb, ',');
tools/perf/util/evlist.c
2539
if (sb->len > max_length) {
tools/perf/util/evlist.c
2540
strbuf_addstr(sb, "...");
tools/perf/util/evlist.c
2544
strbuf_addch(sb, '{');
tools/perf/util/evlist.c
2546
strbuf_addstr(sb, evsel__name(evsel));
tools/perf/util/evlist.c
2551
strbuf_addch(sb, '}');
tools/perf/util/evlist.h
457
void evlist__format_evsels(struct evlist *evlist, struct strbuf *sb, size_t max_length);
tools/perf/util/header.c
2792
struct strbuf sb;
tools/perf/util/header.c
2806
if (strbuf_init(&sb, 128) < 0)
tools/perf/util/header.c
2815
if (strbuf_add(&sb, str, strlen(str) + 1) < 0)
tools/perf/util/header.c
2820
env->sibling_cores = strbuf_detach(&sb, NULL);
tools/perf/util/header.c
2834
if (strbuf_add(&sb, str, strlen(str) + 1) < 0)
tools/perf/util/header.c
2839
env->sibling_threads = strbuf_detach(&sb, NULL);
tools/perf/util/header.c
2883
if (strbuf_add(&sb, str, strlen(str) + 1) < 0)
tools/perf/util/header.c
2888
env->sibling_dies = strbuf_detach(&sb, NULL);
tools/perf/util/header.c
2900
strbuf_release(&sb);
tools/perf/util/header.c
2959
struct strbuf sb;
tools/perf/util/header.c
2970
if (strbuf_init(&sb, 128) < 0)
tools/perf/util/header.c
2981
if (strbuf_addf(&sb, "%u:%s", type, name) < 0)
tools/perf/util/header.c
2984
if (strbuf_add(&sb, "", 1) < 0)
tools/perf/util/header.c
2995
env->pmu_mappings = strbuf_detach(&sb, NULL);
tools/perf/util/header.c
2999
strbuf_release(&sb);
tools/perf/util/hwmon_pmu.c
152
const char * const *sb = b;
tools/perf/util/hwmon_pmu.c
154
return strcmp(sa, *sb);
tools/perf/util/metricgroup.c
451
static int encode_metric_id(struct strbuf *sb, const char *x)
tools/perf/util/metricgroup.c
458
ret = strbuf_addch(sb, '!');
tools/perf/util/metricgroup.c
462
ret = strbuf_addch(sb, '0' + (c - code_characters));
tools/perf/util/metricgroup.c
466
ret = strbuf_addch(sb, *x);
tools/perf/util/metricgroup.c
474
static int decode_metric_id(struct strbuf *sb, const char *x)
tools/perf/util/metricgroup.c
492
ret = strbuf_addch(sb, c);
tools/perf/util/metricgroup.c
502
struct strbuf sb = STRBUF_INIT;
tools/perf/util/metricgroup.c
510
ret = strbuf_setlen(&sb, 0);
tools/perf/util/metricgroup.c
514
ret = decode_metric_id(&sb, ev->metric_id);
tools/perf/util/metricgroup.c
519
ev->metric_id = strdup(sb.buf);
tools/perf/util/metricgroup.c
532
for (cur = strchr(sb.buf, '@') ; cur; cur = strchr(++cur, '@')) {
tools/perf/util/metricgroup.c
538
if (!has_slash && !strchr(sb.buf, ':')) {
tools/perf/util/metricgroup.c
539
ret = strbuf_addch(&sb, ':');
tools/perf/util/metricgroup.c
543
ret = strbuf_addstr(&sb, modifier);
tools/perf/util/metricgroup.c
547
ev->name = strdup(sb.buf);
tools/perf/util/metricgroup.c
554
strbuf_release(&sb);
tools/perf/util/parse-events.c
1504
struct strbuf sb;
tools/perf/util/parse-events.c
1506
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/parse-events.c
1509
strbuf_addf(&sb, "%s//", pmu->name);
tools/perf/util/parse-events.c
1511
strbuf_addf(&sb, "%s/", pmu->name);
tools/perf/util/parse-events.c
1512
parse_events_terms__to_strbuf(const_parsed_terms, &sb);
tools/perf/util/parse-events.c
1513
strbuf_addch(&sb, '/');
tools/perf/util/parse-events.c
1515
fprintf(stderr, "Attempt to add: %s\n", sb.buf);
tools/perf/util/parse-events.c
1516
strbuf_release(&sb);
tools/perf/util/parse-events.c
1558
struct strbuf sb;
tools/perf/util/parse-events.c
1560
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/parse-events.c
1561
parse_events_terms__to_strbuf(&parsed_terms, &sb);
tools/perf/util/parse-events.c
1562
fprintf(stderr, "..after resolving event: %s/%s/\n", pmu->name, sb.buf);
tools/perf/util/parse-events.c
1563
strbuf_release(&sb);
tools/perf/util/parse-events.c
1675
struct strbuf sb;
tools/perf/util/parse-events.c
1677
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/parse-events.c
1678
parse_events_terms__to_strbuf(&parsed_terms, &sb);
tools/perf/util/parse-events.c
1679
pr_debug("%s -> %s/%s/\n", event_name, pmu->name, sb.buf);
tools/perf/util/parse-events.c
1680
strbuf_release(&sb);
tools/perf/util/parse-events.c
1690
struct strbuf sb;
tools/perf/util/parse-events.c
1692
strbuf_init(&sb, /*hint=*/ 0);
tools/perf/util/parse-events.c
1693
parse_events_terms__to_strbuf(&parsed_terms, &sb);
tools/perf/util/parse-events.c
1694
pr_debug("%s -> fake/%s/\n", event_name, sb.buf);
tools/perf/util/parse-events.c
1695
strbuf_release(&sb);
tools/perf/util/parse-events.c
2301
struct strbuf sb = STRBUF_INIT;
tools/perf/util/parse-events.c
2303
evlist__format_evsels(evlist, &sb, 2048);
tools/perf/util/parse-events.c
2304
pr_debug("evlist after sorting/fixing: '%s'\n", sb.buf);
tools/perf/util/parse-events.c
2305
strbuf_release(&sb);
tools/perf/util/parse-events.c
2830
static int parse_events_terms__to_strbuf(const struct parse_events_terms *terms, struct strbuf *sb)
tools/perf/util/parse-events.c
2842
ret = strbuf_addch(sb, ',');
tools/perf/util/parse-events.c
2851
ret = strbuf_addf(sb, "%s", term->config);
tools/perf/util/parse-events.c
2853
ret = strbuf_addf(sb, "%s=%#"PRIx64, term->config, term->val.num);
tools/perf/util/parse-events.c
2856
ret = strbuf_addf(sb, "%s=", term->config);
tools/perf/util/parse-events.c
2860
ret = strbuf_addf(sb, "%s=",
tools/perf/util/parse-events.c
2866
ret = strbuf_addf(sb, "%s", term->val.str);
tools/perf/util/parse-events.c
43
static int parse_events_terms__to_strbuf(const struct parse_events_terms *terms, struct strbuf *sb);
tools/perf/util/python.c
1811
struct strbuf sb = STRBUF_INIT;
tools/perf/util/python.c
1815
strbuf_addstr(&sb, "evlist([");
tools/perf/util/python.c
1818
strbuf_addch(&sb, ',');
tools/perf/util/python.c
1820
strbuf_addstr(&sb, evsel__name(pos));
tools/perf/util/python.c
1822
strbuf_addf(&sb, "%s/%s/", pos->pmu->name, evsel__name(pos));
tools/perf/util/python.c
1825
strbuf_addstr(&sb, "])");
tools/perf/util/python.c
1826
result = PyUnicode_FromString(sb.buf);
tools/perf/util/python.c
1827
strbuf_release(&sb);
tools/perf/util/sort.c
4295
static void add_key(struct strbuf *sb, const char *str, int *llen)
tools/perf/util/sort.c
4301
strbuf_addstr(sb, "\n\t\t\t ");
tools/perf/util/sort.c
4304
strbuf_addf(sb, " %s", str);
tools/perf/util/sort.c
4308
static void add_sort_string(struct strbuf *sb, struct sort_dimension *s, int n,
tools/perf/util/sort.c
4314
add_key(sb, s[i].name, llen);
tools/perf/util/sort.c
4317
static void add_hpp_sort_string(struct strbuf *sb, struct hpp_dimension *s, int n,
tools/perf/util/sort.c
4323
add_key(sb, s[i].name, llen);
tools/perf/util/sort.c
4328
struct strbuf sb;
tools/perf/util/sort.c
4332
strbuf_init(&sb, 300);
tools/perf/util/sort.c
4333
strbuf_addstr(&sb, prefix);
tools/perf/util/sort.c
4334
add_hpp_sort_string(&sb, hpp_sort_dimensions,
tools/perf/util/sort.c
4336
add_sort_string(&sb, common_sort_dimensions,
tools/perf/util/sort.c
4339
add_sort_string(&sb, bstack_sort_dimensions,
tools/perf/util/sort.c
4342
add_sort_string(&sb, memory_sort_dimensions,
tools/perf/util/sort.c
4344
s = strbuf_detach(&sb, NULL);
tools/perf/util/sort.c
4345
strbuf_release(&sb);
tools/perf/util/strbuf.c
100
ret = strbuf_grow(sb, 64);
tools/perf/util/strbuf.c
106
len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap);
tools/perf/util/strbuf.c
111
if (len > strbuf_avail(sb)) {
tools/perf/util/strbuf.c
112
ret = strbuf_grow(sb, len);
tools/perf/util/strbuf.c
117
len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap_saved);
tools/perf/util/strbuf.c
118
if (len > strbuf_avail(sb)) {
tools/perf/util/strbuf.c
125
return strbuf_setlen(sb, sb->len + len);
tools/perf/util/strbuf.c
128
int strbuf_addf(struct strbuf *sb, const char *fmt, ...)
tools/perf/util/strbuf.c
134
ret = strbuf_addv(sb, fmt, ap);
tools/perf/util/strbuf.c
139
ssize_t strbuf_read(struct strbuf *sb, int fd, ssize_t hint)
tools/perf/util/strbuf.c
141
size_t oldlen = sb->len;
tools/perf/util/strbuf.c
142
size_t oldalloc = sb->alloc;
tools/perf/util/strbuf.c
145
ret = strbuf_grow(sb, hint ? hint : 8192);
tools/perf/util/strbuf.c
152
cnt = read(fd, sb->buf + sb->len, sb->alloc - sb->len - 1);
tools/perf/util/strbuf.c
155
strbuf_release(sb);
tools/perf/util/strbuf.c
157
strbuf_setlen(sb, oldlen);
tools/perf/util/strbuf.c
162
sb->len += cnt;
tools/perf/util/strbuf.c
163
ret = strbuf_grow(sb, 8192);
tools/perf/util/strbuf.c
168
sb->buf[sb->len] = '\0';
tools/perf/util/strbuf.c
169
return sb->len - oldlen;
tools/perf/util/strbuf.c
20
int strbuf_init(struct strbuf *sb, ssize_t hint)
tools/perf/util/strbuf.c
22
sb->alloc = sb->len = 0;
tools/perf/util/strbuf.c
23
sb->buf = strbuf_slopbuf;
tools/perf/util/strbuf.c
25
return strbuf_grow(sb, hint);
tools/perf/util/strbuf.c
29
void strbuf_release(struct strbuf *sb)
tools/perf/util/strbuf.c
31
if (sb->alloc) {
tools/perf/util/strbuf.c
32
zfree(&sb->buf);
tools/perf/util/strbuf.c
33
strbuf_init(sb, 0);
tools/perf/util/strbuf.c
37
char *strbuf_detach(struct strbuf *sb, size_t *sz)
tools/perf/util/strbuf.c
39
char *res = sb->alloc ? sb->buf : NULL;
tools/perf/util/strbuf.c
41
*sz = sb->len;
tools/perf/util/strbuf.c
42
strbuf_init(sb, 0);
tools/perf/util/strbuf.c
46
int strbuf_grow(struct strbuf *sb, size_t extra)
tools/perf/util/strbuf.c
49
size_t nr = sb->len + extra + 1;
tools/perf/util/strbuf.c
51
if (nr < sb->alloc)
tools/perf/util/strbuf.c
54
if (nr <= sb->len)
tools/perf/util/strbuf.c
57
if (alloc_nr(sb->alloc) > nr)
tools/perf/util/strbuf.c
58
nr = alloc_nr(sb->alloc);
tools/perf/util/strbuf.c
64
buf = realloc(sb->alloc ? sb->buf : NULL, nr * sizeof(*buf));
tools/perf/util/strbuf.c
68
sb->buf = buf;
tools/perf/util/strbuf.c
69
sb->alloc = nr;
tools/perf/util/strbuf.c
73
int strbuf_addch(struct strbuf *sb, int c)
tools/perf/util/strbuf.c
75
int ret = strbuf_grow(sb, 1);
tools/perf/util/strbuf.c
79
sb->buf[sb->len++] = c;
tools/perf/util/strbuf.c
80
sb->buf[sb->len] = '\0';
tools/perf/util/strbuf.c
84
int strbuf_add(struct strbuf *sb, const void *data, size_t len)
tools/perf/util/strbuf.c
86
int ret = strbuf_grow(sb, len);
tools/perf/util/strbuf.c
90
memcpy(sb->buf + sb->len, data, len);
tools/perf/util/strbuf.c
91
return strbuf_setlen(sb, sb->len + len);
tools/perf/util/strbuf.c
94
static int strbuf_addv(struct strbuf *sb, const char *fmt, va_list ap)
tools/perf/util/strbuf.c
99
if (!strbuf_avail(sb)) {
tools/perf/util/strbuf.h
64
static inline ssize_t strbuf_avail(const struct strbuf *sb) {
tools/perf/util/strbuf.h
65
return sb->alloc ? sb->alloc - sb->len - 1 : 0;
tools/perf/util/strbuf.h
70
static inline int strbuf_setlen(struct strbuf *sb, size_t len) {
tools/perf/util/strbuf.h
71
if (!sb->alloc) {
tools/perf/util/strbuf.h
72
int ret = strbuf_grow(sb, 0);
tools/perf/util/strbuf.h
76
assert(len < sb->alloc);
tools/perf/util/strbuf.h
77
sb->len = len;
tools/perf/util/strbuf.h
78
sb->buf[len] = '\0';
tools/perf/util/strbuf.h
83
int strbuf_addch(struct strbuf *sb, int c);
tools/perf/util/strbuf.h
86
static inline int strbuf_addstr(struct strbuf *sb, const char *s) {
tools/perf/util/strbuf.h
87
return strbuf_add(sb, s, strlen(s));
tools/perf/util/strbuf.h
90
int strbuf_addf(struct strbuf *sb, const char *fmt, ...) __printf(2, 3);
tools/power/x86/turbostat/turbostat.c
6936
struct stat sb;
tools/power/x86/turbostat/turbostat.c
6940
return !stat(pathname, &sb);
tools/power/x86/turbostat/turbostat.c
6945
struct stat sb;
tools/power/x86/turbostat/turbostat.c
6949
return !stat(pathname, &sb);
tools/power/x86/x86_energy_perf_policy/x86_energy_perf_policy.c
1436
struct stat sb;
tools/power/x86/x86_energy_perf_policy/x86_energy_perf_policy.c
1440
if (stat(test_path, &sb) == 0)
tools/power/x86/x86_energy_perf_policy/x86_energy_perf_policy.c
1446
struct stat sb;
tools/power/x86/x86_energy_perf_policy/x86_energy_perf_policy.c
1452
if (stat(pathname, &sb)) {
tools/spi/spidev_test.c
363
struct stat sb;
tools/spi/spidev_test.c
368
if (stat(filename, &sb) == -1)
tools/spi/spidev_test.c
375
tx = malloc(sb.st_size);
tools/spi/spidev_test.c
379
rx = malloc(sb.st_size);
tools/spi/spidev_test.c
383
bytes = read(tx_fd, tx, sb.st_size);
tools/spi/spidev_test.c
384
if (bytes != sb.st_size)
tools/spi/spidev_test.c
387
transfer(fd, tx, rx, sb.st_size);
tools/testing/selftests/alsa/conf.c
143
struct stat sb;
tools/testing/selftests/alsa/conf.c
151
if (lstat(path, &sb) != 0)
tools/testing/selftests/alsa/conf.c
153
if (S_ISLNK(sb.st_mode)) {
tools/testing/selftests/alsa/conf.c
166
if (S_ISDIR(sb.st_mode))
tools/testing/selftests/alsa/conf.c
168
if ((sb.st_mode & S_IRUSR) == 0)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1239
DEFINE_STRBUF(sb, 512);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1244
snappendf(sb, "MISMATCH %s.%s: ", ctx1, ctx2);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1245
snprintf_range(t, sb, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1246
snappendf(sb, " != ");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1247
snprintf_range(t, sb, y);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1249
printf("%s\n", sb->buf);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1426
static void subtest_case_str(struct strbuf *sb, struct subtest_case *t, bool use_op)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1428
snappendf(sb, "(%s)", t_str(t->init_t));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1429
snprintf_range(t->init_t, sb, t->x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1430
snappendf(sb, " (%s)%s ", t_str(t->cond_t), use_op ? op_str(t->op) : "<op>");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1431
snprintf_range(t->init_t, sb, t->y);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1508
DEFINE_STRBUF(sb, 256);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1517
sb->pos = 0; /* reset position in strbuf */
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1518
subtest_case_str(sb, &sub, false /* ignore op */);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1519
if (is_subtest && !test__start_subtest(sb->buf))
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1523
sb->pos = 0; /* reset position in strbuf */
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1524
subtest_case_str(sb, &sub, true /* print op */);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1527
printf("TEST CASE: %s\n", sb->buf);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
1531
ASSERT_OK(err, sb->buf);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
159
static void snprintf_num(enum num_t t, struct strbuf *sb, u64 x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
165
case U64: return snappendf(sb, "%llu", (u64)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
166
case U32: return snappendf(sb, "%u", (u32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
167
case S64: return snappendf(sb, "%lld", (s64)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
168
case S32: return snappendf(sb, "%d", (s32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
175
return snappendf(sb, "U64_MAX");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
177
return snappendf(sb, "U64_MAX-%llu", U64_MAX - x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
179
return snappendf(sb, "%#llx", (u64)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
182
return snappendf(sb, "U32_MAX");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
184
return snappendf(sb, "U32_MAX-%u", U32_MAX - (u32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
186
return snappendf(sb, "%#x", (u32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
189
return snappendf(sb, "S64_MAX");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
191
return snappendf(sb, "S64_MAX-%lld", S64_MAX - (s64)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
193
return snappendf(sb, "S64_MIN");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
195
return snappendf(sb, "S64_MIN+%lld", (s64)x - S64_MIN);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
197
return snappendf(sb, "%#llx", (s64)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
200
return snappendf(sb, "S32_MAX");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
202
return snappendf(sb, "S32_MAX-%d", S32_MAX - (s32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
204
return snappendf(sb, "S32_MIN");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
206
return snappendf(sb, "S32_MIN+%d", (s32)x - S32_MIN);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
208
return snappendf(sb, "%#x", (s32)x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
222
static void snprintf_range(enum num_t t, struct strbuf *sb, struct range x)
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
225
return snprintf_num(t, sb, x.a);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
227
snappendf(sb, "[");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
228
snprintf_num(t, sb, x.a);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
229
snappendf(sb, "; ");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
230
snprintf_num(t, sb, x.b);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
231
snappendf(sb, "]");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
236
DEFINE_STRBUF(sb, 128);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
238
snprintf_range(t, sb, x);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
239
printf("%s%s", sb->buf, sfx);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
700
DEFINE_STRBUF(sb, 512);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
709
snappendf(sb, "scalar(");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
711
snappendf(sb, "%s%s=", cnt++ ? "," : "", t_str(t));
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
712
snprintf_range(t, sb, r->r[t]);
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
714
snappendf(sb, ")");
tools/testing/selftests/bpf/prog_tests/reg_bounds.c
716
printf("%s%s", sb->buf, sfx);
tools/testing/selftests/kvm/guest_memfd_test.c
214
struct stat sb;
tools/testing/selftests/kvm/guest_memfd_test.c
217
ret = fstat(fd, &sb);
tools/testing/selftests/kvm/guest_memfd_test.c
219
TEST_ASSERT_EQ(sb.st_size, total_size);
tools/testing/selftests/kvm/guest_memfd_test.c
220
TEST_ASSERT_EQ(sb.st_blksize, page_size);
tools/testing/selftests/net/mptcp/mptcp_connect.c
836
struct stat sb;
tools/testing/selftests/net/mptcp/mptcp_connect.c
840
err = fstat(fd, &sb);
tools/testing/selftests/net/mptcp/mptcp_connect.c
846
if ((sb.st_mode & S_IFMT) != S_IFREG) {
tools/testing/selftests/net/mptcp/mptcp_connect.c
851
count = sb.st_size;
tools/testing/selftests/resctrl/fill_buf.c
51
sb();
tools/testing/selftests/ring-buffer/map_test.c
270
struct stat sb;
tools/testing/selftests/ring-buffer/map_test.c
276
if (stat(TRACEFS_ROOT"/snapshot", &sb))
tools/testing/selftests/sgx/load.c
184
struct stat sb;
tools/testing/selftests/sgx/load.c
198
ret = stat(device_path, &sb);
tools/testing/selftests/sgx/load.c
46
struct stat sb;
tools/testing/selftests/sgx/load.c
57
ret = stat(path, &sb);
tools/testing/selftests/sgx/load.c
63
bin = mmap(NULL, sb.st_size, PROT_READ, MAP_PRIVATE, fd, 0);
tools/testing/selftests/sgx/load.c
70
encl->bin_size = sb.st_size;
tools/usb/testusb.c
239
static int find_testdev(const char *name, const struct stat *sb, int flag)
tools/usb/testusb.c
245
(void)sb; /* unused */
virt/kvm/guest_memfd.c
937
static struct inode *kvm_gmem_alloc_inode(struct super_block *sb)
virt/kvm/guest_memfd.c
941
gi = alloc_inode_sb(sb, kvm_gmem_inode_cachep, GFP_KERNEL);