BTRFS_SUPER_INFO_SIZE
BTRFS_SUPER_INFO_SIZE - BTRFS_CSUM_SIZE, result);
if (bytenr + BTRFS_SUPER_INFO_SIZE >=
BTRFS_SUPER_INFO_SIZE - BTRFS_CSUM_SIZE, sb->csum);
memcpy(disk_super, sb, BTRFS_SUPER_INFO_SIZE);
bio_add_folio_nofail(bio, folio, BTRFS_SUPER_INFO_SIZE, offset);
if (bytenr + BTRFS_SUPER_INFO_SIZE >=
u64 sb_end = sb_start + BTRFS_SUPER_INFO_SIZE;
!in_range(start, sb_start, BTRFS_SUPER_INFO_SIZE))
static_assert(sizeof(struct btrfs_super_block) == BTRFS_SUPER_INFO_SIZE);
BTRFS_SUPER_INFO_SIZE, REQ_OP_READ);
if (bytenr + BTRFS_SUPER_INFO_SIZE >
fs_info->super_copy = kzalloc(BTRFS_SUPER_INFO_SIZE, GFP_KERNEL);
fs_info->super_for_commit = kzalloc(BTRFS_SUPER_INFO_SIZE, GFP_KERNEL);
if (bytenr + BTRFS_SUPER_INFO_SIZE >= bdev_nr_bytes(bdev))
(bytenr + BTRFS_SUPER_INFO_SIZE) >> PAGE_SHIFT);
ASSERT(BTRFS_SUPER_INFO_SIZE <= fs_info->nodesize);
write_extent_buffer(sb, super_copy, 0, BTRFS_SUPER_INFO_SIZE);
sb_pos + BTRFS_SUPER_INFO_SIZE <= pos)) {
pos = ALIGN(sb_pos + BTRFS_SUPER_INFO_SIZE,
u64 bytenr = ALIGN_DOWN(zone_end, BTRFS_SUPER_INFO_SIZE) -
BTRFS_SUPER_INFO_SIZE;
#define SUPER_INFO_SECTORS ((u64)BTRFS_SUPER_INFO_SIZE >> SECTOR_SHIFT)
BTRFS_SUPER_INFO_SIZE);
wp -= BTRFS_SUPER_INFO_SIZE;