BTRFS_STRIPE_LEN
io_stripe_size = BTRFS_STRIPE_LEN;
const unsigned int stripe_npages = BTRFS_STRIPE_LEN >> PAGE_SHIFT;
BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits;
max(BTRFS_STRIPE_LEN >> PAGE_SHIFT, 1),
rbio->nr_data * BTRFS_STRIPE_LEN,
ASSERT(IS_ALIGNED(offset_in_full_stripe, BTRFS_STRIPE_LEN));
cur_off < offset_in_full_stripe + BTRFS_STRIPE_LEN;
BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits);
if (size != rbio->nr_data * BTRFS_STRIPE_LEN)
BUG_ON(size > rbio->nr_data * BTRFS_STRIPE_LEN);
BTRFS_STRIPE_LEN, false);
#define SCRUB_STRIPE_MAX_FOLIOS (BTRFS_STRIPE_LEN / PAGE_SIZE)
cur_logical < min(stripe->logical + BTRFS_STRIPE_LEN,
stripe->logical = round_down(cur_logical - bg->start, BTRFS_STRIPE_LEN) +
stripe_end = stripe->logical + BTRFS_STRIPE_LEN - 1;
(BTRFS_STRIPE_LEN / BTRFS_MIN_BLOCKSIZE))];
ASSERT(BITS_PER_LONG >= BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits);
return min(BTRFS_STRIPE_LEN,
u64 stripe_len = BTRFS_STRIPE_LEN;
bbio = alloc_scrub_bbio(fs_info, BTRFS_STRIPE_LEN >> min_folio_shift, REQ_OP_READ,
BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits);
BTRFS_STRIPE_LEN, stripe);
cur_logical = found_logical + BTRFS_STRIPE_LEN;
BTRFS_STRIPE_LEN, device, cur_physical,
cur_physical += BTRFS_STRIPE_LEN;
sctx->stat.last_physical = min(physical + BTRFS_STRIPE_LEN,
ret = scrub_simple_mirror(sctx, bg, logical, BTRFS_STRIPE_LEN,
physical += BTRFS_STRIPE_LEN;
ASSERT(fs_info->nodesize <= BTRFS_STRIPE_LEN);
stripe->nr_sectors = BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits;
ASSERT(BTRFS_STRIPE_LEN >> min_folio_shift <= SCRUB_STRIPE_MAX_FOLIOS);
ret = btrfs_alloc_folio_array(BTRFS_STRIPE_LEN >> min_folio_shift,
stripe->csums = kcalloc(BTRFS_STRIPE_LEN >> fs_info->sectorsize_bits,
min_stripe_size = rattr->dev_stripes * BTRFS_STRIPE_LEN;
avail_space = rounddown(avail_space, BTRFS_STRIPE_LEN);
if (out_stripe_len != BTRFS_STRIPE_LEN) {
if (unlikely(!is_power_of_2(stripe_len) || stripe_len != BTRFS_STRIPE_LEN)) {
ctl->stripe_size = round_down(ctl->stripe_size, BTRFS_STRIPE_LEN);
map->io_align = BTRFS_STRIPE_LEN;
map->io_width = BTRFS_STRIPE_LEN;
btrfs_set_stack_chunk_stripe_len(chunk, BTRFS_STRIPE_LEN);
btrfs_set_stack_chunk_io_align(chunk, BTRFS_STRIPE_LEN);
btrfs_set_stack_chunk_io_width(chunk, BTRFS_STRIPE_LEN);
stripe_nr_end = round_up(offset + length, BTRFS_STRIPE_LEN) >>
stripes[i].length += BTRFS_STRIPE_LEN;
return BTRFS_STRIPE_LEN - io_geom->stripe_offset;
logical < stripe_start + BTRFS_STRIPE_LEN)
#define BTRFS_STRIPE_LEN_MASK (BTRFS_STRIPE_LEN - 1)
static_assert(ilog2(BTRFS_STRIPE_LEN) == BTRFS_STRIPE_LEN_SHIFT);
stripe_offset = BTRFS_STRIPE_LEN;
zone_info[i].alloc_offset += BTRFS_STRIPE_LEN;
zone_info[map->num_stripes - 1].alloc_offset > BTRFS_STRIPE_LEN)) {
stripe_offset = BTRFS_STRIPE_LEN;
raid0_allocs[idx] += BTRFS_STRIPE_LEN;
zone_info[map->num_stripes - 1].alloc_offset > BTRFS_STRIPE_LEN)) {
if (!IS_ALIGNED(zone_size, BTRFS_STRIPE_LEN)) {
zone_size, BTRFS_STRIPE_LEN);