block_end
phys_addr_t block_start, block_end, memblock_limit = 0;
for_each_mem_range(i, &block_start, &block_end) {
for_each_mem_range(i, &block_start, &block_end) {
if (block_end > lowmem_limit)
block_end);
else if (!IS_ALIGNED(block_end, PMD_SIZE))
phys_addr_t block_start, block_end;
for_each_mem_range(i, &block_start, &block_end) {
if (block_end > memblock_limit)
memblock_limit = block_end;
block_end = block_start + drm_buddy_block_size(mm, block);
if (range_overflows(new_start, new_size, block_end))
u64 block_start, block_end;
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
if (!contains(start, end, block_start, block_end))
u64 block_end;
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
if (!overlaps(start, end, block_start, block_end))
if (block_start < start || block_end > end) {
u64 adjusted_end = min(block_end, end);
if (contains(start, end, block_start, block_end) &&
u64 block_end;
block_end = block_start + drm_buddy_block_size(mm, block) - 1;
if (!overlaps(start, end, block_start, block_end))
if (contains(start, end, block_start, block_end)) {
u64 block_start, block_end;
if (lhs->block_end <= rhs->block_begin)
if (lhs->block_begin >= rhs->block_end)
if (WARN_ON_ONCE(key->block_end - key->block_begin > BIO_PRISON_MAX_RANGE))
(key->block_end - 1) >> BIO_PRISON_MAX_RANGE_SHIFT))
dm_block_t block_begin, block_end;
if (lhs->block_end <= rhs->block_begin)
if (lhs->block_begin >= rhs->block_end)
dm_block_t block_begin, block_end;
key->block_end = from_oblock(end);
r = dm_thin_remove_range(tc->td, m->cell->key.block_begin, m->cell->key.block_end);
key->block_end = e;
m->virt_end = virt_cell->key.block_end;
break_up_discard_bio(tc, virt_cell->key.block_begin, virt_cell->key.block_end, bio);
u32 block_end;
block_end = imginfo[21];
block_start, block_end);
int block_start, int block_end)
for (i = block_start; i <= block_end; i++)
int block_start, int block_end);
const u64 block_end = block_start + extent_map_block_len(em);
if (block_end < block_start)
return block_end;
u64 block_end;
block_end = block_start + blocksize - 1;
btrfs_lock_extent(io_tree, block_start, block_end, &cached_state);
btrfs_unlock_extent(io_tree, block_start, block_end, &cached_state);
btrfs_clear_extent_bit(&inode->io_tree, block_start, block_end,
ret = btrfs_set_extent_delalloc(inode, block_start, block_end, 0,
btrfs_unlock_extent(io_tree, block_start, block_end, &cached_state);
zero_end = min_t(u64, block_end, end);
block_end + 1 - block_start);
block_end + 1 - block_start);
btrfs_set_extent_bit(&inode->io_tree, block_start, block_end,
btrfs_unlock_extent(io_tree, block_start, block_end, &cached_state);
u64 block_end = ALIGN(size, fs_info->sectorsize);
btrfs_lock_and_flush_ordered_range(inode, hole_start, block_end - 1,
em = btrfs_get_extent(inode, NULL, cur_offset, block_end - cur_offset);
last_byte = min(btrfs_extent_map_end(em), block_end);
if (cur_offset >= block_end)
btrfs_unlock_extent(io_tree, hole_start, block_end - 1, &cached_state);
size_t block_start, block_end;
block_end = block_start + bh->b_size;
if (block_end > from && block_start < to) {
xend = min(to, block_end);
block_start = block_end;
size_t block_start, block_end;
block++, block_start=block_end, bh = bh->b_this_page) {
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
if (block_end > to || block_start < from)
to, block_end,
(block_start < from || block_end > to)) {
size_t block_start, block_end;
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
block_start = block_end;
unsigned block_start, block_end, blocksize;
block_end = block_start + blocksize;
if (block_end > from && block_start < to) {
if (block_end >= to)
block_start = block_end;
unsigned block_start, block_end;
block_start = block_end, bh = next) {
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
unsigned block_start, block_end;
block++, block_start = block_end, bh = bh->b_this_page) {
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
if (block_end > to || block_start < from)
block_end,
(block_start < from || block_end > to)) {
unsigned int block_start = 0, block_end;
block_end = block_start + bh->b_size;
if (block_end > from && block_start < to) {
size = min(to, block_end) - start;
block_start = block_end;
unsigned int blocksize, block_start, block_end;
block_end = 0;
block_start = block_end;
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
size_t block_start, block_end;
block_end = 0;
block_start = block_end;
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to)
sector_t blocknr, block_end;
ops->ent_blocknr(sb, ent_limit - 1, &offset, &block_end);
ra->limit = (block_end + 1) - blocknr;
loff_t block_end = round_up(pos + len, block_size);
block_end - block_start, &poff, &plen);
} while ((block_start += plen) < block_end);
loff_t block_start, block_end;
block_end = le32_to_cpu(*(__le32 *)
if (block_start > block_end) {
ret = zisofs_uncompress_block(inode, block_start, block_end,
block_start = block_end;
loff_t block_end, int pcount,
int i, block_size = block_end - block_start;
block_end(bh), depth);
free_branches(inode, partial->p + 1, block_end(partial->bh),
unsigned int block_start, block_end;
block_start = block_end, bh = bh->b_this_page) {
block_end = block_start + bh->b_size;
if (block_end > from && block_start < to && !buffer_dirty(bh))
u8 * const block_end = out_next + block_size;
if (match_len > (size_t)(block_end - out_next))
block_end, LZX_MIN_MATCH_LEN);
} while (out_next != block_end);
unsigned block_start, block_end;
block_start = block_end, bh = next)
block_end = block_start + blocksize;
if (block_end <= from || block_start >= to) {
unsigned int block_end, block_start;
block_end = block_start + bsize;
if (block_start >= to || block_end <= from) {
(block_start < from || block_end > to)) {
block_end = block_start + bsize;
if (block_end <= from)
block_start = block_end;
unsigned int block_start, block_end;
block_end = block_start + bh->b_size;
if (block_end > from && block_start < to) {
end = min(to, block_end);
block_start = block_end;
unsigned zero_from, zero_to, block_start, block_end;
block_start = block_end) {
block_end = block_start + i_blocksize(inode);
const sector_t block_end = read_end >> msblk->devblksize_log2;
int total_len = (block_end - block) << msblk->devblksize_log2;