range_len
range_len(&kvmppc_uvmem_pgmap.range));
(range_len(&kvmppc_uvmem_pgmap.range) >> PAGE_SHIFT);
int chan, range_len, i;
range_len = s->range_table->length;
range_len = s->range_table_list[chan]->length;
range_len = 0;
CR_RANGE(chanspec) >= range_len) {
i, chanspec, chan, range_len);
resource_size_t size = range_len(hpa);
len = range_len(&ctx->hpa_range);
spa_len = range_len(&hpa_range);
part->range.start, range_len(&part->range),
size = range_len(&cxld->hpa_range);
len = range_len(&info->dvsec_range[which]);
return sysfs_emit(buf, "%#llx\n", range_len(&cxld->hpa_range));
*res = DEFINE_RES_MEM_NAMED(hpa_range->start, range_len(hpa_range),
to_alloc = range_len(&r);
size_t len = range_len(&dev_dax->ranges[i].range);
if (dev_dax->nr_range && range_len(&dev_dax->ranges[0].range)) {
size += range_len(&dev_dax->ranges[i].range);
__release_region(&dax_region->res, range->start, range_len(range));
|| !IS_ALIGNED(range_len(range), align))
pgoff += PHYS_PFN(range_len(&ranges[i].range));
shrink = min_t(u64, to_shrink, range_len(range));
if (shrink >= range_len(range)) {
return adjust_dev_dax_range(dev_dax, adjust, range_len(range)
.size = range_len(&cxlr_dax->hpa_range),
range_len(range), dev_name(dev))) {
pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1;
.size = region_idle ? 0 : range_len(&mri->range),
orig_len += range_len(&dev_dax->ranges[i].range);
total_len += range_len(&range);
res = request_mem_region(range.start, range_len(&range), data->res_name);
range_len(&range), kmem_name, mhp_flags);
rc = remove_memory(range.start, range_len(&range));
.size = range_len(&range),
release_mem_region(chunk->pagemap.range.start, range_len(&chunk->pagemap.range));
unsigned long i, npages = range_len(&chunk->pagemap.range) >> PAGE_SHIFT;
range_len(&chunk->pagemap.range));
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
return _iwl_dump_ini_paging_iter(fwrt, range_ptr, range_len, idx);
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range_ptr, u32 range_len, int idx)
void *range, u32 range_len, int idx);
u32 range_len, void *ptr)
for (i = 0; i < range_len; i++) {
u32 range_len, void *ptr)
for (i = 0; i < range_len; i++) {
u32 range_len;
range_len = ARRAY_SIZE(iwl_prph_dump_addr_ax210);
handler(fwrt, iwl_prph_dump_addr_ax210, range_len, ptr);
range_len = ARRAY_SIZE(iwl_prph_dump_addr_22000);
handler(fwrt, iwl_prph_dump_addr_22000, range_len, ptr);
range_len = ARRAY_SIZE(iwl_prph_dump_addr_comm);
handler(fwrt, iwl_prph_dump_addr_comm, range_len, ptr);
range_len = ARRAY_SIZE(iwl_prph_dump_addr_9000);
handler(fwrt, iwl_prph_dump_addr_9000, range_len, ptr);
len = range->start + range_len(range)
len = range_len(range);
nd_pfn->npfns = PHYS_PFN((range_len(range) - offset));
range_len(&pmem->pgmap.range);
range_len(&pgmap->range), dev_to_node(&pdev->dev),
u64 phys_region_size = range_len(&phys_contig_mem_regions.regions[i]);
slot_add_mem_req.size = range_len(&phys_contig_mem_regions.regions[i]);
KUNIT_EXPECT_EQ(test, range_len(&phys_contig_mem_regions.regions[num - 1]),
range_len = em->len - (cur - em->start);
if (btrfs_test_range_bit_exists(&inode->io_tree, cur, cur + range_len - 1,
range_len = min(btrfs_extent_map_end(em), start + len) - cur;
last->len += range_len;
new->len = range_len;
u32 range_len = entry->len;
range_len = min_t(u32, range_len,
if (entry->start + range_len <= *last_scanned_ret)
((entry->start + range_len - 1) >> PAGE_SHIFT) -
ret = defrag_one_range(inode, entry->start, range_len,
*sectors_defragged += range_len >>
u64 range_len;
u64 range_len = umin(folio_next_pos(folio),
ASSERT(btrfs_folio_test_dirty(fs_info, folio, cur, range_len));
u64 range_len = umin(folio_next_pos(folio),
ASSERT(btrfs_folio_test_uptodate(fs_info, folio, cur, range_len));
*fileoff = cur + range_len;
u32 range_len = min_t(u64, folio_next_pos(folio),
bio_add_folio_nofail(&bbio->bio, folio, range_len,
wbc_account_cgroup_owner(wbc, folio, range_len);
u32 range_len;
range_len = min_t(u64, folio_next_pos(folio), end + 1) - range_start;
btrfs_folio_set_lock(fs_info, folio, range_start, range_len);
processed_end = range_start + range_len - 1;
u32 range_len = min_t(u64, folio_next_pos(folio),
bio_add_folio_nofail(&bbio->bio, folio, range_len,
const u64 range_len = last_byte - cur_offset;
ret = add_falloc_range(&reserve_list, cur_offset, range_len);
&data_reserved, cur_offset, range_len);
qgroup_reserved += range_len;
data_space_needed += range_len;
u32 range_len;
range_len = range_end + 1 - cur;
if (!btrfs_folio_test_ordered(fs_info, folio, cur, range_len)) {
btrfs_folio_clear_ordered(fs_info, folio, cur, range_len);
u64 range_start, range_len;
range_len = bg->length;
range_len = U64_MAX;
range_end = range_start + range_len;
u64 range_len = unode->aux - range_start + 1;
range_start + range_len <= start)
free_len = min(start + len, range_start + range_len) -
range_len(&devmem->pagemap.range));
pfn_last = pfn_first + (range_len(&devmem->pagemap.range) >> PAGE_SHIFT);
range_len(&devmem->pagemap.range));
const long nr_pages = (long)range_len(&r[i]);
PHYS_PFN(range_len(range)), NULL);
arch_remove_memory(range->start, range_len(range),
kasan_remove_zero_shadow(__va(range->start), range_len(range));
pfnmap_untrack(PHYS_PFN(range->start), range_len(range));
is_ram = region_intersects(range->start, range_len(range),
error = pfnmap_track(PHYS_PFN(range->start), range_len(range),
if (!mhp_range_allowed(range->start, range_len(range), !is_private)) {
PHYS_PFN(range_len(range)), params);
error = kasan_add_zero_shadow(__va(range->start), range_len(range));
error = arch_add_memory(nid, range->start, range_len(range),
PHYS_PFN(range_len(range)), params->altmap,
PHYS_PFN(range_len(range)), pgmap);
kasan_remove_zero_shadow(__va(range->start), range_len(range));
pfnmap_untrack(PHYS_PFN(range->start), range_len(range));
return (range->start + range_len(range)) >> PAGE_SHIFT;
PHYS_PFN(range_len(range)));
range_len(&res->range));
pgoff_end = dax_range->pgoff + PHYS_PFN(range_len(range)) - 1;