map_end
u64 map_end = (real_end + 0xfff) & ~0xfff;
map_end - map_start, 12, 0,
bus_addr, map_end - map_start,
unsigned long map_end)
map_end);
real_end = max(map_start, ALIGN_DOWN(map_end, PMD_SIZE));
if (real_end < map_end)
init_range_memory_mapping(real_end, map_end);
unsigned long map_end)
while (start < map_end) {
if (step_size && map_end - start > step_size) {
if (next > map_end)
next = map_end;
next = map_end;
const efi_memory_desc_t *end = efi.memmap.map_end;
return efi.memmap.map_end - efi.memmap.desc_size;
if (entry >= efi.memmap.map_end)
old < old_memmap->map_end;
map.map_end = map.map + data->size;
goto map_end;
map_end:
resource_size_t map_start, map_end, busy = 0;
map_end = map_start + nd_mapping->size - 1;
if (start >= map_start && start < map_end) {
if (end > map_end) {
} else if (end >= map_start && end <= map_end) {
resource_size_t map_end = nd_mapping->start + nd_mapping->size - 1;
nsl_get_dpa(ndd, nd_label) > map_end)
unsigned long map_end = m->base + m->num - 1;
if ((start + size - 1) <= map_end)
return map_end(map, mr) - map_start(map, mr);
unsigned long map_start, map_end;
map_end = map->pages_vm_start + (map->count << PAGE_SHIFT);
if (map_start >= range->end || map_end <= range->start)
mend = min(range->end, map_end);
map->index, map->count, map_start, map_end,
loff_t map_end = (loff_t) map->m_lblk + map->m_len;
if (map_end < ex_end) {
lblk = map_end;
len = ex_end - map_end;
pblk = ext4_ext_pblock(ex) + (map_end - ee_block);
loff_t offset, end, map_end;
map_end = ((loff_t)page_index + 1) << PAGE_SHIFT;
if (map_end > end)
map_end = end;
if (map_end & (PAGE_SIZE - 1))
to = map_end & (PAGE_SIZE - 1);
offset, map_end - 1);
offset = map_end;
void *map_end;
(md) && ((void *)(md) + (m)->desc_size) <= (m)->map_end; \
unsigned char *map_end = map + nr_pages;
} while (++map < map_end);
unsigned char *map_end = map + nr_pages;
} while (++map < map_end);