region_end
unsigned int region_end;
table->end = base_addr + end->region_end;
start->region_end > (start+1)->region_start) {
start->region_end += base_addr;
__start___unwind[i].region_end);
e->region_start, e->region_end, e->Save_SP, e->Save_RP,
else if (addr > e->region_end)
region_start, region_end);
unsigned long region_start, unsigned long region_end)
nid, region_start, region_end);
unsigned long region_start, unsigned long region_end)
if (region_end)
max_addr = region_end;
unsigned long region_start, unsigned long region_end)
region_start, region_end);
region_end);
u64 region_end;
region_end = min(entry->start + entry->size, mem_limit);
if (region.start > region_end)
region.size = region_end - region.start;
u64 start, end, entry_end, region_end;
region_end = region->start + region->size;
entry_end = clamp(region_end, start, end);
region.end = a->region_end;
u64 first_slot, last_slot, region_end;
region_end = min(md->phys_addr + md->num_pages * EFI_PAGE_SIZE - 1,
if (region_end < size)
last_slot = round_down(region_end - size + 1, align);
reg_region_cur <= reg->region_end;
int32_t region_start, region_end;
region_end = NUMBER_REGIONS - MAX_LOW_POINT;
region_end = 1;
for (i = region_end - region_start; i < MAX_REGIONS_NUMBER ; i++)
for (k = 0; k < (region_end - region_start); k++) {
start_index = (region_end + MAX_LOW_POINT) * NUMBER_SW_SEGMENTS;
dc_fixpt_from_int(region_end));
int32_t region_start, region_end;
region_end = 0;
for (i = region_end - region_start; i < MAX_REGIONS_NUMBER ; i++)
for (k = 0; k < (region_end - region_start); k++) {
start_index = (region_end + MAX_LOW_POINT) * NUMBER_SW_SEGMENTS;
dc_fixpt_from_int(region_end));
uint32_t region_end
int32_t region_start, region_end;
region_end = NUMBER_REGIONS - MAX_LOW_POINT;
region_end = 1;
for (i = region_end - region_start; i < MAX_REGIONS_NUMBER ; i++)
for (k = 0; k < (region_end - region_start); k++) {
start_index = (region_end + MAX_LOW_POINT) * NUMBER_SW_SEGMENTS;
dc_fixpt_from_int(region_end));
reg_region_cur <= reg->region_end;
gam_regs.region_end = REG(CM_RGAM_RAMA_REGION_32_33);
gam_regs.region_end = REG(CM_RGAM_RAMB_REGION_32_33);
gam_regs.region_end = REG(CM_DGAM_RAMB_REGION_14_15);
gam_regs.region_end = REG(CM_DGAM_RAMA_REGION_14_15);
gam_regs.region_end = REG(CM_BLNDGAM_RAMA_REGION_32_33);
gam_regs.region_end = REG(CM_BLNDGAM_RAMB_REGION_32_33);
gam_regs.region_end = REG(CM_BLNDGAM_RAMA_REGION_32_33);
gam_regs.region_end = REG(CM_BLNDGAM_RAMB_REGION_32_33);
gam_regs.region_end = REG(CM_GAMCOR_RAMB_REGION_32_33);
gam_regs.region_end = REG(CM_GAMCOR_RAMA_REGION_32_33);
gam_regs.region_end = REG(DWB_OGAM_RAMA_REGION_32_33);
gam_regs.region_end = REG(DWB_OGAM_RAMB_REGION_32_33);
int32_t region_start, region_end;
region_end = region_start + NUMBER_REGIONS;
region_end = 1;
for (k = 0; k < (region_end - region_start); k++) {
start_index = (region_end + MAX_LOW_POINT) * NUMBER_SW_SEGMENTS;
dc_fixpt_from_int(region_end));
gam_regs.region_end = REG(MPCC_OGAM_RAMB_REGION_32_33[mpcc_id]);
gam_regs.region_end = REG(MPCC_OGAM_RAMA_REGION_32_33[mpcc_id]);
gam_regs.region_end = REG(MPCC_OGAM_RAMA_REGION_32_33[mpcc_id]);
gam_regs.region_end = REG(MPCC_OGAM_RAMB_REGION_32_33[mpcc_id]);
gam_regs.region_end = REG(MPCC_MCM_1DLUT_RAMA_REGION_32_33[mpcc_id]);
gam_regs.region_end = REG(MPCC_MCM_1DLUT_RAMB_REGION_32_33[mpcc_id]);
u64 region_end = region_start + size;
region_width = max(fls64(region_start ^ (region_end - 1)),
u64 region_end = *region_start + *size;
region_width = max(fls64(*region_start ^ (region_end - 1)),
unsigned long region_start, region_end;
region_end = region_start + nr_regions - 1;
if (region_end == clone->nr_regions - 1) {
loff_t region_end = 0;
region_end = region->offset + region->erasesize * region->numblocks;
if (region && addr == region_end) {
region_end = region->offset + region->erasesize * region->numblocks;
u32 region_end;
region_end = region_start + region_size;
region_end += (mtd->erasesize - 1);
region_end &= ~(mtd->erasesize -1);
region_start, region_end);
part->size = region_end - region_start;
u64 region_end;
region_end = region->offset + region->size;
while (len && addr >= region->offset && addr < region_end) {
__u64 region_end; /* available since kernel version 2.6.36 */
unsigned long region_start, region_end;
region_end = ALIGN_DOWN(end, KASAN_MEMORY_PER_SHADOW_PAGE);
if (end != region_end &&
free_region_end > region_end)
region_end += KASAN_MEMORY_PER_SHADOW_PAGE;
shadow_end = kasan_mem_to_shadow((void *)region_end);
phys_addr_t region_end;
region_end = memblock_start_of_DRAM();
while (region_end < memblock_end_of_DRAM()) {
memblock_reserve(region_end + gap_size, region_size);
region_end += gap_size + region_size;
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
req_node_end = region_end(req_node);
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
max_addr = region_end(max_node);
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
ASSERT_LE(region_end(new_rgn), region_end(req_node));
ASSERT_LE(region_end(new_rgn), region_end(req_node));
req_node_end = region_end(req_node);
ASSERT_LE(region_end(new_rgn), req_node_end);
ASSERT_LE(region_end(new_rgn), region_end(req_node));
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
max_addr = region_end(max_node);
ASSERT_LE(region_end(new_rgn), region_end(req_node));
max_addr = region_end(max_node);
ASSERT_LE(region_end(req_node), region_end(new_rgn));
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
ASSERT_EQ(new_rgn->base, region_end(exp_node) - size);
ASSERT_EQ(new_rgn->base, region_end(exp_node) - size);
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
ASSERT_EQ(new_rgn->base, region_end(exp_node) - size);
req_node_end = region_end(req_node);
exp_node_end = region_end(exp_node);
ASSERT_EQ(new_rgn->base, region_end(req_node) - size);
max_addr = region_end(max_node);
max_addr = region_end(max_node);
ASSERT_LE(region_end(new_rgn), region_end(req_node));
ASSERT_LE(region_end(new_rgn), region_end(exp_node));
ASSERT_LE(region_end(new_rgn), region_end(exp_node));
ASSERT_LE(region_end(new_rgn), region_end(req_node));
ASSERT_LE(region_end(new_rgn), region_end(exp_node));
req_node_end = region_end(req_node);
ASSERT_LE(region_end(new_rgn), req_node_end);
exp_node_end = region_end(req_node);
ASSERT_LE(region_end(new_rgn), exp_node_end);
ASSERT_LE(region_end(new_rgn), region_end(req_node));
max_addr = region_end(max_node);
ASSERT_LE(region_end(new_rgn), region_end(min_node));
max_addr = region_end(max_node);
ASSERT_LE(region_end(new_rgn), region_end(min_node));
ASSERT_LE(region_end(req_node), region_end(new_rgn));