Symbol: SECTION_SIZE
arch/arm/include/asm/memory.h
87
#define FDT_FIXED_SIZE (2 * SECTION_SIZE)
arch/arm/include/asm/memory.h
88
#define FDT_VIRT_BASE(physbase) ((void *)(FDT_FIXED_BASE | (physbase) % SECTION_SIZE))
arch/arm/include/asm/pgtable-2level.h
97
#define SECTION_MASK (~(SECTION_SIZE-1))
arch/arm/kernel/efi.c
42
if (round_down(base + size, SECTION_SIZE) <
arch/arm/kernel/efi.c
43
round_up(base, SECTION_SIZE) + SECTION_SIZE)
arch/arm/kernel/setup.c
1028
start = memblock_phys_alloc_range(crash_size, SECTION_SIZE,
arch/arm/mach-omap2/omap-secure.c
115
size = ALIGN(size, SECTION_SIZE);
arch/arm/mach-omap2/omap-secure.c
116
omap_secure_memblock_base = arm_memblock_steal(size, SECTION_SIZE);
arch/arm/mach-tegra/iomap.h
107
#define IO_PPSB_SIZE SECTION_SIZE
arch/arm/mach-tegra/iomap.h
111
#define IO_APB_SIZE SECTION_SIZE
arch/arm/mm/dump.c
357
if (SECTION_SIZE < PMD_SIZE && pmd_leaf(pmd[1])) {
arch/arm/mm/dump.c
358
addr += SECTION_SIZE;
arch/arm/mm/idmap.c
62
addr += SECTION_SIZE;
arch/arm/mm/init.c
321
if (addr & SECTION_SIZE)
arch/arm/mm/init.c
327
local_flush_tlb_kernel_range(addr, addr + SECTION_SIZE);
arch/arm/mm/init.c
349
if (!IS_ALIGNED(perms[i].start, SECTION_SIZE) ||
arch/arm/mm/init.c
350
!IS_ALIGNED(perms[i].end, SECTION_SIZE)) {
arch/arm/mm/init.c
353
SECTION_SIZE);
arch/arm/mm/init.c
359
addr += SECTION_SIZE)
arch/arm/mm/mmu.c
1074
vm_reserve_area_early(addr, SECTION_SIZE, pmd_empty_section_gap);
arch/arm/mm/mmu.c
1095
if ((addr & ~PMD_MASK) == SECTION_SIZE) {
arch/arm/mm/mmu.c
1107
if ((addr & ~PMD_MASK) == SECTION_SIZE) {
arch/arm/mm/mmu.c
1587
phys_addr_t kernel_x_end = round_up(__pa(__init_end), SECTION_SIZE);
arch/arm/mm/mmu.c
1760
kernel_sec_end = round_up(__pa(_end), SECTION_SIZE);
arch/arm/mm/mmu.c
791
if (addr & SECTION_SIZE)
arch/arm/mm/mmu.c
796
phys += SECTION_SIZE;
arch/arm/mm/mmu.c
797
} while (pmd++, addr += SECTION_SIZE, addr != end);