SZ_256M
dma_direct_set_offset(data, PHYS_OFFSET, BUS_OFFSET, SZ_256M);
dma_direct_set_offset(&isa_dma_dev, PHYS_OFFSET, BUS_OFFSET, SZ_256M);
.size = SZ_256M - (SZ_4M + SZ_64M),
.dma_zone_size = SZ_256M,
#define MODULES_END (MODULES_VADDR + SZ_256M)
{ 0, SZ_256M + SZ_128M },
if (size <= SZ_256M)
size -= SZ_256M;
mem_array[2] = cpu_to_be32(PHYS_OFFSET + SZ_2G + SZ_256M);
size_preio = min_t(unsigned long, size, SZ_256M);
if (n << 28 < ALIGN(TASK_SIZE, SZ_256M))
#define KASAN_KERN_START ALIGN_DOWN(PAGE_OFFSET - SZ_256M, SZ_256M)
#define MODULES_END (ASM_CONST(CONFIG_PAGE_OFFSET) & ~(UL(SZ_256M) - 1))
#define MODULES_BASE (MODULES_VADDR & ~(UL(SZ_256M) - 1))
DEFINE(NUM_USER_SEGMENTS, ALIGN(TASK_SIZE, SZ_256M) >> 28);
unsigned int max_size = SZ_256M;
if (addr < ALIGN_DOWN(MODULES_VADDR, SZ_256M))
if (addr > ALIGN(MODULES_END, SZ_256M) - 1)
for (i = ALIGN(TASK_SIZE, SZ_256M) >> 28; i < 16; i++) {
memblock_set_current_limit(min_t(u64, first_memblock_size, SZ_256M));
if (*block_size > SZ_256M)
*block_size = SZ_256M;
for (i = 0; i < ALIGN(TASK_SIZE, SZ_256M) >> 28; i++)
mask |= SZ_16M | SZ_256M;
__builtin_ctzll(SZ_256M), __builtin_ctzll(SZ_16G), __builtin_ctzll(SZ_2M)
(SZ_256M), (SZ_16G), (SZ_2M)
.end = 0x20000000 + SZ_256M - 1,
.end = 0x30000000 + SZ_256M - 1,
max_gap_start = SZ_256M;
ivpu_hw_range_init(vdev, &vdev->hw->ranges.global, 0x90000000, SZ_256M);
ivpu_hw_range_init(vdev, &vdev->hw->ranges.global, 0x90000000, SZ_256M);
if (align >= SZ_256M) {
if (!IS_ALIGNED(cfmws->base_hpa, SZ_256M)) {
if (!IS_ALIGNED(cfmws->window_size, SZ_256M)) {
hpa_range.start = ALIGN_DOWN(hpa_range.start, SZ_256M);
hpa_range.end = ALIGN(hpa_range.end, SZ_256M) - 1;
if (!IS_ALIGNED(size, SZ_256M))
div64_u64_rem(p->res->start, (u64)hbiw * SZ_256M, &rem);
div64_u64_rem(size, (u64)SZ_256M * p->interleave_ways, &remainder);
res = alloc_free_mem_region(cxlrd->res, size, SZ_256M,
#define CXL_CAPACITY_MULTIPLIER SZ_256M
GEM_BUG_ON(pci_resource_len(pdev, GEN12_LMEM_BAR) != SZ_256M);
io_size = max_t(u64, io_size, SZ_256M); /* 256M seems to be the common lower limit */
#define ROGUE_RGNHDR_HEAP_SIZE SZ_256M
#define ROGUE_RISCVFW_REGION_SIZE SZ_256M
#define PANTHOR_VM_MIN_KERNEL_VA_SIZE SZ_256M
KUNIT_ASSERT_EQ(test, SZ_256M, pf_profile_fair_ggtt(gt, num_vfs));
dom->domain.geometry.aperture_end = SZ_256M - 1;
case SZ_256M:
.size = SZ_256M,
.size = SZ_256M,
.size = SZ_256M,
if (params->size == SZ_256M)
if (params->size == SZ_256M)
#define IS_256MB_ALIGNED(x) IS_ALIGNED(x, SZ_256M)
case SZ_256M:
window_size = SZ_256M;
case SZ_256M:
if (resource_size(mem) != SZ_256M) {
(mem->start != v3->non_pre_mem + SZ_256M)) {
if (resource_size(mem) != SZ_256M) {
case SZ_256M:
if (hp->dxfer_len >= SZ_256M) {
#define STM32_OSPI_MAX_MMAP_SZ SZ_256M
#define STM32_QSPI_MAX_MMAP_SZ SZ_256M
cache_size = div_u64(block_group->length, SZ_256M);
cur_bytes = min_t(u64, num_bytes, SZ_256M);
if (new_size < SZ_256M) {
return SZ_256M;
val &= ~((u64)SZ_256M - 1);
if (val < SZ_256M)
.data_stripe_size = SZ_256M,
{SZ_64M - SZ_4M, SZ_64M - SZ_4M + SZ_256M},
.data_stripe_size = SZ_256M,
.data_stripe_phys_start = {SZ_256M},
.window_size = SZ_256M * 4UL,
.window_size = SZ_256M * 8UL,
.window_size = SZ_256M * 4UL,
.window_size = SZ_256M * 8UL,
.window_size = SZ_256M * 4UL,
.window_size = SZ_256M,
.window_size = SZ_256M * 8UL,
.window_size = SZ_256M * 8UL,
res = alloc_mock_res(window->window_size, SZ_256M);
cpu_to_le64(SZ_256M / CXL_CAPACITY_MULTIPLIER),
.size = SZ_256M
.base = SZ_256M,
.base = SZ_256M,
.base = SZ_256M,
#define MEM_GPA SZ_256M