SZ_2G
.dma_zone_size = SZ_2G,
.dma_zone_size = SZ_2G,
.dma_zone_size = SZ_2G,
.dma_zone_size = SZ_2G,
#define MODULES_VSIZE (SZ_2G)
rlimit(RLIMIT_STACK) / 2, SZ_2G));
BUILD_BUG_ON(MODULES_VSIZE < SZ_2G);
if (kernel_size < SZ_2G)
module_plt_base = kernel_end - SZ_2G;
module_plt_base = random_bounding_box(SZ_2G, min, max);
module_plt_base ? (SZ_2G - kernel_size) / PAGE_SIZE : 0);
fallback_end = module_plt_base + SZ_2G;
end = module_plt_base + SZ_2G;
mem_array[2] = cpu_to_be32(PHYS_OFFSET + SZ_2G + SZ_256M);
#define MIN_DDW_VPMEM_DMA_WINDOW SZ_2G
if (TASK_SIZE <= UL(SZ_2G) && border >= UL(SZ_2G))
if (reladdr >= (long)SZ_2G || reladdr < -(long long)SZ_2G) {
if (reladdr >= SZ_2G || reladdr < -(long)SZ_2G) {
if (phb->dma_window_base_cur + phb->dma_window_size > SZ_2G) {
MEMBLOCK_LOW_LIMIT, SZ_2G,
#define KERNEL_LINK_ADDR (ADDRESS_SPACE_END - SZ_2G + 1)
#define MODULES_LOWEST_VADDR (KERNEL_LINK_ADDR - SZ_2G)
#define MODULES_VADDR (PFN_ALIGN((unsigned long)&_end) - SZ_2G)
kasan_mem_to_shadow((const void *)MODULES_VADDR + SZ_2G));
.end = SZ_2G,
vsize = round_up(SZ_2G + max_mappable, rte_size) +
amode31_lma = randomize_within_range(vmlinux.amode31_size, PAGE_SIZE, amode31_min, SZ_2G);
uv_info.uv_base_stor_len, SZ_1M, SZ_2G,
gmap_pv_destroy_range(kvm->arch.gmap, 0, gpa_to_gfn(SZ_2G), false);
min((u64)SZ_2G, prop->dram_end_address - cur_addr);
num_of_pkts = div64_u64(round_up(size, SZ_2G), SZ_2G);
chunk_size = min_t(u64, SZ_2G, end_addr - cur_addr);
lin_dma_pkts_cnt = DIV_ROUND_UP_ULL(size, SZ_2G);
lin_dma_pkt->tsize = cpu_to_le32(SZ_2G);
size -= SZ_2G;
addr += SZ_2G;
ivpu_hw_range_init(vdev, &vdev->hw->ranges.shave, 0x180000000, SZ_2G);
ivpu_hw_range_init(vdev, &vdev->hw->ranges.shave, 0x80000000, SZ_2G);
linear_addr = linear_addr - SZ_32G + SZ_2G;
dma_set_max_seg_size(dev, SZ_2G);
if (cmdbuf_paddr >= SZ_2G)
priv->mmu_global->memory_base = SZ_2G;
} else if (cmdbuf_paddr + SZ_128M >= SZ_2G) {
priv->mmu_global->memory_base = SZ_2G;
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
mem = mock_region_create(i915, 0, SZ_2G, I915_GTT_PAGE_SIZE_4K, 0, 0);
#define ROGUE_PM_MAX_FREELIST_SIZE SZ_2G
u64 mm_size = SZ_8G + SZ_2G, size = SZ_8G + SZ_1G, min_block_size = SZ_8G;
KUNIT_EXPECT_EQ(test, SZ_2G, pf_profile_fair_ggtt(gt, 1));
KUNIT_EXPECT_EQ(test, SZ_2G + SZ_1G + SZ_512M, pf_profile_fair_ggtt(gt, 1));
u64 shareable = SZ_2G + SZ_1G + SZ_512M;
KUNIT_ASSERT_EQ(test, SZ_2G, pf_profile_fair_ggtt(gt, num_vfs));
xe_res_next(&cursor, min_t(u64, cursor.size, SZ_2G));
size_t size = min_t(u64, cursor.size, SZ_2G);
SZ_2G - SZ_4K, false);
dma_set_max_seg_size(dev, SZ_2G);
caps->page_size_cap |= (SZ_4M | SZ_1G | SZ_2G);
dma_set_max_seg_size(&pdev->dev, SZ_2G);
dma_set_max_seg_size(&dev->dev, SZ_2G);
if (len <= SZ_2G)
priv->info.pgsize_bitmap % (IS_32BIT ? SZ_2G : SZ_16G);
if (ops->iova_to_phys(ops, SZ_2G + 42))
if (ops->iova_to_phys(ops, SZ_2G + 42))
mapping = arm_iommu_create_mapping(dev, SZ_1G, SZ_2G);
.mminfra_offset = SZ_2G,
#define MCU_CACHE_OFFSET SZ_2G
vpu->enable_4GB = !!(totalram_pages() > (SZ_2G >> PAGE_SHIFT));
mapping = arm_iommu_create_mapping(isp->dev, SZ_1G, SZ_2G);
if (si.totalram * si.mem_unit > SZ_2G) {
if (chipsize == SZ_2G) {
[RP_BAR0] = _ULL(128 * SZ_2G),
[RP_BAR1] = SZ_2G,
case SZ_2G:
case SZ_2G:
case SZ_2G:
(pci_offset < SZ_4G && pci_offset > SZ_2G)) {
#define NR_PAGES_2GB (SZ_2G / PAGE_SIZE)
#define BTRFS_MAX_TRIM_LENGTH SZ_2G
.expected_len = SZ_2G,
.min_hole_size = SZ_2G,
.expected_start = SZ_2G,
.min_hole_size = SZ_2G,
{ .start = SZ_2G, .len = SZ_1G },
.expected_len = SZ_2G,
.min_hole_size = SZ_2G,
.expected_start = SZ_2G,
.min_hole_size = SZ_2G,
.expected_len = SZ_2G,
.min_hole_size = SZ_2G,
{ .start = SZ_2G, .len = SZ_1G },
.expected_len = SZ_2G,
.expected_start = SZ_2G,
.expected_len = SZ_2G,
.expected_len = SZ_2G,
{ .start = 9ULL * SZ_1G, .len = SZ_2G },
.expected_pending_end = SZ_2G - 1,
.pending_extent = { 0, SZ_2G },
.expected_pending_end = SZ_2G - 1,
.pending_extent = { SZ_2G, SZ_1G },
.expected_pending_start = SZ_2G,
.pending_extent = { SZ_2G, SZ_1G },
.hole_len = SZ_2G,
.pending_extent = { SZ_1G, SZ_2G },
{ .start = 0, .len = SZ_2G },
.expected_start = SZ_2G,
{ .start = SZ_2G, .len = SZ_1G },
#define DEV_SIZE SZ_2G
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_2G, SZ_1M));
.base = SZ_2G,
slot_size = SZ_2G;