arch/arm/include/asm/ptdump.h
24
#define EFI_RUNTIME_MAP_END SZ_1G
arch/arm/mach-bcm/bcm2711.c
21
.dma_zone_size = SZ_1G,
arch/arm/mach-highbank/highbank.c
166
.dma_zone_size = (4ULL * SZ_1G),
arch/arm64/include/asm/efi.h
85
return (image_addr & ~(SZ_1G - 1UL)) + (1UL << (VA_BITS_MIN - 1));
arch/arm64/include/asm/kvm_pkvm.h
142
res += __hyp_pgtable_max_pages(SZ_1G >> PAGE_SHIFT);
arch/arm64/include/asm/kvm_pkvm.h
158
res += __hyp_pgtable_max_pages(SZ_1G >> PAGE_SHIFT);
arch/arm64/include/asm/memory.h
51
#define VMEMMAP_END (-UL(SZ_1G))
arch/arm64/kernel/machine_kexec_file.c
148
kbuf.buf_max = round_down(kernel_load_addr, SZ_1G)
arch/arm64/kernel/machine_kexec_file.c
149
+ (unsigned long)SZ_1G * 32;
arch/arm64/kvm/nested.c
1044
scope->size = SZ_1G;
arch/arm64/kvm/nested.c
1072
scope->size = SZ_1G;
arch/arm64/kvm/nested.c
440
max_size = SZ_1G;
arch/arm64/kvm/nested.c
542
else if (sz < SZ_1G) sz = SZ_1G;
arch/arm64/kvm/nested.c
603
max_size = SZ_1G;
arch/loongarch/kernel/machine_kexec_file.c
208
kbuf.buf_max = round_down(kernel_load_addr, SZ_1G) + (unsigned long)SZ_1G * 32;
arch/powerpc/mm/book3s64/hash_utils.c
2483
return randomize_page(max_t(unsigned long, mm->brk, SZ_1T), SZ_1G);
arch/powerpc/mm/book3s64/hash_utils.c
2485
return randomize_page(mm->brk, SZ_1G);
arch/powerpc/platforms/powernv/memtrace.c
74
#define FLUSH_CHUNK_SIZE SZ_1G
arch/riscv/mm/init.c
293
dma32_phys_limit = min(4UL * SZ_1G, (unsigned long)PFN_PHYS(max_low_pfn));
arch/riscv/mm/init.c
94
#define LOG2_SZ_1G ilog2(SZ_1G)
arch/riscv/mm/ptdump.c
116
{ SZ_1G, "UEFI runtime end" },
arch/s390/include/asm/extmem.h
16
#define MAX_DCSS_ADDR (512UL * SZ_1G)
arch/x86/kernel/process.c
1031
return randomize_page(mm->brk, SZ_1G);
arch/x86/mm/mem_encrypt.c
135
size = clamp_val(size, IO_TLB_DEFAULT_SIZE, SZ_1G);
arch/x86/virt/vmx/tdx/tdx.c
357
#define TDMR_ALIGNMENT SZ_1G
drivers/accel/habanalabs/common/command_buffer.c
15
#define CB_VA_POOL_SIZE (4UL * SZ_1G)
drivers/accel/habanalabs/common/debugfs.c
675
if (len_bytes == 0 || len_bytes > SZ_1G) {
drivers/accel/habanalabs/common/device.c
2354
hdev->asic_prop.dram_size / SZ_1G);
drivers/dma/bcm2835-dma.c
172
#define MAX_DMA_LEN SZ_1G
drivers/dma/tegra186-gpc-dma.c
1310
.max_dma_count = SZ_1G,
drivers/dma/tegra186-gpc-dma.c
1318
.max_dma_count = SZ_1G,
drivers/dma/tegra186-gpc-dma.c
1326
.max_dma_count = SZ_1G,
drivers/edac/bluefield_edac.c
324
(SZ_1G / PAGE_SIZE);
drivers/edac/versal_edac.c
579
size *= SZ_1G;
drivers/firmware/qcom/qcom_scm.c
818
if (!ret && res.result[2] > SZ_1G) {
drivers/gpu/drm/amd/amdgpu/amdgpu_device.c
418
if (memory_carved >= SZ_1G/SZ_1M) {
drivers/gpu/drm/etnaviv/etnaviv_iommu_v2.c
305
drm_mm_init(&context->mm, SZ_4K, (u64)SZ_1G * 4 - SZ_4K);
drivers/gpu/drm/i915/gem/i915_gem_stolen.c
940
lmem_size *= SZ_1G;
drivers/gpu/drm/i915/gem/selftests/huge_pages.c
1383
{ igt_create_local, SZ_64K, SZ_1G, },
drivers/gpu/drm/i915/gt/intel_region_lmem.c
211
lmem_size *= SZ_1G;
drivers/gpu/drm/i915/gt/selftest_tlb.c
202
resource_size_t size = SZ_1G;
drivers/gpu/drm/i915/selftests/i915_gem_gtt.c
1008
unsigned int sizes[] = { SZ_2M, SZ_1G };
drivers/gpu/drm/i915/selftests/intel_memory_region.c
567
total = max_t(u64, total, SZ_1G);
drivers/gpu/drm/i915/selftests/intel_memory_region.c
809
for (ps = PAGE_SIZE; ps <= SZ_1G; ps <<= 1) {
drivers/gpu/drm/panfrost/panfrost_mmu.c
378
blk_offset = -addr % SZ_1G ?: SZ_1G;
drivers/gpu/drm/panthor/panthor_mmu.c
849
blk_offset = -addr % SZ_1G ?: SZ_1G;
drivers/gpu/drm/tests/drm_buddy_test.c
862
u64 mm_size = SZ_8G + SZ_2G, size = SZ_8G + SZ_1G, min_block_size = SZ_8G;
drivers/gpu/drm/vc4/vc4_kms.c
686
if (load_state->membus_load > SZ_1G + SZ_512M)
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
159
KUNIT_EXPECT_EQ(test, SZ_2G + SZ_1G + SZ_512M, pf_profile_fair_ggtt(gt, 1));
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
168
u64 shareable = SZ_2G + SZ_1G + SZ_512M;
drivers/gpu/drm/xe/tests/xe_gt_sriov_pf_config_kunit.c
187
KUNIT_ASSERT_EQ(test, SZ_1G, pf_profile_fair_ggtt(gt, num_vfs));
drivers/gpu/drm/xe/xe_devcoredump.c
171
#define XE_DEVCOREDUMP_CHUNK_MAX (SZ_512M + SZ_1G)
drivers/gpu/drm/xe/xe_migrate.c
135
(xe->mem.vram), SZ_1G);
drivers/gpu/drm/xe/xe_migrate.c
164
pos += SZ_1G, ofs += 8) {
drivers/gpu/drm/xe/xe_migrate.c
165
if (pos + SZ_1G >= vram_limit) {
drivers/gpu/drm/xe/xe_migrate.c
321
xe_assert(xe, actual_phy_size <= (MAX_NUM_PTE - IDENTITY_OFFSET) * SZ_1G);
drivers/gpu/drm/xe/xe_migrate.c
330
DIV_ROUND_UP_ULL(actual_phy_size, SZ_1G);
drivers/gpu/drm/xe/xe_migrate.c
334
IDENTITY_OFFSET / 2) * SZ_1G);
drivers/gpu/drm/xe/xe_vm.c
2500
return SZ_1G;
drivers/gpu/drm/xe/xe_vm.c
2508
return SZ_1G; /* Uninitialized, used max size */
drivers/gpu/drm/xe/xe_vm.c
2514
case SZ_1G:
drivers/gpu/drm/xe/xe_vram.c
156
*tile_size = (u64)REG_FIELD_GET(GENMASK(17, 8), reg) * SZ_1G;
drivers/gpu/drm/xe/xe_vram.c
157
*tile_offset = (u64)REG_FIELD_GET(GENMASK(7, 1), reg) * SZ_1G;
drivers/infiniband/hw/irdma/icrdma_hw.c
193
dev->hw_attrs.page_size_cap = SZ_4K | SZ_2M | SZ_1G;
drivers/infiniband/hw/irdma/ig3rdma_hw.c
128
dev->hw_attrs.page_size_cap = SZ_4K | SZ_2M | SZ_1G;
drivers/infiniband/hw/mana/main.c
715
caps->page_size_cap |= (SZ_4M | SZ_1G | SZ_2G);
drivers/iommu/arm/arm-smmu-v3/arm-smmu-v3.c
4531
smmu->pgsize_bitmap |= SZ_4K | SZ_2M | SZ_1G;
drivers/iommu/arm/arm-smmu/arm-smmu.c
1921
smmu->pgsize_bitmap |= SZ_4K | SZ_2M | SZ_1G;
drivers/iommu/generic_pt/kunit_iommu_pt.h
315
if (range->last_va - range->va > SZ_1G)
drivers/iommu/generic_pt/kunit_iommu_pt.h
316
range->last_va = range->va + SZ_1G;
drivers/iommu/intel/iommu.c
2865
dmar_domain->domain.pgsize_bitmap &= ~(u64)SZ_1G;
drivers/iommu/intel/iommu.c
2963
dmar_domain->domain.pgsize_bitmap &= ~(u64)SZ_1G;
drivers/iommu/intel/iommu.c
3036
(dmar_domain->domain.pgsize_bitmap & SZ_1G))
drivers/iommu/intel/iommu.c
3081
if (!(sslps & BIT(1)) && (dmar_domain->domain.pgsize_bitmap & SZ_1G))
drivers/iommu/io-pgtable-arm-selftests.c
106
iova += SZ_1G;
drivers/iommu/io-pgtable-arm-selftests.c
128
iova += SZ_1G;
drivers/iommu/io-pgtable-arm-selftests.c
157
SZ_4K | SZ_2M | SZ_1G,
drivers/iommu/io-pgtable-arm-selftests.c
78
if (ops->iova_to_phys(ops, SZ_1G + 42))
drivers/iommu/io-pgtable-arm-v7s.c
843
if (ops->iova_to_phys(ops, SZ_1G + 42))
drivers/iommu/io-pgtable-arm.c
1166
cfg->pgsize_bitmap &= (SZ_4K | SZ_2M | SZ_1G);
drivers/iommu/io-pgtable-arm.c
1176
cfg->pgsize_bitmap &= (SZ_4K | SZ_2M | SZ_1G);
drivers/iommu/io-pgtable-arm.c
1192
cfg->pgsize_bitmap &= (SZ_4K | SZ_2M | SZ_1G);
drivers/iommu/io-pgtable-arm.c
896
page_sizes = (SZ_4K | SZ_2M | SZ_1G);
drivers/iommu/ipmmu-vmsa.c
574
domain->io_domain.pgsize_bitmap = SZ_1G | SZ_2M | SZ_4K;
drivers/iommu/ipmmu-vmsa.c
810
mapping = arm_iommu_create_mapping(dev, SZ_1G, SZ_2G);
drivers/iommu/mtk_iommu.c
361
{ .iova_base = SZ_1G, .size = 0xc0000000}, /* APU CODE */
drivers/iommu/riscv/iommu.c
1422
domain->domain.pgsize_bitmap = va_mask & (SZ_4K | SZ_2M | SZ_1G | SZ_512G);
drivers/media/platform/ti/omap3isp/isp.c
1938
mapping = arm_iommu_create_mapping(isp->dev, SZ_1G, SZ_2G);
drivers/mtd/nand/raw/loongson-nand-controller.c
696
case SZ_1G:
drivers/pci/controller/dwc/pcie-designware.c
1006
pci->region_align / SZ_1K, (pci->region_limit + 1) / SZ_1G);
drivers/pci/controller/pci-ftpci100.c
158
case SZ_1G:
drivers/pci/controller/pci-rcar-gen2.c
188
window_size = SZ_1G;
drivers/pci/controller/pci-rcar-gen2.c
216
case SZ_1G:
drivers/pci/controller/pci-v3-semi.c
650
case SZ_1G:
drivers/pci/controller/pcie-iproc.c
210
.size_unit = SZ_1G,
drivers/pci/controller/pcie-iproc.c
220
.size_unit = SZ_1G,
drivers/pci/controller/plda/pcie-microchip-host.c
677
MPFS_NC_BOUNCE_ADDR, SZ_1G);
drivers/pci/controller/plda/pcie-microchip-host.c
678
mc_pcie_setup_inbound_atr(port, 1, SZ_1G,
drivers/pci/controller/plda/pcie-microchip-host.c
679
MPFS_NC_BOUNCE_ADDR + SZ_1G, SZ_1G);
drivers/vfio/pci/nvgrace-gpu/main.c
1044
resmem_size = SZ_1G;
fs/btrfs/block-group.c
2954
u64 div = SZ_1G;
fs/btrfs/block-group.c
2961
if (btrfs_super_total_bytes(fs_info->super_copy) <= (SZ_1G * 10ULL))
fs/btrfs/ctree.c
685
search_start = round_down(buf->start, SZ_1G);
fs/btrfs/free-space-cache.c
716
if (size < SZ_1G)
fs/btrfs/free-space-cache.c
719
max_bytes = MAX_CACHE_BYTES_PER_GIG * div_u64(size, SZ_1G);
fs/btrfs/space-info.c
222
if (fs_info->fs_devices->total_rw_bytes > 50ULL * SZ_1G)
fs/btrfs/space-info.c
223
return SZ_1G;
fs/btrfs/space-info.c
440
return min_t(u64, data_chunk_size, SZ_1G);
fs/btrfs/tests/chunk-allocation-tests.c
106
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
109
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
113
.expected_len = 8ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
118
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
121
{ .start = SZ_2G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
122
{ .start = 4ULL * SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
131
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
134
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
135
{ .start = 5ULL * SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
139
.expected_len = 3ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
144
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
147
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
148
{ .start = 3ULL * SZ_1G, .len = 5ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
151
.expected_start = 8ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
157
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
160
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
161
{ .start = 3ULL * SZ_1G, .len = 6ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
165
.expected_len = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
170
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
171
.min_hole_size = 3ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
173
{ .start = SZ_2G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
174
{ .start = 4ULL * SZ_1G, .len = 5ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
183
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
184
.min_hole_size = 3ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
186
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
187
{ .start = 4ULL * SZ_1G, .len = 5ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
196
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
197
.min_hole_size = 3ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
199
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
200
{ .start = 3ULL * SZ_1G, .len = 5ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
203
.expected_start = 8ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
209
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
210
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
212
{ .start = 0, .len = 10ULL * SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
215
.expected_start = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
221
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
222
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
224
{ .start = 9ULL * SZ_1G, .len = SZ_2G },
fs/btrfs/tests/chunk-allocation-tests.c
228
.expected_len = 9ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
232
.hole_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
234
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
237
.expected_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
338
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
344
.hole_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
345
.hole_len = 9ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
346
.pending_extent = { SZ_1G, SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
348
.expected_pending_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
353
.hole_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
354
.hole_len = 9ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
363
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
364
.pending_extent = { SZ_2G, SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
367
.expected_pending_end = 3ULL * SZ_1G - 1,
fs/btrfs/tests/chunk-allocation-tests.c
372
.hole_len = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
373
.pending_extent = { SZ_2G, SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
380
.pending_extent = { SZ_1G, SZ_2G },
fs/btrfs/tests/chunk-allocation-tests.c
382
.expected_pending_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
383
.expected_pending_end = 3ULL * SZ_1G - 1,
fs/btrfs/tests/chunk-allocation-tests.c
48
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
49
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
53
.expected_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
58
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
59
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
61
{ .start = 0, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
64
.expected_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
65
.expected_len = 9ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
69
.hole_start = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
70
.hole_len = 9ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
71
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
77
.expected_len = 8ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
82
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
83
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
85
{ .start = SZ_1G, .len = SZ_1G },
fs/btrfs/tests/chunk-allocation-tests.c
89
.expected_len = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
94
.hole_len = 10ULL * SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
95
.min_hole_size = SZ_1G,
fs/btrfs/tests/chunk-allocation-tests.c
97
{ .start = SZ_2G, .len = SZ_1G },
fs/btrfs/tests/raid-stripe-tree-tests.c
1027
stripe->physical = logical + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
119
stripe->physical = logical3 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
253
stripe->physical = logical1 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
274
stripe->physical = logical2 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
295
stripe->physical = logical3 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
396
stripe->physical = logical1 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
536
stripe->physical = logical1 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
557
stripe->physical = logical2 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
676
stripe->physical = logical + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
77
stripe->physical = logical1 + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
785
stripe->physical = logical + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
907
stripe->physical = logical + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
953
stripe->physical = SZ_1G + logical + i * SZ_1G;
fs/btrfs/tests/raid-stripe-tree-tests.c
969
if (io_stripe.physical != logical + SZ_1G) {
fs/btrfs/tests/raid-stripe-tree-tests.c
971
logical + SZ_1G, io_stripe.physical);
fs/btrfs/tests/raid-stripe-tree-tests.c
98
stripe->physical = logical2 + i * SZ_1G;
fs/btrfs/volumes.c
5449
ctl->max_stripe_size = min_t(u64, ctl->max_chunk_size, SZ_1G);
fs/btrfs/volumes.c
5637
ctl->stripe_size = min_t(u64, ctl->stripe_size, SZ_1G);
fs/btrfs/volumes.h
35
#define BTRFS_MAX_DATA_CHUNK_SIZE (10ULL * SZ_1G)
fs/btrfs/volumes.h
41
#define BTRFS_MAX_DISCARD_CHUNK_SIZE (SZ_1G)
fs/btrfs/zoned.c
37
#define BTRFS_SB_LOG_FIRST_OFFSET (512ULL * SZ_1G)
fs/btrfs/zoned.c
38
#define BTRFS_SB_LOG_SECOND_OFFSET (4096ULL * SZ_1G)
include/linux/mm.h
2501
#define MAX_FOLIO_ORDER get_order(IS_ENABLED(CONFIG_64BIT) ? SZ_16G : SZ_1G)
io_uring/mock_file.c
240
if (mc.file_size > SZ_1G)
io_uring/rsrc.c
92
if (ulen > SZ_1G || !ulen)
kernel/dma/pool.c
202
unsigned long pages = totalram_pages() / (SZ_1G / SZ_128K);
mm/hugetlb_cgroup.c
665
if (hsize >= SZ_1G)
mm/hugetlb_cgroup.c
666
snprintf(buf, size, "%luGB", hsize / SZ_1G);
mm/util.c
396
return randomize_page(mm->brk, SZ_1G);
tools/testing/memblock/tests/basic_api.c
125
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
129
.base = SZ_1G + SZ_16K,
tools/testing/memblock/tests/basic_api.c
1341
.size = SZ_1G
tools/testing/memblock/tests/basic_api.c
179
.size = SZ_1G
tools/testing/memblock/tests/basic_api.c
235
.size = SZ_1G
tools/testing/memblock/tests/basic_api.c
2393
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
2407
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_1M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2408
ASSERT_FALSE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_4M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
2411
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_1M, SZ_2M));
tools/testing/memblock/tests/basic_api.c
2412
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_2M, SZ_2M));
tools/testing/memblock/tests/basic_api.c
2415
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G, SZ_4M));
tools/testing/memblock/tests/basic_api.c
2416
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G - SZ_2M, SZ_8M));
tools/testing/memblock/tests/basic_api.c
2417
ASSERT_TRUE(memblock_overlaps_region(&memblock.memory, SZ_1G + SZ_1M, SZ_1M));
tools/testing/memblock/tests/basic_api.c
350
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
354
.base = SZ_1G + SZ_16K,
tools/testing/memblock/tests/basic_api.c
358
.base = SZ_1G + SZ_8K,
tools/testing/memblock/tests/basic_api.c
48
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
643
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
644
.size = SZ_1G
tools/testing/memblock/tests/basic_api.c
648
.size = SZ_1G
tools/testing/memblock/tests/basic_api.c
819
.base = SZ_1G,
tools/testing/memblock/tests/basic_api.c
823
.base = SZ_1G + SZ_16K,
tools/testing/memblock/tests/basic_api.c
827
.base = SZ_1G + SZ_8K,
tools/testing/selftests/kvm/coalesced_io_test.c
221
.mmio_gpa = 4ull * SZ_1G,
tools/testing/selftests/kvm/coalesced_io_test.c
222
.mmio = (uint64_t *)(4ull * SZ_1G),
tools/testing/selftests/kvm/mmu_stress_test.c
304
max_mem = 128ull * SZ_1G;
tools/testing/selftests/kvm/mmu_stress_test.c
314
max_mem = 1ull * atoi_positive("Memory size", optarg) * SZ_1G;
tools/testing/selftests/kvm/mmu_stress_test.c
317
slot_size = 1ull * atoi_positive("Slot size", optarg) * SZ_1G;
tools/testing/selftests/kvm/mmu_stress_test.c
334
max_mem / SZ_1G,
tools/testing/selftests/kvm/mmu_stress_test.c
378
(gpa - start_gpa) / SZ_1G, nr_vcpus);
tools/testing/selftests/vfio/lib/vfio_pci_device.c
154
align = min_t(size_t, size, SZ_1G);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
126
size_t size = min_t(size_t, self->bar->info.size, SZ_1G);
tools/testing/selftests/vfio/vfio_dma_mapping_mmio_test.c
129
vaddr = mmap_reserve(size, SZ_1G, getpagesize());
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
116
FIXTURE_VARIANT_ADD_ALL_IOMMU_MODES(anonymous_hugetlb_1gb, SZ_1G, MAP_HUGETLB | MAP_HUGE_1GB);
tools/testing/selftests/vfio/vfio_dma_mapping_test.c
183
case SZ_1G:
tools/testing/selftests/vfio/vfio_pci_driver_test.c
223
total_size = 250UL * SZ_1G;
tools/testing/selftests/vfio/vfio_pci_driver_test.c
82
region_setup(self->iommu, self->iova_allocator, &self->memcpy_region, SZ_1G);
tools/testing/selftests/vfio/vfio_pci_driver_test.c
86
self->unmapped_iova = iova_allocator_alloc(self->iova_allocator, SZ_1G);