SZ_512K
.size = 3 * SZ_1M + SZ_512K },
.offset = 3 * SZ_1M + SZ_512K,
.offset = 3 * SZ_1M + SZ_512K + SZ_256K,
#define MV2120_NOR_BOOT_SIZE SZ_512K
#define NET2BIG_NOR_BOOT_SIZE SZ_512K
#define INTEGRATOR_BOOT_ROM_SIZE SZ_512K
#define INTEGRATOR_MBRD_SSRAM_SIZE SZ_512K
ret = l2x0_cache_size_of_parse(np, aux_val, aux_mask, &assoc, SZ_512K);
if (imm < -SZ_512K || imm >= SZ_512K) {
if (imm < -SZ_512K || imm >= SZ_512K) {
if (arch_vmap_try_size(addr, end, pfn, max_page_shift, SZ_512K))
return SZ_512K;
if (size >= SZ_512K)
return SZ_512K;
return SZ_512K / SZ_4K;
if (k_cur < ALIGN_DOWN(k_end, SZ_512K))
for (; p < ALIGN(p, SZ_512K) && p < top && !err; p += SZ_16K, v += SZ_16K)
for (; p < ALIGN(p, SZ_8M) && p < top && !err; p += SZ_512K, v += SZ_512K)
for (; p < ALIGN_DOWN(top, SZ_512K) && p < top && !err; p += SZ_512K, v += SZ_512K)
.size = SZ_512K,
.size = SZ_512K,
CACHE_ENTRY(0x22, CACHE_L3, SZ_512K ), /* 4-way set assoc, sectored cache, 64 byte line size */
CACHE_ENTRY(0x3e, CACHE_L2, SZ_512K ), /* 4-way set assoc, sectored cache, 64 byte line size */
CACHE_ENTRY(0x43, CACHE_L2, SZ_512K ), /* 4-way set assoc, 32 byte line size */
CACHE_ENTRY(0x7b, CACHE_L2, SZ_512K ), /* 8-way set assoc, sectored cache, 64 byte line size */
CACHE_ENTRY(0x7f, CACHE_L2, SZ_512K ), /* 2-way set assoc, 64 byte line size */
CACHE_ENTRY(0x80, CACHE_L2, SZ_512K ), /* 8-way set assoc, 64 byte line size */
CACHE_ENTRY(0x83, CACHE_L2, SZ_512K ), /* 8-way set assoc, 32 byte line size */
CACHE_ENTRY(0x86, CACHE_L2, SZ_512K ), /* 4-way set assoc, 64 byte line size */
CACHE_ENTRY(0xd0, CACHE_L3, SZ_512K ), /* 4-way set assoc, 64 byte line size */
#define IVPU_FW_CRITICAL_BUFFER_SIZE SZ_512K
#define FW_SHARED_MEM_ALIGNMENT SZ_512K /* VPU MTRR limitation */
mhi_cntrl->seg_len = SZ_512K;
pci_resource_len(pdev, bar) < SZ_512K ||
#define IOAT_CHUNK_SIZE (SZ_512K)
const size_t sz = SZ_512K;
#define SVC_BUF_SIZE SZ_512K
#define SUBALLOC_SIZE SZ_512K
ce->ring_size = SZ_512K;
return intel_engine_create_pinned_context(engine, engine->gt->vm, SZ_512K,
ce = intel_engine_create_pinned_context(engine, vm, SZ_512K,
err = create_watcher(&watcher[1], engine, SZ_512K);
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = (SZ_1M + SZ_512K),
.gmem = (SZ_1M + SZ_512K),
.gmem = SZ_1M + SZ_512K,
.gmem = SZ_1M + SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_1M + SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
.gmem = SZ_512K,
u32 small = SZ_8K, medium = SZ_512K,
u32 size = SZ_512K;
static size_t per_xecore_buf_size = SZ_512K;
IS_DGFX(xe) ? SZ_1M : SZ_512K, 16);
xe_tile_assert(tile, xe_bo_size(batch) == SZ_512K);
SZ_512K | SZ_1M | SZ_2M)
#define MAX_FW_SIZE_V6 (SZ_512K) /* 512KB */
#define MAX_FW_SIZE_V7 (SZ_512K) /* 512KB */
#define MAX_FW_SIZE_V8 (SZ_512K) /* 512KB */
SZ_128K / 512, SZ_256K / 512, SZ_512K / 512, SZ_1M / 512,
memorg->pages_per_eraseblock = (SZ_512K + SZ_256K) /
mtd->erasesize = SZ_512K + SZ_256K;
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.sector_size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.sector_size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
.size = SZ_512K,
mhi_ctrl->sbl_size = SZ_512K;
mhi_ctrl->seg_len = SZ_512K;
mhi_ctrl->sbl_size = SZ_512K;
mhi_ctrl->seg_len = SZ_512K;
pcie_ecam->dbi_base = cfg->win + SZ_512K;
ispi->chip0_size = SZ_512K;
nvm->active_size = SZ_512K;
u64 num_chunks = DIV_ROUND_UP(end - start, SZ_512K);
u64 cur_end = min(end, start + SZ_512K - 1);
btrfs_find_first_clear_extent_bit(&tree, SZ_512K, &start, &end,
ret = btrfs_remove_free_space(cache, SZ_512K, 3 * SZ_1M);
if (test_check_exists(cache, SZ_512K, 3 * SZ_1M)) {
ret = test_add_free_space_entry(cache, SZ_128M + SZ_512K,
SZ_128M - SZ_512K, 1);
if (!test_check_exists(cache, SZ_128M + SZ_512K, SZ_256K)) {
ret = btrfs_add_free_space(cache, SZ_128M, SZ_512K);
if (!test_check_exists(cache, SZ_128M, SZ_512K)) {
ret = test_add_free_space_entry(cache, 0, SZ_128M - SZ_512K, 1);
if (test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) {
ret = btrfs_add_free_space(cache, SZ_128M - SZ_512K, SZ_512K);
if (!test_check_exists(cache, SZ_128M - SZ_512K, SZ_512K)) {
ret = btrfs_delete_raid_extent(trans, logical1 + SZ_512K, SZ_1M);
logical1 + SZ_512K, (u64)SZ_1M);
if (len != SZ_512K) {
(u64)SZ_512K, len);
ret = btrfs_get_raid_extent_offset(fs_info, logical2 + SZ_512K, &len,
logical2 + SZ_512K, logical2 + len);
if (io_stripe.physical != logical2 + SZ_512K) {
logical2 + SZ_512K, io_stripe.physical);
if (len != SZ_512K) {
(u64)SZ_512K, len);
ret = btrfs_get_raid_extent_offset(fs_info, logical1 + SZ_512K, &len,
logical1 + SZ_512K, logical1 + SZ_512K + len);
ret = btrfs_delete_raid_extent(trans, logical1, SZ_512K);
ret = btrfs_delete_raid_extent(trans, logical2 + SZ_512K, SZ_512K);
.base = SZ_512K,