xe_bo_size
size = xe_bo_size(bo);
xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo));
return xe_bo_size(node->bo);
offset = xe_device_ccs_bytes(tile_to_xe(tile), xe_bo_size(bo));
xe_bo_size(bo), SZ_4K));
xe_assert(xe, (buf - base + G2G_BUFFER_SIZE) <= xe_bo_size(guc->g2g.bo));
xe_map_memset(xe, &remote->vmap, 0, 0xd0, xe_bo_size(remote));
retval = xe_map_rd(xe, &remote->vmap, xe_bo_size(remote) - 8, u64);
xe_map_memset(xe, &remote->vmap, 0, 0xc0, xe_bo_size(remote));
xe_map_memset(xe, &bo->vmap, 0, 0xd0, xe_bo_size(bo));
retval = xe_map_rd(xe, &bo->vmap, xe_bo_size(bo) - 8, u64);
xe_map_memset(xe, &remote->vmap, 0, 0xd0, xe_bo_size(remote));
xe_map_memset(xe, &bo->vmap, 0, 0xc0, xe_bo_size(bo));
retval = xe_map_rd(xe, &remote->vmap, xe_bo_size(bo) - 8, u64);
xe_res_first(pt->ttm.resource, 0, xe_bo_size(pt), &src_it);
xe_res_first_sg(xe_bo_sg(pt), 0, xe_bo_size(pt), &src_it);
xe_map_memset(xe, &tiny->vmap, 0, 0x22, xe_bo_size(tiny));
retval = xe_map_rd(xe, &tiny->vmap, xe_bo_size(tiny) - 4, u32);
xe_map_memset(xe, &big->vmap, 0, 0x11, xe_bo_size(big));
retval = xe_map_rd(xe, &big->vmap, xe_bo_size(big) - 4, u32);
u64 size = xe_bo_size(src_bo);
xe_map_memset(xe, &sys_bo->vmap, 0, 0xd0, xe_bo_size(sys_bo));
retval = xe_map_rd(xe, &vram_bo->vmap, xe_bo_size(vram_bo) - 8, u64);
retval = xe_map_rd(xe, &ccs_bo->vmap, xe_bo_size(ccs_bo) - 8, u64);
retval = xe_map_rd(xe, &vram_bo->vmap, xe_bo_size(vram_bo) - 8, u64);
retval = xe_map_rd(xe, &ccs_bo->vmap, xe_bo_size(ccs_bo) - 8, u64);
xe_map_memset(xe, &sys_bo->vmap, 0, 0xd0, xe_bo_size(sys_bo));
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
retval = xe_map_rd(xe, &sys_bo->vmap, xe_bo_size(sys_bo) - 8, u64);
bool big = xe_bo_size(bo) >= SZ_2M;
xe_bo_size(bo),
backup = xe_bo_init_locked(xe, NULL, NULL, bo->ttm.base.resv, NULL, xe_bo_size(bo),
xe_bo_size(bo));
xe_bo_size(bo),
xe_bo_size(bo));
xe_bo_size(bo) - (offset & PAGE_MASK), &cursor);
start + xe_bo_size(bo), U64_MAX,
xe_bo_size(*src), dst_flags);
ret = ttm_bo_kmap(&bo->ttm, 0, xe_bo_size(bo) >> PAGE_SHIFT, &bo->kmap);
extra_pages = DIV_ROUND_UP(xe_device_ccs_bytes(xe, xe_bo_size(bo)),
XE_WARN_ON(offset + xe_bo_size(bo) > (1ull << 32));
return PAGE_ALIGN(xe_bo_size(bo));
u64 sz = xe_bo_size(bo);
xe_map_memset(xe, &ggtt->scratch->vmap, 0, 0, xe_bo_size(ggtt->scratch));
end = start + xe_bo_size(bo);
for (xe_res_first_sg(xe_bo_sg(bo), 0, xe_bo_size(bo), &cur);
for (xe_res_first(bo->ttm.resource, 0, xe_bo_size(bo), &cur);
xe_tile_assert(ggtt->tile, bo->ggtt_node[tile_id]->base.size == xe_bo_size(bo));
xe_bo_size(bo), alignment, 0, start, end, 0);
xe_tile_assert(ggtt->tile, bo->ggtt_node[tile_id]->base.size == xe_bo_size(bo));
xe_bo_size(gsc->private) - fw_size);
bb->cs[bb->len++] = (xe_bo_size(gsc->private) / SZ_4K) |
return bo ? xe_bo_size(bo) : 0;
offset += xe_bo_size(bo);
err = pf_push_vf_cfg_lmem(gt, vfid, xe_bo_size(bo));
vfid, xe_bo_size(bo), xe_bo_size(bo) / SZ_1M);
string_get_size(xe_bo_size(config->lmem_obj), 1, STRING_UNITS_2,
n, xe_bo_size(config->lmem_obj), buf);
cfg[n++] = lower_32_bits(xe_bo_size(config->lmem_obj));
cfg[n++] = upper_32_bits(xe_bo_size(config->lmem_obj));
vram_size = xe_bo_size(vram);
size = xe_bo_size(vram);
xe_assert(xe, (buf - base + G2G_BUFFER_SIZE) <= xe_bo_size(g2g_bo));
xe_assert(xe, xe_bo_size(bo) <= GUC_GGTT_TOP - addr);
xe_map_memset(ads_to_xe(ads), ads_to_map(ads), 0, 0, xe_bo_size(ads->bo));
xe_map_memset(ads_to_xe(ads), ads_to_map(ads), 0, 0, xe_bo_size(ads->bo));
snapshot->ctb_size = xe_bo_size(ct->bo);
xe_map_memset(xe, &ct->bo->vmap, 0, 0, xe_bo_size(ct->bo));
snapshot->size = xe_bo_size(log->bo);
xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo));
xe_bo_size(huc->fw.bo));
lmtt_assert(lmtt, IS_ALIGNED(xe_bo_size(bo), page_size));
xe_res_first(bo->ttm.resource, 0, xe_bo_size(bo), &cur);
xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo));
u32 offset = xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE -
return xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE - LRC_INDIRECT_CTX_BO_SIZE;
return xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE;
u64 size = xe_bo_size(src_bo);
size = xe_bo_size(src_bo);
u64 size = xe_bo_size(bo);
xe_res_first_sg(xe_bo_sg(bo), 0, xe_bo_size(bo), &src_it);
xe_res_first(src, 0, xe_bo_size(bo), &src_it);
xe_tile_assert(tile, xe_bo_size(pt_bo) == SZ_4K);
xe_tile_assert(tile, m->batch_base_ofs + xe_bo_size(batch) < SZ_2M);
pt29_ofs = xe_bo_size(bo) - 3 * XE_PAGE_SIZE;
for (i = 0; i < xe_bo_size(batch);
xe_res_first(bo->ttm.resource, offset, xe_bo_size(bo) - offset, &cursor);
xe_tile_assert(tile, xe_bo_size(batch) == SZ_1M);
xe_tile_assert(tile, xe_bo_size(batch) == SZ_512K);
for (i = 0; i < xe_bo_size(batch);
u64 pt30_ofs = xe_bo_size(bo) - 2 * XE_PAGE_SIZE;
u64 pt31_ofs = xe_bo_size(bo) - XE_PAGE_SIZE;
u64 size = xe_bo_size(src_bo);
if (src_bo != dst_bo && XE_WARN_ON(xe_bo_size(src_bo) != xe_bo_size(dst_bo)))
xe_bo_size(stream->oa_buffer.bo) > SZ_16M ?
struct drm_xe_oa_stream_info info = { .oa_buf_size = xe_bo_size(stream->oa_buffer.bo), };
if (vma->vm_end - vma->vm_start != xe_bo_size(stream->oa_buffer.bo)) {
int size_exponent = __ffs(xe_bo_size(stream->oa_buffer.bo));
memset(stream->oa_buffer.vaddr, 0, xe_bo_size(stream->oa_buffer.bo));
size, id, bo ? xe_bo_size(bo) : 0);
*val = xe_bo_size(bo);
__entry->size = xe_bo_size(bo);
__entry->size = xe_bo_size(bo);
if (XE_IOCTL_DBG(xe, range > xe_bo_size(bo)) ||
xe_bo_size(bo) - range)) {
ops = vm_bind_ioctl_ops_create(vm, &vops, bo, 0, addr, xe_bo_size(bo),