amdgpu_bo_size
*bo_size = amdgpu_bo_size(bo);
amdgpu_bo_size(entry->bo_va->base.bo),
*size = amdgpu_bo_size(bo);
*size = amdgpu_bo_size(bo);
total_size += amdgpu_bo_size(bo);
failed_size += amdgpu_bo_size(bo);
u64 size = amdgpu_bo_size(bo);
total_size += amdgpu_bo_size(bo);
size = amdgpu_bo_size(p->uf_bo);
bo_entry->size = amdgpu_bo_size(bo);
addr, amdgpu_bo_size(bo),
amdgpu_bo_size(bo),
job->gds_size = amdgpu_bo_size(gds) >> PAGE_SHIFT;
job->gws_size = amdgpu_bo_size(gws) >> PAGE_SHIFT;
job->oa_size = amdgpu_bo_size(oa) >> PAGE_SHIFT;
atomic64_sub(amdgpu_bo_size(bo), &adev->vram_pin_size);
atomic64_sub(amdgpu_bo_size(bo), &adev->gart_pin_size);
size = amdgpu_bo_size(bo);
atomic64_add(amdgpu_bo_size(bo), &adev->vram_pin_size);
atomic64_add(amdgpu_bo_size(bo), &adev->gart_pin_size);
size = amdgpu_bo_size(tmr_bo);
amdgpu_bo_size(psp->tmr_bo), psp->tmr_mc_addr);
__entry->bo_size = amdgpu_bo_size(bo);
__entry->bo_size = amdgpu_bo_size(bo);
amdgpu_res_first(bo->tbo.resource, 0, amdgpu_bo_size(bo), &cursor);
amdgpu_res_first(bo->tbo.resource, 0, amdgpu_bo_size(bo), &dst);
amdgpu_bo_size(db_obj->obj)) {
size = amdgpu_bo_size(adev->uvd.inst[j].vcpu_bo);
size = amdgpu_bo_size(adev->uvd.inst[i].vcpu_bo);
memset_io(adev->vce.cpu_addr, 0, amdgpu_bo_size(adev->vce.vcpu_bo));
size = amdgpu_bo_size(adev->vcn.inst[i].vcpu_bo);
size = amdgpu_bo_size(adev->vcn.inst[i].vcpu_bo);
if (bo && offset + size > amdgpu_bo_size(bo))
uint64_t size = amdgpu_bo_size(bo);
int64_t size = sign * amdgpu_bo_size(bo);
entries = amdgpu_bo_size(bo) / 8;
return amdgpu_bo_size(bo);
u64 bo_size = amdgpu_bo_size(adev->vce.vcpu_bo);
unsigned size = amdgpu_bo_size(adev->vce.vcpu_bo);
unsigned size = amdgpu_bo_size(adev->vce.vcpu_bo);
unsigned size = amdgpu_bo_size(adev->vce.vcpu_bo);
bo_bucket->size = amdgpu_bo_size(dumper_bo);
atomic64_sub(amdgpu_bo_size(svm_bo->bo), &pdd->vram_usage);
atomic64_add(amdgpu_bo_size(bo), &pdd->vram_usage);