ivpu_bo_size
ivpu_bo_size(vdev->fw->mem_log_crit) + ivpu_bo_size(vdev->fw->mem_log_verb);
u64 size = ivpu_bo_size(fw->mem) - image_end_offset;
boot_params->ipc_header_area_size = ivpu_bo_size(ipc_mem_rx) / 2;
boot_params->ipc_payload_area_start = ipc_mem_rx->vpu_addr + ivpu_bo_size(ipc_mem_rx) / 2;
boot_params->ipc_payload_area_size = ivpu_bo_size(ipc_mem_rx) / 2;
boot_params->crit_tracing_buff_size = ivpu_bo_size(vdev->fw->mem_log_crit);
boot_params->verbose_tracing_buff_size = ivpu_bo_size(vdev->fw->mem_log_verb);
if ((*offset + sizeof(*log)) > ivpu_bo_size(bo))
if ((char *)log + log->size > (char *)ivpu_bo_vaddr(bo) + ivpu_bo_size(bo)) {
ret = ivpu_mmu_context_insert_node(ctx, range, ivpu_bo_size(bo), &bo->mm_node);
ivpu_dbg(vdev, BO, "import: bo %8p size %9zu\n", bo, ivpu_bo_size(bo));
drm_WARN_ON(&vdev->drm, ivpu_bo_size(bo) == 0);
action, bo, ivpu_bo_size(bo), bo->ctx_id, bo->vpu_addr,
ret = ivpu_mmu_context_map_sgt(vdev, bo->ctx, bo->vpu_addr, sgt, ivpu_bo_size(bo),
if (cpu_addr >= (ivpu_bo_vaddr(bo) + ivpu_bo_size(bo)))
if (vpu_addr >= (bo->vpu_addr + ivpu_bo_size(bo)))
ret = gen_pool_add(ipc->mm_tx, ipc->mem_tx->vpu_addr, ivpu_bo_size(ipc->mem_tx), -1);
memset(ivpu_bo_vaddr(ipc->mem_tx), 0, ivpu_bo_size(ipc->mem_tx));
memset(ivpu_bo_vaddr(ipc->mem_rx), 0, ivpu_bo_size(ipc->mem_rx));
size_t size = ivpu_bo_size(cmdq->mem) - sizeof(struct vpu_job_queue_header);
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
cmdq->mem->vpu_addr, ivpu_bo_size(cmdq->mem));
entry->primary_preempt_buf_size = ivpu_bo_size(job->primary_preempt_buf);
entry->secondary_preempt_buf_size = ivpu_bo_size(job->secondary_preempt_buf);
if (commands_offset >= ivpu_bo_size(bo)) {
commands_offset, ivpu_bo_size(bo));
if (ivpu_bo_size(preempt_bo) < ivpu_fw_preempt_buf_size(vdev)) {
args->max_data_size = ivpu_bo_size(ms->bo);
ivpu_bo_size(bo), NULL, &info_size);
ms->buff_size = ivpu_bo_size(ms->bo) / MS_NUM_BUFFERS;
total_npu_memory += ivpu_bo_size(bo);