root/drivers/gpu/drm/etnaviv/etnaviv_dump.c
// SPDX-License-Identifier: GPL-2.0
/*
 * Copyright (C) 2015-2018 Etnaviv Project
 */

#include <linux/devcoredump.h>
#include <linux/moduleparam.h>

#include "etnaviv_cmdbuf.h"
#include "etnaviv_dump.h"
#include "etnaviv_gem.h"
#include "etnaviv_gpu.h"
#include "etnaviv_mmu.h"
#include "etnaviv_sched.h"
#include "state.xml.h"
#include "state_hi.xml.h"

static bool etnaviv_dump_core = true;
module_param_named(dump_core, etnaviv_dump_core, bool, 0600);

struct core_dump_iterator {
        void *start;
        struct etnaviv_dump_object_header *hdr;
        void *data;
};

static const unsigned short etnaviv_dump_registers[] = {
        VIVS_HI_AXI_STATUS,
        VIVS_HI_CLOCK_CONTROL,
        VIVS_HI_IDLE_STATE,
        VIVS_HI_AXI_CONFIG,
        VIVS_HI_INTR_ENBL,
        VIVS_HI_CHIP_IDENTITY,
        VIVS_HI_CHIP_FEATURE,
        VIVS_HI_CHIP_MODEL,
        VIVS_HI_CHIP_REV,
        VIVS_HI_CHIP_DATE,
        VIVS_HI_CHIP_TIME,
        VIVS_HI_CHIP_MINOR_FEATURE_0,
        VIVS_HI_CACHE_CONTROL,
        VIVS_HI_AXI_CONTROL,
        VIVS_PM_POWER_CONTROLS,
        VIVS_PM_MODULE_CONTROLS,
        VIVS_PM_MODULE_STATUS,
        VIVS_PM_PULSE_EATER,
        VIVS_MC_MMU_FE_PAGE_TABLE,
        VIVS_MC_MMU_TX_PAGE_TABLE,
        VIVS_MC_MMU_PE_PAGE_TABLE,
        VIVS_MC_MMU_PEZ_PAGE_TABLE,
        VIVS_MC_MMU_RA_PAGE_TABLE,
        VIVS_MC_DEBUG_MEMORY,
        VIVS_MC_MEMORY_BASE_ADDR_RA,
        VIVS_MC_MEMORY_BASE_ADDR_FE,
        VIVS_MC_MEMORY_BASE_ADDR_TX,
        VIVS_MC_MEMORY_BASE_ADDR_PEZ,
        VIVS_MC_MEMORY_BASE_ADDR_PE,
        VIVS_MC_MEMORY_TIMING_CONTROL,
        VIVS_MC_BUS_CONFIG,
        VIVS_FE_DMA_STATUS,
        VIVS_FE_DMA_DEBUG_STATE,
        VIVS_FE_DMA_ADDRESS,
        VIVS_FE_DMA_LOW,
        VIVS_FE_DMA_HIGH,
        VIVS_FE_AUTO_FLUSH,
};

static void etnaviv_core_dump_header(struct core_dump_iterator *iter,
        u32 type, void *data_end)
{
        struct etnaviv_dump_object_header *hdr = iter->hdr;

        hdr->magic = cpu_to_le32(ETDUMP_MAGIC);
        hdr->type = cpu_to_le32(type);
        hdr->file_offset = cpu_to_le32(iter->data - iter->start);
        hdr->file_size = cpu_to_le32(data_end - iter->data);

        iter->hdr++;
        iter->data += le32_to_cpu(hdr->file_size);
}

static void etnaviv_core_dump_registers(struct core_dump_iterator *iter,
        struct etnaviv_gpu *gpu)
{
        struct etnaviv_dump_registers *reg = iter->data;
        unsigned int i;
        u32 read_addr;

        for (i = 0; i < ARRAY_SIZE(etnaviv_dump_registers); i++, reg++) {
                read_addr = etnaviv_dump_registers[i];
                if (read_addr >= VIVS_PM_POWER_CONTROLS &&
                    read_addr <= VIVS_PM_PULSE_EATER)
                        read_addr = gpu_fix_power_address(gpu, read_addr);
                reg->reg = cpu_to_le32(etnaviv_dump_registers[i]);
                reg->value = cpu_to_le32(gpu_read(gpu, read_addr));
        }

        etnaviv_core_dump_header(iter, ETDUMP_BUF_REG, reg);
}

static void etnaviv_core_dump_mmu(struct core_dump_iterator *iter,
        struct etnaviv_iommu_context *mmu, size_t mmu_size)
{
        etnaviv_iommu_dump(mmu, iter->data);

        etnaviv_core_dump_header(iter, ETDUMP_BUF_MMU, iter->data + mmu_size);
}

static void etnaviv_core_dump_mem(struct core_dump_iterator *iter, u32 type,
        void *ptr, size_t size, u64 iova)
{
        memcpy(iter->data, ptr, size);

        iter->hdr->iova = cpu_to_le64(iova);

        etnaviv_core_dump_header(iter, type, iter->data + size);
}

void etnaviv_core_dump(struct etnaviv_gem_submit *submit)
{
        struct etnaviv_gpu *gpu = submit->gpu;
        struct core_dump_iterator iter;
        struct etnaviv_gem_object *obj;
        unsigned int n_obj, n_bomap_pages;
        size_t file_size, mmu_size;
        __le64 *bomap, *bomap_start;
        int i;

        /* Only catch the first event, or when manually re-armed */
        if (!etnaviv_dump_core)
                return;
        etnaviv_dump_core = false;

        mutex_lock(&submit->mmu_context->lock);

        mmu_size = etnaviv_iommu_dump_size(submit->mmu_context);

        /* We always dump registers, mmu, ring, hanging cmdbuf and end marker */
        n_obj = 5;
        n_bomap_pages = 0;
        file_size = ARRAY_SIZE(etnaviv_dump_registers) *
                        sizeof(struct etnaviv_dump_registers) +
                    mmu_size + gpu->buffer.size + submit->cmdbuf.size;

        /* Add in the active buffer objects */
        for (i = 0; i < submit->nr_bos; i++) {
                obj = submit->bos[i].obj;
                file_size += obj->base.size;
                n_bomap_pages += obj->base.size >> PAGE_SHIFT;
                n_obj++;
        }

        /* If we have any buffer objects, add a bomap object */
        if (n_bomap_pages) {
                file_size += n_bomap_pages * sizeof(__le64);
                n_obj++;
        }

        /* Add the size of the headers */
        file_size += sizeof(*iter.hdr) * n_obj;

        /* Allocate the file in vmalloc memory, it's likely to be big */
        iter.start = __vmalloc(file_size, GFP_NOWAIT);
        if (!iter.start) {
                mutex_unlock(&submit->mmu_context->lock);
                dev_warn(gpu->dev, "failed to allocate devcoredump file\n");
                return;
        }

        /* Point the data member after the headers */
        iter.hdr = iter.start;
        iter.data = &iter.hdr[n_obj];

        memset(iter.hdr, 0, iter.data - iter.start);

        etnaviv_core_dump_registers(&iter, gpu);
        etnaviv_core_dump_mmu(&iter, submit->mmu_context, mmu_size);
        etnaviv_core_dump_mem(&iter, ETDUMP_BUF_RING, gpu->buffer.vaddr,
                              gpu->buffer.size,
                              etnaviv_cmdbuf_get_va(&gpu->buffer,
                                        &submit->mmu_context->cmdbuf_mapping));

        etnaviv_core_dump_mem(&iter, ETDUMP_BUF_CMD,
                              submit->cmdbuf.vaddr, submit->cmdbuf.size,
                              etnaviv_cmdbuf_get_va(&submit->cmdbuf,
                                        &submit->mmu_context->cmdbuf_mapping));

        mutex_unlock(&submit->mmu_context->lock);

        /* Reserve space for the bomap */
        if (n_bomap_pages) {
                bomap_start = bomap = iter.data;
                memset(bomap, 0, sizeof(*bomap) * n_bomap_pages);
                etnaviv_core_dump_header(&iter, ETDUMP_BUF_BOMAP,
                                         bomap + n_bomap_pages);
        } else {
                /* Silence warning */
                bomap_start = bomap = NULL;
        }

        for (i = 0; i < submit->nr_bos; i++) {
                struct etnaviv_vram_mapping *vram;
                struct page **pages;
                void *vaddr;

                obj = submit->bos[i].obj;
                vram = submit->bos[i].mapping;

                mutex_lock(&obj->lock);
                pages = etnaviv_gem_get_pages(obj);
                mutex_unlock(&obj->lock);
                if (!IS_ERR(pages)) {
                        int j;

                        iter.hdr->data[0] = cpu_to_le32((bomap - bomap_start));

                        for (j = 0; j < obj->base.size >> PAGE_SHIFT; j++)
                                *bomap++ = cpu_to_le64(page_to_phys(*pages++));
                }

                iter.hdr->iova = cpu_to_le64(vram->iova);

                vaddr = etnaviv_gem_vmap(&obj->base);
                if (vaddr)
                        memcpy(iter.data, vaddr, obj->base.size);

                etnaviv_core_dump_header(&iter, ETDUMP_BUF_BO, iter.data +
                                         obj->base.size);
        }

        etnaviv_core_dump_header(&iter, ETDUMP_BUF_END, iter.data);

        dev_coredumpv(gpu->dev, iter.start, iter.data - iter.start, GFP_NOWAIT);
}