root/drivers/gpu/drm/amd/ras/ras_mgr/amdgpu_ras_sys.c
// SPDX-License-Identifier: MIT
/*
 * Copyright 2025 Advanced Micro Devices, Inc.
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice shall be included in
 * all copies or substantial portions of the Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR
 * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
 * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
 * OTHER DEALINGS IN THE SOFTWARE.
 *
 */
#include "ras_sys.h"
#include "amdgpu_ras_mgr.h"
#include "amdgpu_ras.h"
#include "amdgpu_reset.h"

static int amdgpu_ras_sys_detect_fatal_event(struct ras_core_context *ras_core, void *data)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        int ret;
        uint64_t seq_no;

        ret = amdgpu_ras_global_ras_isr(adev);
        if (ret)
                return ret;

        seq_no = amdgpu_ras_mgr_gen_ras_event_seqno(adev, RAS_SEQNO_TYPE_UE);
        RAS_DEV_INFO(adev,
                "{%llu} Uncorrectable hardware error(ERREVENT_ATHUB_INTERRUPT) detected!\n",
                seq_no);

        return amdgpu_ras_process_handle_unexpected_interrupt(adev, data);
}

static int amdgpu_ras_sys_poison_consumption_event(struct ras_core_context *ras_core,
                                void *data)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        struct ras_event_req *req = (struct ras_event_req *)data;
        pasid_notify pasid_fn;

        if (!req)
                return -EINVAL;

        if (req->pasid_fn) {
                pasid_fn = (pasid_notify)req->pasid_fn;
                pasid_fn(adev, req->pasid, req->data);
        }

        return 0;
}

static int amdgpu_ras_sys_gen_seqno(struct ras_core_context *ras_core,
                        enum ras_seqno_type seqno_type, uint64_t *seqno)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        struct amdgpu_ras_mgr *ras_mgr = amdgpu_ras_mgr_get_context(adev);
        struct ras_event_manager *event_mgr;
        struct ras_event_state *event_state;
        struct amdgpu_hive_info *hive;
        enum ras_event_type event_type;
        uint64_t seq_no;

        if (!ras_mgr || !seqno ||
                (seqno_type >= RAS_SEQNO_TYPE_COUNT_MAX))
                return -EINVAL;

        switch (seqno_type) {
        case RAS_SEQNO_TYPE_UE:
                event_type = RAS_EVENT_TYPE_FATAL;
                break;
        case RAS_SEQNO_TYPE_CE:
        case RAS_SEQNO_TYPE_DE:
                event_type = RAS_EVENT_TYPE_POISON_CREATION;
                break;
        case RAS_SEQNO_TYPE_POISON_CONSUMPTION:
                event_type = RAS_EVENT_TYPE_POISON_CONSUMPTION;
                break;
        default:
                event_type = RAS_EVENT_TYPE_INVALID;
                break;
        }

        hive = amdgpu_get_xgmi_hive(adev);
        event_mgr = hive ? &hive->event_mgr : &ras_mgr->ras_event_mgr;
        event_state = &event_mgr->event_state[event_type];
        if ((event_type == RAS_EVENT_TYPE_FATAL) && amdgpu_ras_in_recovery(adev)) {
                seq_no = event_state->last_seqno;
        } else {
                seq_no = atomic64_inc_return(&event_mgr->seqno);
                event_state->last_seqno = seq_no;
                atomic64_inc(&event_state->count);
        }
        amdgpu_put_xgmi_hive(hive);

        *seqno = seq_no;
        return 0;

}

static int amdgpu_ras_sys_event_notifier(struct ras_core_context *ras_core,
                                   enum ras_notify_event event_id, void *data)
{
        struct amdgpu_ras_mgr *ras_mgr = amdgpu_ras_mgr_get_context(ras_core->dev);
        int ret = 0;

        switch (event_id) {
        case RAS_EVENT_ID__BAD_PAGE_DETECTED:
                schedule_delayed_work(&ras_mgr->retire_page_dwork, 0);
                break;
        case RAS_EVENT_ID__POISON_CONSUMPTION:
                amdgpu_ras_sys_poison_consumption_event(ras_core, data);
                break;
        case RAS_EVENT_ID__RESERVE_BAD_PAGE:
                ret = amdgpu_ras_reserve_page(ras_core->dev, *(uint64_t *)data);
                break;
        case RAS_EVENT_ID__FATAL_ERROR_DETECTED:
                ret = amdgpu_ras_sys_detect_fatal_event(ras_core, data);
                break;
        case RAS_EVENT_ID__UPDATE_BAD_PAGE_NUM:
                ret = amdgpu_dpm_send_hbm_bad_pages_num(ras_core->dev, *(uint32_t *)data);
                break;
        case RAS_EVENT_ID__UPDATE_BAD_CHANNEL_BITMAP:
                ret = amdgpu_dpm_send_hbm_bad_channel_flag(ras_core->dev, *(uint32_t *)data);
                break;
        case RAS_EVENT_ID__DEVICE_RMA:
                ras_log_ring_add_log_event(ras_core, RAS_LOG_EVENT_RMA, NULL, NULL);
                ret = amdgpu_dpm_send_rma_reason(ras_core->dev);
                break;
        case RAS_EVENT_ID__RESET_GPU:
                ret = amdgpu_ras_mgr_reset_gpu(ras_core->dev, *(uint32_t *)data);
                break;
        case RAS_EVENT_ID__RAS_EVENT_PROC_BEGIN:
                ret = amdgpu_ras_process_begin(ras_core->dev);
                break;
        case RAS_EVENT_ID__RAS_EVENT_PROC_END:
                ret = amdgpu_ras_process_end(ras_core->dev);
                break;
        default:
                RAS_DEV_WARN(ras_core->dev, "Invalid ras notify event:%d\n", event_id);
                break;
        }

        return ret;
}

static u64 amdgpu_ras_sys_get_utc_second_timestamp(struct ras_core_context *ras_core)
{
        return ktime_get_real_seconds();
}

static int amdgpu_ras_sys_check_gpu_status(struct ras_core_context *ras_core,
                                uint32_t *status)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        uint32_t gpu_status = 0;

        if (amdgpu_in_reset(adev) || amdgpu_ras_in_recovery(adev))
                gpu_status |= RAS_GPU_STATUS__IN_RESET;

        if (amdgpu_sriov_vf(adev))
                gpu_status |= RAS_GPU_STATUS__IS_VF;

        *status = gpu_status;

        return 0;
}

static int amdgpu_ras_sys_get_device_system_info(struct ras_core_context *ras_core,
                        struct device_system_info *dev_info)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;

        dev_info->device_id = adev->pdev->device;
        dev_info->vendor_id = adev->pdev->vendor;
        dev_info->socket_id = adev->smuio.funcs->get_socket_id(adev);

        return 0;
}

static int amdgpu_ras_sys_gpu_reset_lock(struct ras_core_context *ras_core,
                        bool down, bool try)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        int ret = 0;

        if (down && try)
                ret = down_read_trylock(&adev->reset_domain->sem);
        else if (down)
                down_read(&adev->reset_domain->sem);
        else
                up_read(&adev->reset_domain->sem);

        return ret;
}

static bool amdgpu_ras_sys_detect_ras_interrupt(struct ras_core_context *ras_core)
{
        return !!atomic_read(&amdgpu_ras_in_intr);
}

static int amdgpu_ras_sys_get_gpu_mem(struct ras_core_context *ras_core,
        enum gpu_mem_type mem_type, struct gpu_mem_block *gpu_mem)
{
        struct amdgpu_device *adev = (struct amdgpu_device *)ras_core->dev;
        struct psp_context *psp = &adev->psp;
        struct psp_ring *psp_ring;
        struct ta_mem_context *mem_ctx;

        if (mem_type == GPU_MEM_TYPE_RAS_PSP_RING) {
                psp_ring = &psp->km_ring;
                gpu_mem->mem_bo = adev->firmware.rbuf;
                gpu_mem->mem_size = psp_ring->ring_size;
                gpu_mem->mem_mc_addr = psp_ring->ring_mem_mc_addr;
                gpu_mem->mem_cpu_addr = psp_ring->ring_mem;
        } else if (mem_type == GPU_MEM_TYPE_RAS_PSP_CMD) {
                gpu_mem->mem_bo = psp->cmd_buf_bo;
                gpu_mem->mem_size = PSP_CMD_BUFFER_SIZE;
                gpu_mem->mem_mc_addr = psp->cmd_buf_mc_addr;
                gpu_mem->mem_cpu_addr = psp->cmd_buf_mem;
        } else if (mem_type == GPU_MEM_TYPE_RAS_PSP_FENCE) {
                gpu_mem->mem_bo = psp->fence_buf_bo;
                gpu_mem->mem_size = PSP_FENCE_BUFFER_SIZE;
                gpu_mem->mem_mc_addr = psp->fence_buf_mc_addr;
                gpu_mem->mem_cpu_addr = psp->fence_buf;
        } else if (mem_type == GPU_MEM_TYPE_RAS_TA_FW) {
                gpu_mem->mem_bo = psp->fw_pri_bo;
                gpu_mem->mem_size = PSP_1_MEG;
                gpu_mem->mem_mc_addr = psp->fw_pri_mc_addr;
                gpu_mem->mem_cpu_addr = psp->fw_pri_buf;
        } else if (mem_type == GPU_MEM_TYPE_RAS_TA_CMD) {
                mem_ctx = &psp->ras_context.context.mem_context;
                gpu_mem->mem_bo = mem_ctx->shared_bo;
                gpu_mem->mem_size = mem_ctx->shared_mem_size;
                gpu_mem->mem_mc_addr = mem_ctx->shared_mc_addr;
                gpu_mem->mem_cpu_addr = mem_ctx->shared_buf;
        } else {
                return -EINVAL;
        }

        if (!gpu_mem->mem_bo || !gpu_mem->mem_size ||
                !gpu_mem->mem_mc_addr || !gpu_mem->mem_cpu_addr) {
                RAS_DEV_ERR(ras_core->dev, "The ras psp gpu memory is invalid!\n");
                return -ENOMEM;
        }

        return 0;
}

static int amdgpu_ras_sys_put_gpu_mem(struct ras_core_context *ras_core,
        enum gpu_mem_type mem_type, struct gpu_mem_block *gpu_mem)
{

        return 0;
}

const struct ras_sys_func amdgpu_ras_sys_fn = {
        .ras_notifier = amdgpu_ras_sys_event_notifier,
        .get_utc_second_timestamp = amdgpu_ras_sys_get_utc_second_timestamp,
        .gen_seqno = amdgpu_ras_sys_gen_seqno,
        .check_gpu_status = amdgpu_ras_sys_check_gpu_status,
        .get_device_system_info = amdgpu_ras_sys_get_device_system_info,
        .gpu_reset_lock = amdgpu_ras_sys_gpu_reset_lock,
        .detect_ras_interrupt = amdgpu_ras_sys_detect_ras_interrupt,
        .get_gpu_mem = amdgpu_ras_sys_get_gpu_mem,
        .put_gpu_mem = amdgpu_ras_sys_put_gpu_mem,
};