root/drivers/gpu/drm/xe/xe_hw_error.c
// SPDX-License-Identifier: MIT
/*
 * Copyright © 2025 Intel Corporation
 */

#include <linux/fault-inject.h>

#include "regs/xe_gsc_regs.h"
#include "regs/xe_hw_error_regs.h"
#include "regs/xe_irq_regs.h"

#include "xe_device.h"
#include "xe_hw_error.h"
#include "xe_mmio.h"
#include "xe_survivability_mode.h"

#define  HEC_UNCORR_FW_ERR_BITS 4
extern struct fault_attr inject_csc_hw_error;

/* Error categories reported by hardware */
enum hardware_error {
        HARDWARE_ERROR_CORRECTABLE = 0,
        HARDWARE_ERROR_NONFATAL = 1,
        HARDWARE_ERROR_FATAL = 2,
        HARDWARE_ERROR_MAX,
};

static const char * const hec_uncorrected_fw_errors[] = {
        "Fatal",
        "CSE Disabled",
        "FD Corruption",
        "Data Corruption"
};

static const char *hw_error_to_str(const enum hardware_error hw_err)
{
        switch (hw_err) {
        case HARDWARE_ERROR_CORRECTABLE:
                return "CORRECTABLE";
        case HARDWARE_ERROR_NONFATAL:
                return "NONFATAL";
        case HARDWARE_ERROR_FATAL:
                return "FATAL";
        default:
                return "UNKNOWN";
        }
}

static bool fault_inject_csc_hw_error(void)
{
        return IS_ENABLED(CONFIG_DEBUG_FS) && should_fail(&inject_csc_hw_error, 1);
}

static void csc_hw_error_work(struct work_struct *work)
{
        struct xe_tile *tile = container_of(work, typeof(*tile), csc_hw_error_work);
        struct xe_device *xe = tile_to_xe(tile);
        int ret;

        ret = xe_survivability_mode_runtime_enable(xe);
        if (ret)
                drm_err(&xe->drm, "Failed to enable runtime survivability mode\n");
}

static void csc_hw_error_handler(struct xe_tile *tile, const enum hardware_error hw_err)
{
        const char *hw_err_str = hw_error_to_str(hw_err);
        struct xe_device *xe = tile_to_xe(tile);
        struct xe_mmio *mmio = &tile->mmio;
        u32 base, err_bit, err_src;
        unsigned long fw_err;

        if (xe->info.platform != XE_BATTLEMAGE)
                return;

        base = BMG_GSC_HECI1_BASE;
        lockdep_assert_held(&xe->irq.lock);
        err_src = xe_mmio_read32(mmio, HEC_UNCORR_ERR_STATUS(base));
        if (!err_src) {
                drm_err_ratelimited(&xe->drm, HW_ERR "Tile%d reported HEC_ERR_STATUS_%s blank\n",
                                    tile->id, hw_err_str);
                return;
        }

        if (err_src & UNCORR_FW_REPORTED_ERR) {
                fw_err = xe_mmio_read32(mmio, HEC_UNCORR_FW_ERR_DW0(base));
                for_each_set_bit(err_bit, &fw_err, HEC_UNCORR_FW_ERR_BITS) {
                        drm_err_ratelimited(&xe->drm, HW_ERR
                                            "%s: HEC Uncorrected FW %s error reported, bit[%d] is set\n",
                                             hw_err_str, hec_uncorrected_fw_errors[err_bit],
                                             err_bit);

                        schedule_work(&tile->csc_hw_error_work);
                }
        }

        xe_mmio_write32(mmio, HEC_UNCORR_ERR_STATUS(base), err_src);
}

static void hw_error_source_handler(struct xe_tile *tile, const enum hardware_error hw_err)
{
        const char *hw_err_str = hw_error_to_str(hw_err);
        struct xe_device *xe = tile_to_xe(tile);
        unsigned long flags;
        u32 err_src;

        if (xe->info.platform != XE_BATTLEMAGE)
                return;

        spin_lock_irqsave(&xe->irq.lock, flags);
        err_src = xe_mmio_read32(&tile->mmio, DEV_ERR_STAT_REG(hw_err));
        if (!err_src) {
                drm_err_ratelimited(&xe->drm, HW_ERR "Tile%d reported DEV_ERR_STAT_%s blank!\n",
                                    tile->id, hw_err_str);
                goto unlock;
        }

        if (err_src & XE_CSC_ERROR)
                csc_hw_error_handler(tile, hw_err);

        xe_mmio_write32(&tile->mmio, DEV_ERR_STAT_REG(hw_err), err_src);

unlock:
        spin_unlock_irqrestore(&xe->irq.lock, flags);
}

/**
 * xe_hw_error_irq_handler - irq handling for hw errors
 * @tile: tile instance
 * @master_ctl: value read from master interrupt register
 *
 * Xe platforms add three error bits to the master interrupt register to support error handling.
 * These three bits are used to convey the class of error FATAL, NONFATAL, or CORRECTABLE.
 * To process the interrupt, determine the source of error by reading the Device Error Source
 * Register that corresponds to the class of error being serviced.
 */
void xe_hw_error_irq_handler(struct xe_tile *tile, const u32 master_ctl)
{
        enum hardware_error hw_err;

        if (fault_inject_csc_hw_error())
                schedule_work(&tile->csc_hw_error_work);

        for (hw_err = 0; hw_err < HARDWARE_ERROR_MAX; hw_err++)
                if (master_ctl & ERROR_IRQ(hw_err))
                        hw_error_source_handler(tile, hw_err);
}

/*
 * Process hardware errors during boot
 */
static void process_hw_errors(struct xe_device *xe)
{
        struct xe_tile *tile;
        u32 master_ctl;
        u8 id;

        for_each_tile(tile, xe, id) {
                master_ctl = xe_mmio_read32(&tile->mmio, GFX_MSTR_IRQ);
                xe_hw_error_irq_handler(tile, master_ctl);
                xe_mmio_write32(&tile->mmio, GFX_MSTR_IRQ, master_ctl);
        }
}

/**
 * xe_hw_error_init - Initialize hw errors
 * @xe: xe device instance
 *
 * Initialize and check for errors that occurred during boot
 * prior to driver load
 */
void xe_hw_error_init(struct xe_device *xe)
{
        struct xe_tile *tile = xe_device_get_root_tile(xe);

        if (!IS_DGFX(xe) || IS_SRIOV_VF(xe))
                return;

        INIT_WORK(&tile->csc_hw_error_work, csc_hw_error_work);

        process_hw_errors(xe);
}