root/arch/loongarch/kernel/acpi.c
// SPDX-License-Identifier: GPL-2.0
/*
 * acpi.c - Architecture-Specific Low-Level ACPI Boot Support
 *
 * Author: Jianmin Lv <lvjianmin@loongson.cn>
 *         Huacai Chen <chenhuacai@loongson.cn>
 * Copyright (C) 2020-2022 Loongson Technology Corporation Limited
 */

#include <linux/init.h>
#include <linux/acpi.h>
#include <linux/efi-bgrt.h>
#include <linux/export.h>
#include <linux/irq.h>
#include <linux/irqdomain.h>
#include <linux/memblock.h>
#include <linux/of_fdt.h>
#include <linux/serial_core.h>
#include <asm/io.h>
#include <asm/numa.h>
#include <asm/loongson.h>

int acpi_disabled;
EXPORT_SYMBOL(acpi_disabled);
int acpi_noirq;
int acpi_pci_disabled;
EXPORT_SYMBOL(acpi_pci_disabled);
int acpi_strict = 1; /* We have no workarounds on LoongArch */
int num_processors;
int disabled_cpus;

u64 acpi_saved_sp;

#define PREFIX                  "ACPI: "

struct acpi_madt_core_pic acpi_core_pic[MAX_CORE_PIC];

void __init __iomem * __acpi_map_table(unsigned long phys, unsigned long size)
{

        if (!phys || !size)
                return NULL;

        return early_memremap(phys, size);
}
void __init __acpi_unmap_table(void __iomem *map, unsigned long size)
{
        if (!map || !size)
                return;

        early_memunmap(map, size);
}

void __iomem *acpi_os_ioremap(acpi_physical_address phys, acpi_size size)
{
        if (!memblock_is_memory(phys))
                return ioremap(phys, size);
        else
                return ioremap_cache(phys, size);
}

#ifdef CONFIG_SMP
static int set_processor_mask(u32 id, u32 pass)
{
        int cpu = -1, cpuid = id;

        if (num_processors >= NR_CPUS) {
                pr_warn(PREFIX "nr_cpus limit of %i reached."
                        " processor 0x%x ignored.\n", NR_CPUS, cpuid);

                return -ENODEV;

        }

        if (cpuid == loongson_sysconf.boot_cpu_id)
                cpu = 0;

        switch (pass) {
        case 1: /* Pass 1 handle enabled processors */
                if (cpu < 0)
                        cpu = find_first_zero_bit(cpumask_bits(cpu_present_mask), NR_CPUS);
                num_processors++;
                set_cpu_present(cpu, true);
                break;
        case 2: /* Pass 2 handle disabled processors */
                if (cpu < 0)
                        cpu = find_first_zero_bit(cpumask_bits(cpu_possible_mask), NR_CPUS);
                disabled_cpus++;
                break;
        default:
                return cpu;
        }

        set_cpu_possible(cpu, true);
        __cpu_number_map[cpuid] = cpu;
        __cpu_logical_map[cpu] = cpuid;

        return cpu;
}
#endif

static int __init
acpi_parse_p1_processor(union acpi_subtable_headers *header, const unsigned long end)
{
        struct acpi_madt_core_pic *processor = NULL;

        processor = (struct acpi_madt_core_pic *)header;
        if (BAD_MADT_ENTRY(processor, end))
                return -EINVAL;

        acpi_table_print_madt_entry(&header->common);
#ifdef CONFIG_SMP
        acpi_core_pic[processor->core_id] = *processor;
        if (processor->flags & ACPI_MADT_ENABLED)
                set_processor_mask(processor->core_id, 1);
#endif

        return 0;
}

static int __init
acpi_parse_p2_processor(union acpi_subtable_headers *header, const unsigned long end)
{
        struct acpi_madt_core_pic *processor = NULL;

        processor = (struct acpi_madt_core_pic *)header;
        if (BAD_MADT_ENTRY(processor, end))
                return -EINVAL;

#ifdef CONFIG_SMP
        if (!(processor->flags & ACPI_MADT_ENABLED))
                set_processor_mask(processor->core_id, 2);
#endif

        return 0;
}
static int __init
acpi_parse_eio_master(union acpi_subtable_headers *header, const unsigned long end)
{
        static int core = 0;
        struct acpi_madt_eio_pic *eiointc = NULL;

        eiointc = (struct acpi_madt_eio_pic *)header;
        if (BAD_MADT_ENTRY(eiointc, end))
                return -EINVAL;

        core = eiointc->node * CORES_PER_EIO_NODE;
        set_bit(core, loongson_sysconf.cores_io_master);

        return 0;
}

static void __init acpi_process_madt(void)
{
#ifdef CONFIG_SMP
        int i;

        for (i = 0; i < NR_CPUS; i++) {
                __cpu_number_map[i] = -1;
                __cpu_logical_map[i] = -1;
        }
#endif
        acpi_table_parse_madt(ACPI_MADT_TYPE_CORE_PIC,
                        acpi_parse_p1_processor, MAX_CORE_PIC);

        acpi_table_parse_madt(ACPI_MADT_TYPE_CORE_PIC,
                        acpi_parse_p2_processor, MAX_CORE_PIC);

        acpi_table_parse_madt(ACPI_MADT_TYPE_EIO_PIC,
                        acpi_parse_eio_master, MAX_IO_PICS);

        loongson_sysconf.nr_cpus = num_processors;
}

int pptt_enabled;

int __init parse_acpi_topology(void)
{
        int cpu, topology_id;

        for_each_possible_cpu(cpu) {
                topology_id = find_acpi_cpu_topology(cpu, 0);
                if (topology_id < 0) {
                        pr_warn("Invalid BIOS PPTT\n");
                        return -ENOENT;
                }

                if (acpi_pptt_cpu_is_thread(cpu) <= 0)
                        cpu_data[cpu].core = topology_id;
                else {
                        topology_id = find_acpi_cpu_topology(cpu, 1);
                        if (topology_id < 0)
                                return -ENOENT;

                        cpu_data[cpu].core = topology_id;
                }
        }

        pptt_enabled = 1;

        return 0;
}

#ifndef CONFIG_SUSPEND
int (*acpi_suspend_lowlevel)(void);
#else
int (*acpi_suspend_lowlevel)(void) = loongarch_acpi_suspend;
#endif

void __init acpi_boot_table_init(void)
{
        /*
         * If acpi_disabled, bail out
         */
        if (acpi_disabled)
                goto fdt_earlycon;

        /*
         * Initialize the ACPI boot-time table parser.
         */
        if (acpi_table_init()) {
                disable_acpi();
                goto fdt_earlycon;
        }

        loongson_sysconf.boot_cpu_id = read_csr_cpuid();

        /*
         * Process the Multiple APIC Description Table (MADT), if present
         */
        acpi_process_madt();

        /* Do not enable ACPI SPCR console by default */
        acpi_parse_spcr(earlycon_acpi_spcr_enable, false);

        if (IS_ENABLED(CONFIG_ACPI_BGRT))
                acpi_table_parse(ACPI_SIG_BGRT, acpi_parse_bgrt);

        return;

fdt_earlycon:
        if (earlycon_acpi_spcr_enable)
                early_init_dt_scan_chosen_stdout();
}

#ifdef CONFIG_ACPI_NUMA

/* Callback for Proximity Domain -> CPUID mapping */
void __init
acpi_numa_processor_affinity_init(struct acpi_srat_cpu_affinity *pa)
{
        int pxm, node;

        if (srat_disabled())
                return;
        if (pa->header.length != sizeof(struct acpi_srat_cpu_affinity)) {
                bad_srat();
                return;
        }
        if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0)
                return;
        pxm = pa->proximity_domain_lo;
        if (acpi_srat_revision >= 2) {
                pxm |= (pa->proximity_domain_hi[0] << 8);
                pxm |= (pa->proximity_domain_hi[1] << 16);
                pxm |= (pa->proximity_domain_hi[2] << 24);
        }
        node = acpi_map_pxm_to_node(pxm);
        if (node < 0) {
                pr_err("SRAT: Too many proximity domains %x\n", pxm);
                bad_srat();
                return;
        }

        if (pa->apic_id >= CONFIG_NR_CPUS) {
                pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u skipped apicid that is too big\n",
                                pxm, pa->apic_id, node);
                return;
        }

        early_numa_add_cpu(pa->apic_id, node);

        set_cpuid_to_node(pa->apic_id, node);
        node_set(node, numa_nodes_parsed);
        pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u\n", pxm, pa->apic_id, node);
}

void __init
acpi_numa_x2apic_affinity_init(struct acpi_srat_x2apic_cpu_affinity *pa)
{
        int pxm, node;

        if (srat_disabled())
                return;
        if (pa->header.length < sizeof(struct acpi_srat_x2apic_cpu_affinity)) {
                bad_srat();
                return;
        }
        if ((pa->flags & ACPI_SRAT_CPU_ENABLED) == 0)
                return;
        pxm = pa->proximity_domain;
        node = acpi_map_pxm_to_node(pxm);
        if (node < 0) {
                pr_err("SRAT: Too many proximity domains %x\n", pxm);
                bad_srat();
                return;
        }

        if (pa->apic_id >= CONFIG_NR_CPUS) {
                pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u skipped apicid that is too big\n",
                                pxm, pa->apic_id, node);
                return;
        }

        early_numa_add_cpu(pa->apic_id, node);

        set_cpuid_to_node(pa->apic_id, node);
        node_set(node, numa_nodes_parsed);
        pr_info("SRAT: PXM %u -> CPU 0x%02x -> Node %u\n", pxm, pa->apic_id, node);
}

#endif

void __init arch_reserve_mem_area(acpi_physical_address addr, size_t size)
{
        memblock_reserve(addr, size);
}

#ifdef CONFIG_ACPI_HOTPLUG_CPU

#include <acpi/processor.h>

static int __ref acpi_map_cpu2node(acpi_handle handle, int cpu, int physid)
{
#ifdef CONFIG_ACPI_NUMA
        int nid;

        nid = acpi_get_node(handle);

        if (nid != NUMA_NO_NODE)
                nid = early_cpu_to_node(cpu);

        if (nid != NUMA_NO_NODE) {
                set_cpuid_to_node(physid, nid);
                node_set(nid, numa_nodes_parsed);
                set_cpu_numa_node(cpu, nid);
                cpumask_set_cpu(cpu, cpumask_of_node(nid));
        }
#endif
        return 0;
}

int acpi_map_cpu(acpi_handle handle, phys_cpuid_t physid, u32 acpi_id, int *pcpu)
{
        int cpu;

        cpu = cpu_number_map(physid);
        if (cpu < 0 || cpu >= nr_cpu_ids) {
                pr_info(PREFIX "Unable to map lapic to logical cpu number\n");
                return -ERANGE;
        }

        num_processors++;
        set_cpu_present(cpu, true);
        acpi_map_cpu2node(handle, cpu, physid);

        *pcpu = cpu;

        return 0;
}
EXPORT_SYMBOL(acpi_map_cpu);

int acpi_unmap_cpu(int cpu)
{
#ifdef CONFIG_ACPI_NUMA
        set_cpuid_to_node(cpu_logical_map(cpu), NUMA_NO_NODE);
#endif
        set_cpu_present(cpu, false);
        num_processors--;

        pr_info("cpu%d hot remove!\n", cpu);

        return 0;
}
EXPORT_SYMBOL(acpi_unmap_cpu);

#endif /* CONFIG_ACPI_HOTPLUG_CPU */