root/src/system/kernel/arch/riscv64/arch_vm_translation_map.cpp
/*
 * Copyright 2007-2010, François Revol, revol@free.fr.
 * Copyright 2008-2010, Ingo Weinhold, ingo_weinhold@gmx.de.
 * Copyright 2002-2007, Axel Dörfler, axeld@pinc-software.de. All rights
 *   reserved.
 * Copyright 2019, Adrien Destugues, pulkomandy@pulkomandy.tk.
 * Distributed under the terms of the MIT License.
 *
 * Copyright 2001, Travis Geiselbrecht. All rights reserved.
 * Distributed under the terms of the NewOS License.
 */


#include <arch_cpu_defs.h>
#include <boot/kernel_args.h>
#include <KernelExport.h>
#include <kernel.h>
#include <vm/vm.h>
#include <vm/vm_priv.h>
#include <vm/VMAddressSpace.h>
#include <Clint.h>
#include <Htif.h>
#include <Plic.h>

#include "RISCV64VMTranslationMap.h"


#define TRACE_VM_TMAP
#ifdef TRACE_VM_TMAP
#       define TRACE(x...) dprintf(x)
#else
#       define TRACE(x...) ;
#endif


ssize_t gVirtFromPhysOffset = 0;

phys_addr_t sPageTable = 0;
char sPhysicalPageMapperData[sizeof(RISCV64VMPhysicalPageMapper)];


// TODO: Consolidate function with RISCV64VMTranslationMap

static Pte*
LookupPte(addr_t virtAdr, bool alloc, kernel_args* args)
{
        Pte *pte = (Pte*)VirtFromPhys(sPageTable);
        for (int level = 2; level > 0; level --) {
                pte += VirtAdrPte(virtAdr, level);
                if (!pte->isValid) {
                        if (!alloc)
                                return NULL;
                        page_num_t ppn = vm_allocate_early_physical_page(args);
                        if (ppn == 0)
                                return NULL;
                        memset((Pte*)VirtFromPhys(B_PAGE_SIZE * ppn), 0, B_PAGE_SIZE);
                        Pte newPte {
                                .isValid = true,
                                .isGlobal = true,
                                .ppn = ppn
                        };
                        pte->val = newPte.val;
                }
                pte = (Pte*)VirtFromPhys(B_PAGE_SIZE * pte->ppn);
        }
        pte += VirtAdrPte(virtAdr, 0);
        return pte;
}


static void
Map(addr_t virtAdr, phys_addr_t physAdr, uint64 flags, kernel_args* args)
{
        // dprintf("Map(0x%" B_PRIxADDR ", 0x%" B_PRIxADDR ")\n", virtAdr, physAdr);
        Pte* pte = LookupPte(virtAdr, true, args);
        if (pte == NULL) panic("can't allocate page table");

        Pte newPte {
                .isValid = true,
                .isGlobal = true, // we map only kernel pages here so always set global flag
                .isAccessed = true,
                .isDirty = true,
                .ppn = physAdr / B_PAGE_SIZE
        };
        newPte.val |= flags;

        pte->val = newPte.val;

        FlushTlbPage(virtAdr);
}


//#pragma mark -

status_t
arch_vm_translation_map_init(kernel_args *args,
        VMPhysicalPageMapper** _physicalPageMapper)
{
        TRACE("vm_translation_map_init: entry\n");

#ifdef TRACE_VM_TMAP
        TRACE("physical memory ranges:\n");
        for (uint32 i = 0; i < args->num_physical_memory_ranges; i++) {
                phys_addr_t start = args->physical_memory_range[i].start;
                phys_addr_t end = start + args->physical_memory_range[i].size;
                TRACE("  %" B_PRIxPHYSADDR " - %" B_PRIxPHYSADDR "\n", start, end);
        }

        TRACE("allocated physical ranges:\n");
        for (uint32 i = 0; i < args->num_physical_allocated_ranges; i++) {
                phys_addr_t start = args->physical_allocated_range[i].start;
                phys_addr_t end = start + args->physical_allocated_range[i].size;
                TRACE("  %" B_PRIxPHYSADDR " - %" B_PRIxPHYSADDR "\n", start, end);
        }

        TRACE("allocated virtual ranges:\n");
        for (uint32 i = 0; i < args->num_virtual_allocated_ranges; i++) {
                addr_t start = args->virtual_allocated_range[i].start;
                addr_t end = start + args->virtual_allocated_range[i].size;
                TRACE("  %" B_PRIxADDR " - %" B_PRIxADDR "\n", start, end);
        }

        TRACE("kernel args ranges:\n");
        for (uint32 i = 0; i < args->num_kernel_args_ranges; i++) {
                phys_addr_t start = args->kernel_args_range[i].start;
                phys_addr_t end = start + args->kernel_args_range[i].size;
                TRACE("  %" B_PRIxPHYSADDR " - %" B_PRIxPHYSADDR "\n", start, end);
        }
#endif

        sPageTable = SatpReg{.val = Satp()}.ppn * B_PAGE_SIZE;

        dprintf("physMapBase: %#" B_PRIxADDR "\n", args->arch_args.physMap.start);
        dprintf("physMemBase: %#" B_PRIxADDR "\n", args->physical_memory_range[0].start);
        gVirtFromPhysOffset = args->arch_args.physMap.start - args->physical_memory_range[0].start;

        arch_cpu_disable_user_access();

        *_physicalPageMapper = new(&sPhysicalPageMapperData)
                RISCV64VMPhysicalPageMapper();

        return B_OK;
}


status_t
arch_vm_translation_map_init_post_sem(kernel_args *args)
{
        return B_OK;
}


status_t
arch_vm_translation_map_init_post_area(kernel_args *args)
{
        TRACE("vm_translation_map_init_post_area: entry\n");
        return B_OK;
}


status_t
arch_vm_translation_map_early_map(kernel_args *args,
        addr_t virtAdr, phys_addr_t physAdr, uint8 attributes)
{
        //dprintf("early_map(%#" B_PRIxADDR ", %#" B_PRIxADDR ")\n", virtAdr, physAdr);
        Pte flags {
                .isRead  = (attributes & B_KERNEL_READ_AREA)    != 0,
                .isWrite = (attributes & B_KERNEL_WRITE_AREA)   != 0,
                .isExec  = (attributes & B_KERNEL_EXECUTE_AREA) != 0,
        };
        Map(virtAdr, physAdr, flags.val, args);
        return B_OK;
}


status_t
arch_vm_translation_map_create_map(bool kernel, VMTranslationMap** _map)
{
        *_map = new(std::nothrow) RISCV64VMTranslationMap(kernel,
                (kernel) ? sPageTable : 0);

        if (*_map == NULL)
                return B_NO_MEMORY;

        return B_OK;
}


bool
arch_vm_translation_map_is_kernel_page_accessible(addr_t virtualAddress,
        uint32 protection)
{
        return virtualAddress != 0;
}