root/sys/arch/arm/arm/vm_machdep.c
/*      $OpenBSD: vm_machdep.c,v 1.30 2025/05/21 09:06:58 mpi Exp $     */
/*      $NetBSD: vm_machdep.c,v 1.31 2004/01/04 11:33:29 jdolecek Exp $ */

/*
 * Copyright (c) 1994-1998 Mark Brinicombe.
 * Copyright (c) 1994 Brini.
 * All rights reserved.
 *
 * This code is derived from software written for Brini by Mark Brinicombe
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. All advertising materials mentioning features or use of this software
 *    must display the following acknowledgement:
 *      This product includes software developed by Brini.
 * 4. The name of the company nor the name of the author may be used to
 *    endorse or promote products derived from this software without specific
 *    prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED
 * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
 * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT,
 * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES
 * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
 * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 *
 * RiscBSD kernel project
 *
 * vm_machdep.h
 *
 * vm machine specific bits
 *
 * Created      : 08/10/94
 */

#include <sys/param.h>
#include <sys/systm.h>
#include <sys/proc.h>
#include <sys/buf.h>
#include <sys/user.h>

#include <uvm/uvm_extern.h>

#include <machine/cpu.h>
#include <machine/pmap.h>

#include <arm/vfp.h>

extern pv_addr_t systempage;

int process_read_regs   (struct proc *p, struct reg *regs);
int process_read_fpregs (struct proc *p, struct fpreg *regs);

extern void proc_trampoline     (void);

/*
 * Finish a fork operation, with process p2 nearly set up.
 * Copy and update the pcb and trap frame, making the child ready to run.
 *
 * Rig the child's kernel stack so that it will start out in
 * proc_trampoline() and call 'func' with 'arg' as an argument.
 * For normal processes this is child_return(), which causes the
 * child to go directly to user level with an apparent return value
 * of 0 from fork(), while the parent process returns normally.
 * For kernel threads this will be a function that never returns.
 *
 * An alternate user-level stack or TCB can be requested by passing
 * a non-NULL value; these are poked into the PCB so they're in
 * effect at the initial return to userspace.
 */
void
cpu_fork(struct proc *p1, struct proc *p2, void *stack, void *tcb,
    void (*func)(void *), void *arg)
{
        struct pcb *pcb = &p2->p_addr->u_pcb;
        struct trapframe *tf;
        struct switchframe *sf;

        if (p1 == curproc) {
                /* Sync the PCB before we copy it. */
                savectx(curpcb);
        }

        /* Copy the pcb */
        *pcb = p1->p_addr->u_pcb;

        /* 
         * Set up the undefined stack for the process.
         * Note: this stack is not in use if we are forking from p1
         */
        pcb->pcb_un.un_32.pcb32_und_sp = (u_int)p2->p_addr +
            USPACE_UNDEF_STACK_TOP;
        pcb->pcb_un.un_32.pcb32_sp = (u_int)p2->p_addr + USPACE_SVC_STACK_TOP;

        pmap_activate(p2);

        pcb->pcb_tf = tf = (struct trapframe *)pcb->pcb_un.un_32.pcb32_sp - 1;
        *tf = *p1->p_addr->u_pcb.pcb_tf;

        /*
         * If specified, give the child a different stack and/or TCB.
         * Enforce 8-byte alignment on the stack.
         */
        if (stack != NULL)
                tf->tf_usr_sp = (vaddr_t)stack & -8;
        if (tcb != NULL)
                p2->p_addr->u_pcb.pcb_tcb = tcb;

        sf = (struct switchframe *)tf - 1;
        sf->sf_r4 = (u_int)func;
        sf->sf_r5 = (u_int)arg;
        sf->sf_pc = (u_int)proc_trampoline;
        pcb->pcb_un.un_32.pcb32_sp = (u_int)sf;
}

void
cpu_exit(struct proc *p)
{
        /* If we were using the FPU, forget about it. */
        if (p->p_addr->u_pcb.pcb_fpcpu != NULL)
                vfp_discard(p);
}

struct kmem_va_mode kv_physwait = {
        .kv_map = &phys_map,
        .kv_wait = 1,
};

/*
 * Map a user I/O request into kernel virtual address space.
 * Note: the pages are already locked by uvm_vslock(), so we
 * do not need to pass an access_type to pmap_enter().
 */
void
vmapbuf(struct buf *bp, vsize_t len)
{
        vaddr_t faddr, taddr, off;
        paddr_t fpa;

        if ((bp->b_flags & B_PHYS) == 0)
                panic("vmapbuf");
        faddr = trunc_page((vaddr_t)(bp->b_saveaddr = bp->b_data));
        off = (vaddr_t)bp->b_data - faddr;
        len = round_page(off + len);
        taddr = (vaddr_t)km_alloc(len, &kv_physwait, &kp_none, &kd_waitok);
        bp->b_data = (caddr_t)(taddr + off);
        /*
         * The region is locked, so we expect that pmap_pte() will return
         * non-NULL.
         * XXX: unwise to expect this in a multithreaded environment.
         * anything can happen to a pmap between the time we lock a
         * region, release the pmap lock, and then relock it for
         * the pmap_extract().
         *
         * no need to flush TLB since we expect nothing to be mapped
         * where we just allocated (TLB will be flushed when our
         * mapping is removed).
         */
        while (len) {
                (void) pmap_extract(vm_map_pmap(&bp->b_proc->p_vmspace->vm_map),
                    faddr, &fpa);
                pmap_kenter_pa(taddr, fpa, PROT_READ | PROT_WRITE);
                faddr += PAGE_SIZE;
                taddr += PAGE_SIZE;
                len -= PAGE_SIZE;
        }
        pmap_update(pmap_kernel());
}

/*
 * Unmap a previously-mapped user I/O request.
 */
void
vunmapbuf(struct buf *bp, vsize_t len)
{
        vaddr_t addr, off;

        if ((bp->b_flags & B_PHYS) == 0)
                panic("vunmapbuf");
        addr = trunc_page((vaddr_t)bp->b_data);
        off = (vaddr_t)bp->b_data - addr;
        len = round_page(off + len);
        pmap_kremove(addr, len);
        pmap_update(pmap_kernel());
        km_free((void *)addr, len, &kv_physwait, &kp_none);
        bp->b_data = bp->b_saveaddr;
        bp->b_saveaddr = NULL;
}