root/sys/powerpc/include/pmap.h
/*-
 * SPDX-License-Identifier: BSD-3-Clause AND BSD-4-Clause
 *
 * Copyright (C) 2006 Semihalf, Marian Balakowicz <m8@semihalf.com>
 * All rights reserved.
 *
 * Adapted for Freescale's e500 core CPUs.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. The name of the author may not be used to endorse or promote products
 *    derived from this software without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.  IN
 * NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED
 * TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
 * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
 * LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
 * NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
 * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 */
/*-
 * Copyright (C) 1995, 1996 Wolfgang Solfrank.
 * Copyright (C) 1995, 1996 TooLs GmbH.
 * All rights reserved.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. All advertising materials mentioning features or use of this software
 *    must display the following acknowledgement:
 *      This product includes software developed by TooLs GmbH.
 * 4. The name of TooLs GmbH may not be used to endorse or promote products
 *    derived from this software without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY TOOLS GMBH ``AS IS'' AND ANY EXPRESS OR
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
 * IN NO EVENT SHALL TOOLS GMBH BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
 * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS;
 * OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
 * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
 * OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
 * ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 *
 *      from: $NetBSD: pmap.h,v 1.17 2000/03/30 16:18:24 jdolecek Exp $
 */

#ifndef _MACHINE_PMAP_H_
#define _MACHINE_PMAP_H_

#include <sys/queue.h>
#include <sys/tree.h>
#include <sys/_cpuset.h>
#include <sys/_lock.h>
#include <sys/_mutex.h>
#include <machine/sr.h>
#include <machine/pte.h>
#include <machine/slb.h>
#include <machine/tlb.h>
#include <machine/vmparam.h>
#ifdef __powerpc64__
#include <vm/_vm_radix.h>
#endif

/*
 * The radix page table structure is described by levels 1-4.
 * See Fig 33. on p. 1002 of Power ISA v3.0B
 *
 * Page directories and tables must be size aligned.
 */

/* Root page directory - 64k   -- each entry covers 512GB */
typedef uint64_t pml1_entry_t;
/* l2 page directory - 4k      -- each entry covers 1GB */
typedef uint64_t pml2_entry_t;
/* l3 page directory - 4k      -- each entry covers 2MB */
typedef uint64_t pml3_entry_t;
/* l4 page directory - 256B/4k -- each entry covers 64k/4k */
typedef uint64_t pml4_entry_t;

typedef uint64_t pt_entry_t;

struct pmap;
typedef struct pmap *pmap_t;

#define PMAP_ENTER_QUICK_LOCKED 0x10000000

#if !defined(NPMAPS)
#define NPMAPS          32768
#endif /* !defined(NPMAPS) */

struct  slbtnode;

struct pvo_entry {
        LIST_ENTRY(pvo_entry) pvo_vlink;        /* Link to common virt page */
#ifndef __powerpc64__
        LIST_ENTRY(pvo_entry) pvo_olink;        /* Link to overflow entry */
#endif
        union {
                RB_ENTRY(pvo_entry) pvo_plink;  /* Link to pmap entries */
                SLIST_ENTRY(pvo_entry) pvo_dlink; /* Link to delete enty */
        };
        struct {
#ifndef __powerpc64__
                /* 32-bit fields */
                pte_t       pte;
#endif
                /* 64-bit fields */
                uintptr_t   slot;
                vm_paddr_t  pa;
                vm_prot_t   prot;
        } pvo_pte;
        pmap_t          pvo_pmap;               /* Owning pmap */
        vm_offset_t     pvo_vaddr;              /* VA of entry */
        uint64_t        pvo_vpn;                /* Virtual page number */
};
LIST_HEAD(pvo_head, pvo_entry);
SLIST_HEAD(pvo_dlist, pvo_entry);
RB_HEAD(pvo_tree, pvo_entry);
int pvo_vaddr_compare(struct pvo_entry *, struct pvo_entry *);
RB_PROTOTYPE(pvo_tree, pvo_entry, pvo_plink, pvo_vaddr_compare);

/* Used by 32-bit PMAP */
#define PVO_PTEGIDX_MASK        0x007UL         /* which PTEG slot */
#define PVO_PTEGIDX_VALID       0x008UL         /* slot is valid */
/* Used by 64-bit PMAP */
#define PVO_HID                 0x008UL         /* PVO entry in alternate hash*/
/* Used by both */
#define PVO_WIRED               0x010UL         /* PVO entry is wired */
#define PVO_MANAGED             0x020UL         /* PVO entry is managed */
#define PVO_BOOTSTRAP           0x080UL         /* PVO entry allocated during
                                                   bootstrap */
#define PVO_DEAD                0x100UL         /* waiting to be deleted */
#define PVO_LARGE               0x200UL         /* large page */
#define PVO_VADDR(pvo)          ((pvo)->pvo_vaddr & ~ADDR_POFF)
#define PVO_PTEGIDX_GET(pvo)    ((pvo)->pvo_vaddr & PVO_PTEGIDX_MASK)
#define PVO_PTEGIDX_ISSET(pvo)  ((pvo)->pvo_vaddr & PVO_PTEGIDX_VALID)
#define PVO_PTEGIDX_CLR(pvo)    \
        ((void)((pvo)->pvo_vaddr &= ~(PVO_PTEGIDX_VALID|PVO_PTEGIDX_MASK)))
#define PVO_PTEGIDX_SET(pvo, i) \
        ((void)((pvo)->pvo_vaddr |= (i)|PVO_PTEGIDX_VALID))
#define PVO_VSID(pvo)           ((pvo)->pvo_vpn >> 16)

struct  pmap {
        struct          pmap_statistics pm_stats;
        struct  mtx     pm_mtx;
        cpuset_t        pm_active;
        union {
                struct {
                    #ifdef __powerpc64__
                        struct slbtnode *pm_slb_tree_root;
                        struct slb      **pm_slb;
                        int             pm_slb_len;
                    #else
                        register_t      pm_sr[16];
                    #endif

                        struct pmap     *pmap_phys;
                        struct pvo_tree pmap_pvo;
                };
#ifdef __powerpc64__
                /* Radix support */
                struct {
                        pml1_entry_t    *pm_pml1;       /* KVA of root page directory */
                        struct vm_radix  pm_radix;      /* spare page table pages */
                        TAILQ_HEAD(,pv_chunk)   pm_pvchunk;     /* list of mappings in pmap */
                        uint64_t        pm_pid; /* PIDR value */
                        int pm_flags;
                };
#endif
                struct {
                        /* TID to identify this pmap entries in TLB */
                        tlbtid_t        pm_tid[MAXCPU];

#ifdef __powerpc64__
                        /*
                         * Page table directory,
                         * array of pointers to page directories.
                         */
                        pte_t ****pm_root;
#else
                        /*
                         * Page table directory,
                         * array of pointers to page tables.
                         */
                        pte_t           **pm_pdir;

                        /* List of allocated ptbl bufs (ptbl kva regions). */
                        TAILQ_HEAD(, ptbl_buf)  pm_ptbl_list;
#endif
                };
        } __aligned(CACHE_LINE_SIZE);
};

/*
 * pv_entries are allocated in chunks per-process.  This avoids the
 * need to track per-pmap assignments.
 */
#define _NPCPV  126
#define _NPCM   howmany(_NPCPV, 64)

#define PV_CHUNK_HEADER                                                 \
        pmap_t                  pc_pmap;                                \
        TAILQ_ENTRY(pv_chunk)   pc_list;                                \
        uint64_t                pc_map[_NPCM];  /* bitmap; 1 = free */  \
        TAILQ_ENTRY(pv_chunk)   pc_lru;

struct pv_entry {
        pmap_t pv_pmap;
        vm_offset_t pv_va;
        TAILQ_ENTRY(pv_entry) pv_link;
};
typedef struct pv_entry *pv_entry_t;

struct pv_chunk_header {
        PV_CHUNK_HEADER
};
struct pv_chunk {
        PV_CHUNK_HEADER
        uint64_t        reserved;
        struct pv_entry         pc_pventry[_NPCPV];
};

struct  md_page {
        union {
                struct {
                        volatile int32_t mdpg_attrs;
                        vm_memattr_t     mdpg_cache_attrs;
                        struct  pvo_head mdpg_pvoh;
                        int             pv_gen;   /* (p) */
                };
                struct {
                        int                     pv_tracked;
                };
        };
        TAILQ_HEAD(, pv_entry)  pv_list;  /* (p) */
};

#ifdef AIM
#define pmap_page_get_memattr(m)        ((m)->md.mdpg_cache_attrs)
#else
#define pmap_page_get_memattr(m)        VM_MEMATTR_DEFAULT
#endif /* AIM */

/*
 * Return the VSID corresponding to a given virtual address.
 * If no VSID is currently defined, it will allocate one, and add
 * it to a free slot if available.
 *
 * NB: The PMAP MUST be locked already.
 */
uint64_t va_to_vsid(pmap_t pm, vm_offset_t va);

/* Lock-free, non-allocating lookup routines */
uint64_t kernel_va_to_slbv(vm_offset_t va);
struct slb *user_va_to_slb_entry(pmap_t pm, vm_offset_t va);

uint64_t allocate_user_vsid(pmap_t pm, uint64_t esid, int large);
void    free_vsid(pmap_t pm, uint64_t esid, int large);
void    slb_insert_user(pmap_t pm, struct slb *slb);
void    slb_insert_kernel(uint64_t slbe, uint64_t slbv);

struct slbtnode *slb_alloc_tree(void);
void     slb_free_tree(pmap_t pm);
struct slb **slb_alloc_user_cache(void);
void    slb_free_user_cache(struct slb **);

extern  struct pmap kernel_pmap_store;
#define kernel_pmap     (&kernel_pmap_store)

#ifdef _KERNEL

#define PMAP_LOCK(pmap)         mtx_lock(&(pmap)->pm_mtx)
#define PMAP_LOCK_ASSERT(pmap, type) \
                                mtx_assert(&(pmap)->pm_mtx, (type))
#define PMAP_LOCK_DESTROY(pmap) mtx_destroy(&(pmap)->pm_mtx)
#define PMAP_LOCK_INIT(pmap)    mtx_init(&(pmap)->pm_mtx, \
                                    (pmap == kernel_pmap) ? "kernelpmap" : \
                                    "pmap", NULL, MTX_DEF | MTX_DUPOK)
#define PMAP_LOCKED(pmap)       mtx_owned(&(pmap)->pm_mtx)
#define PMAP_MTX(pmap)          (&(pmap)->pm_mtx)
#define PMAP_TRYLOCK(pmap)      mtx_trylock(&(pmap)->pm_mtx)
#define PMAP_UNLOCK(pmap)       mtx_unlock(&(pmap)->pm_mtx)

#define pmap_page_is_write_mapped(m)    (((m)->a.flags & PGA_WRITEABLE) != 0)

#define pmap_vm_page_alloc_check(m)

void            pmap_bootstrap(vm_offset_t, vm_offset_t);
void            pmap_kenter(vm_offset_t va, vm_paddr_t pa);
void            pmap_kenter_attr(vm_offset_t va, vm_paddr_t pa, vm_memattr_t);
void            pmap_kremove(vm_offset_t);
void            *pmap_mapdev(vm_paddr_t, vm_size_t);
void            *pmap_mapdev_attr(vm_paddr_t, vm_size_t, vm_memattr_t);
void            pmap_unmapdev(void *, vm_size_t);
void            pmap_page_set_memattr(vm_page_t, vm_memattr_t);
int             pmap_change_attr(vm_offset_t, vm_size_t, vm_memattr_t);
int             pmap_map_user_ptr(pmap_t pm, volatile const void *uaddr,
                    void **kaddr, size_t ulen, size_t *klen);
int             pmap_decode_kernel_ptr(vm_offset_t addr, int *is_user,
                    vm_offset_t *decoded_addr);
void            pmap_deactivate(struct thread *);
vm_paddr_t      pmap_kextract(vm_offset_t);
int             pmap_dev_direct_mapped(vm_paddr_t, vm_size_t);
bool            pmap_mmu_install(char *name, int prio);
void            pmap_mmu_init(void);
const char      *pmap_mmu_name(void);
bool            pmap_ps_enabled(pmap_t pmap);
int             pmap_nofault(pmap_t pmap, vm_offset_t va, vm_prot_t flags);
bool            pmap_page_is_mapped(vm_page_t m);
#define pmap_map_delete(pmap, sva, eva) pmap_remove(pmap, sva, eva)

void            pmap_page_array_startup(long count);

#define vtophys(va)     pmap_kextract((vm_offset_t)(va))

extern  vm_offset_t virtual_avail;
extern  vm_offset_t virtual_end;
extern  caddr_t crashdumpmap;

extern  vm_offset_t msgbuf_phys;

extern  int pmap_bootstrapped;
extern  int radix_mmu;
extern  int superpages_enabled;

#ifdef AIM
void pmap_early_io_map_init(void);
#endif
vm_offset_t pmap_early_io_map(vm_paddr_t pa, vm_size_t size);
void pmap_early_io_unmap(vm_offset_t va, vm_size_t size);
void pmap_track_page(pmap_t pmap, vm_offset_t va);
void pmap_page_print_mappings(vm_page_t m);
void pmap_tlbie_all(void);

static inline int
pmap_vmspace_copy(pmap_t dst_pmap __unused, pmap_t src_pmap __unused)
{

        return (0);
}

#endif

#endif /* !_MACHINE_PMAP_H_ */