root/arch/powerpc/include/asm/book3s/32/mmu-hash.h
/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_
#define _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_

/*
 * 32-bit hash table MMU support
 */

/*
 * BATs
 */

/* Block size masks */
#define BL_128K 0x000
#define BL_256K 0x001
#define BL_512K 0x003
#define BL_1M   0x007
#define BL_2M   0x00F
#define BL_4M   0x01F
#define BL_8M   0x03F
#define BL_16M  0x07F
#define BL_32M  0x0FF
#define BL_64M  0x1FF
#define BL_128M 0x3FF
#define BL_256M 0x7FF

/* BAT Access Protection */
#define BPP_XX  0x00            /* No access */
#define BPP_RX  0x01            /* Read only */
#define BPP_RW  0x02            /* Read/write */

#ifndef __ASSEMBLER__
/* Contort a phys_addr_t into the right format/bits for a BAT */
#ifdef CONFIG_PHYS_64BIT
#define BAT_PHYS_ADDR(x) ((u32)((x & 0x00000000fffe0000ULL) | \
                                ((x & 0x0000000e00000000ULL) >> 24) | \
                                ((x & 0x0000000100000000ULL) >> 30)))
#define PHYS_BAT_ADDR(x) (((u64)(x) & 0x00000000fffe0000ULL) | \
                          (((u64)(x) << 24) & 0x0000000e00000000ULL) | \
                          (((u64)(x) << 30) & 0x0000000100000000ULL))
#else
#define BAT_PHYS_ADDR(x) (x)
#define PHYS_BAT_ADDR(x) ((x) & 0xfffe0000)
#endif

struct ppc_bat {
        u32 batu;
        u32 batl;
};
#endif /* !__ASSEMBLER__ */

/*
 * Hash table
 */

/* Values for PP (assumes Ks=0, Kp=1) */
#define PP_RWXX 0       /* Supervisor read/write, User none */
#define PP_RWRX 1       /* Supervisor read/write, User read */
#define PP_RWRW 2       /* Supervisor read/write, User read/write */
#define PP_RXRX 3       /* Supervisor read,       User read */

/* Values for Segment Registers */
#define SR_NX   0x10000000      /* No Execute */
#define SR_KP   0x20000000      /* User key */
#define SR_KS   0x40000000      /* Supervisor key */

#ifdef __ASSEMBLER__

#include <asm/asm-offsets.h>

.macro uus_addi sr reg1 reg2 imm
        .if NUM_USER_SEGMENTS > \sr
        addi    \reg1,\reg2,\imm
        .endif
.endm

.macro uus_mtsr sr reg1
        .if NUM_USER_SEGMENTS > \sr
        mtsr    \sr, \reg1
        .endif
.endm

/*
 * This isync() shouldn't be necessary as the kernel is not excepted to run
 * any instruction in userspace soon after the update of segments and 'rfi'
 * instruction is used to return to userspace, but hash based cores
 * (at least G3) seem to exhibit a random behaviour when the 'isync' is not
 * there. 603 cores don't have this behaviour so don't do the 'isync' as it
 * saves several CPU cycles.
 */
.macro uus_isync
#ifdef CONFIG_PPC_BOOK3S_604
BEGIN_MMU_FTR_SECTION
        isync
END_MMU_FTR_SECTION_IFSET(MMU_FTR_HPTE_TABLE)
#endif
.endm

.macro update_user_segments_by_4 tmp1 tmp2 tmp3 tmp4
        uus_addi        1, \tmp2, \tmp1, 0x111
        uus_addi        2, \tmp3, \tmp1, 0x222
        uus_addi        3, \tmp4, \tmp1, 0x333

        uus_mtsr        0, \tmp1
        uus_mtsr        1, \tmp2
        uus_mtsr        2, \tmp3
        uus_mtsr        3, \tmp4

        uus_addi        4, \tmp1, \tmp1, 0x444
        uus_addi        5, \tmp2, \tmp2, 0x444
        uus_addi        6, \tmp3, \tmp3, 0x444
        uus_addi        7, \tmp4, \tmp4, 0x444

        uus_mtsr        4, \tmp1
        uus_mtsr        5, \tmp2
        uus_mtsr        6, \tmp3
        uus_mtsr        7, \tmp4

        uus_addi        8, \tmp1, \tmp1, 0x444
        uus_addi        9, \tmp2, \tmp2, 0x444
        uus_addi        10, \tmp3, \tmp3, 0x444
        uus_addi        11, \tmp4, \tmp4, 0x444

        uus_mtsr        8, \tmp1
        uus_mtsr        9, \tmp2
        uus_mtsr        10, \tmp3
        uus_mtsr        11, \tmp4

        uus_addi        12, \tmp1, \tmp1, 0x444
        uus_addi        13, \tmp2, \tmp2, 0x444
        uus_addi        14, \tmp3, \tmp3, 0x444
        uus_addi        15, \tmp4, \tmp4, 0x444

        uus_mtsr        12, \tmp1
        uus_mtsr        13, \tmp2
        uus_mtsr        14, \tmp3
        uus_mtsr        15, \tmp4

        uus_isync
.endm

#else

/*
 * This macro defines the mapping from contexts to VSIDs (virtual
 * segment IDs).  We use a skew on both the context and the high 4 bits
 * of the 32-bit virtual address (the "effective segment ID") in order
 * to spread out the entries in the MMU hash table.  Note, if this
 * function is changed then hash functions will have to be
 * changed to correspond.
 */
#define CTX_TO_VSID(c, id)      ((((c) * (897 * 16)) + (id * 0x111)) & 0xffffff)

/*
 * Hardware Page Table Entry
 * Note that the xpn and x bitfields are used only by processors that
 * support extended addressing; otherwise, those bits are reserved.
 */
struct hash_pte {
        unsigned long v:1;      /* Entry is valid */
        unsigned long vsid:24;  /* Virtual segment identifier */
        unsigned long h:1;      /* Hash algorithm indicator */
        unsigned long api:6;    /* Abbreviated page index */
        unsigned long rpn:20;   /* Real (physical) page number */
        unsigned long xpn:3;    /* Real page number bits 0-2, optional */
        unsigned long r:1;      /* Referenced */
        unsigned long c:1;      /* Changed */
        unsigned long w:1;      /* Write-thru cache mode */
        unsigned long i:1;      /* Cache inhibited */
        unsigned long m:1;      /* Memory coherence */
        unsigned long g:1;      /* Guarded */
        unsigned long x:1;      /* Real page number bit 3, optional */
        unsigned long pp:2;     /* Page protection */
};

typedef struct {
        unsigned long id;
        unsigned long sr0;
        void __user *vdso;
} mm_context_t;

#ifdef CONFIG_PPC_KUEP
#define INIT_MM_CONTEXT(mm) .context.sr0 = SR_NX
#endif

void update_bats(void);
static inline void cleanup_cpu_mmu_context(void) { }

/* patch sites */
extern s32 patch__hash_page_A0, patch__hash_page_A1, patch__hash_page_A2;
extern s32 patch__hash_page_B, patch__hash_page_C;
extern s32 patch__flush_hash_A0, patch__flush_hash_A1, patch__flush_hash_A2;
extern s32 patch__flush_hash_B;

#include <linux/sizes.h>
#include <linux/align.h>

#include <asm/reg.h>
#include <asm/task_size_32.h>

static __always_inline void update_user_segment(u32 n, u32 val)
{
        if (n << 28 < ALIGN(TASK_SIZE, SZ_256M))
                mtsr(val + n * 0x111, n << 28);
}

static __always_inline void update_user_segments(u32 val)
{
        val &= 0xf0ffffff;

        update_user_segment(0, val);
        update_user_segment(1, val);
        update_user_segment(2, val);
        update_user_segment(3, val);
        update_user_segment(4, val);
        update_user_segment(5, val);
        update_user_segment(6, val);
        update_user_segment(7, val);
        update_user_segment(8, val);
        update_user_segment(9, val);
        update_user_segment(10, val);
        update_user_segment(11, val);
        update_user_segment(12, val);
        update_user_segment(13, val);
        update_user_segment(14, val);
        update_user_segment(15, val);
}

int __init find_free_bat(void);
unsigned int bat_block_size(unsigned long base, unsigned long top);
#endif /* !__ASSEMBLER__ */

/* We happily ignore the smaller BATs on 601, we don't actually use
 * those definitions on hash32 at the moment anyway
 */
#define mmu_virtual_psize       MMU_PAGE_4K
#define mmu_linear_psize        MMU_PAGE_256M

#endif /* _ASM_POWERPC_BOOK3S_32_MMU_HASH_H_ */