root/sys/arch/sh/include/cpu.h
/*      $OpenBSD: cpu.h,v 1.42 2024/11/06 18:59:09 miod Exp $   */
/*      $NetBSD: cpu.h,v 1.41 2006/01/21 04:24:12 uwe Exp $     */

/*-
 * Copyright (c) 2002 The NetBSD Foundation, Inc. All rights reserved.
 * Copyright (c) 1990 The Regents of the University of California.
 * All rights reserved.
 *
 * This code is derived from software contributed to Berkeley by
 * William Jolitz.
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 * 1. Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 * 2. Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in the
 *    documentation and/or other materials provided with the distribution.
 * 3. Neither the name of the University nor the names of its contributors
 *    may be used to endorse or promote products derived from this software
 *    without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
 * SUCH DAMAGE.
 *
 *      @(#)cpu.h       5.4 (Berkeley) 5/9/91
 */

/*
 * SH3/SH4 support.
 *
 *  T.Horiuchi    Brains Corp.   5/22/98
 */

#ifndef _SH_CPU_H_
#define _SH_CPU_H_

#include <sh/psl.h>
#include <sh/frame.h>

#ifdef _KERNEL

/*
 * Per-CPU information.
 */

#include <machine/intr.h>
#include <sys/clockintr.h>
#include <sys/sched.h>

struct cpu_info {
        struct proc *ci_curproc;

        struct schedstate_percpu ci_schedstate; /* scheduler state */
        u_int32_t ci_randseed;
#ifdef DIAGNOSTIC
        int     ci_mutex_level;
#endif
#ifdef GPROF
        struct gmonparam *ci_gmon;
        struct clockintr ci_gmonclock;
#endif

        int     ci_want_resched;
        int     ci_idepth;

        struct clockqueue ci_queue;

        char    ci_panicbuf[512];
};

extern struct cpu_info cpu_info_store;
#define curcpu()        (&cpu_info_store)
#define cpu_number()    0
#define CPU_IS_PRIMARY(ci)      1
#define CPU_IS_RUNNING(ci)      1
#define CPU_INFO_ITERATOR       int
#define CPU_INFO_FOREACH(cii, ci) \
        for (cii = 0, ci = curcpu(); ci != NULL; ci = NULL)
#define CPU_INFO_UNIT(ci)       0
#define MAXCPUS 1
#define cpu_unidle(ci)

#define CPU_BUSY_CYCLE()        __asm volatile ("" ::: "memory")


/*
 * Arguments to clockintr_dispatch encapsulate the previous
 * machine state in an opaque clockframe.
 */
struct clockframe {
        int     ssr;    /* status register at time of interrupt */
        int     spc;    /* program counter at time of interrupt */
};

#define CLKF_USERMODE(cf)       (!KERNELMODE((cf)->ssr))
#define CLKF_PC(cf)             ((cf)->spc)
#define CLKF_INTR(cf)           (curcpu()->ci_idepth > 1)

/*
 * This is used during profiling to integrate system time.  It can safely
 * assume that the process is resident.
 */
#define PROC_PC(p)      ((p)->p_md.md_regs->tf_spc)
#define PROC_STACK(p)   ((p)->p_md.md_regs->tf_r15)

/*
 * Preempt the current process if in interrupt from user mode,
 * or after the current trap/syscall if in system mode.
 */
void need_resched(struct cpu_info *);
#define clear_resched(ci)       (ci)->ci_want_resched = 0

/*
 * Give a profiling tick to the current process when the user profiling
 * buffer pages are invalid.  On the MIPS, request an ast to send us
 * through trap, marking the proc as needing a profiling tick.
 */
#define need_proftick(p)        aston(p)

/*
 * Notify the current process (p) that it has a signal pending,
 * process as soon as possible.
 */
#define signotify(p)    aston(p)

#define aston(p)        ((p)->p_md.md_astpending = 1)

/*
 * We need a machine-independent name for this.
 */
#define DELAY(x)                delay(x)

#define cpu_idle_enter()        do { /* nothing */ } while (0)
#define cpu_idle_cycle()        __asm volatile("sleep")
#define cpu_idle_leave()        do { /* nothing */ } while (0)

#endif /* _KERNEL */

/*
 * Logical address space of SH3/SH4 CPU.
 */
#define SH3_PHYS_MASK   0x1fffffff

#define SH3_P0SEG_BASE  0x00000000      /* TLB mapped, also U0SEG */
#define SH3_P0SEG_END   0x7fffffff
#define SH3_P1SEG_BASE  0x80000000      /* pa == va */
#define SH3_P1SEG_END   0x9fffffff
#define SH3_P2SEG_BASE  0xa0000000      /* pa == va, non-cacheable */
#define SH3_P2SEG_END   0xbfffffff
#define SH3_P3SEG_BASE  0xc0000000      /* TLB mapped, kernel mode */
#define SH3_P3SEG_END   0xdfffffff
#define SH3_P4SEG_BASE  0xe0000000      /* peripheral space */
#define SH3_P4SEG_END   0xffffffff

#define SH3_P1SEG_TO_PHYS(x)    ((uint32_t)(x) & SH3_PHYS_MASK)
#define SH3_P2SEG_TO_PHYS(x)    ((uint32_t)(x) & SH3_PHYS_MASK)
#define SH3_PHYS_TO_P1SEG(x)    ((uint32_t)(x) | SH3_P1SEG_BASE)
#define SH3_PHYS_TO_P2SEG(x)    ((uint32_t)(x) | SH3_P2SEG_BASE)
#define SH3_P1SEG_TO_P2SEG(x)   ((uint32_t)(x) | 0x20000000)
#define SH3_P2SEG_TO_P1SEG(x)   ((uint32_t)(x) & ~0x20000000)

#ifdef _KERNEL
#ifndef __lint__

/*
 * Switch from P1 (cached) to P2 (uncached).  This used to be written
 * using gcc's assigned goto extension, but gcc4 aggressive optimizations
 * tend to optimize that away under certain circumstances.
 */
#define RUN_P2                                          \
        do {                                            \
                register uint32_t r0 asm("r0");         \
                uint32_t pc;                            \
                __asm volatile(                         \
                        "       mov.l   1f, %1  ;"      \
                        "       mova    2f, %0  ;"      \
                        "       or      %0, %1  ;"      \
                        "       jmp     @%1     ;"      \
                        "        nop            ;"      \
                        "       .align 2        ;"      \
                        "1:     .long   0x20000000;"    \
                        "2:;"                           \
                        : "=r"(r0), "=r"(pc));          \
        } while (0)

/*
 * Switch from P2 (uncached) back to P1 (cached).  We need to be
 * running on P2 to access cache control, memory-mapped cache and TLB
 * arrays, etc. and after touching them at least 8 instructions are
 * necessary before jumping to P1, so provide that padding here.
 */
#define RUN_P1                                          \
        do {                                            \
                register uint32_t r0 asm("r0");         \
                uint32_t pc;                            \
                __asm volatile(                         \
                /*1*/   "       mov.l   1f, %1  ;"      \
                /*2*/   "       mova    2f, %0  ;"      \
                /*3*/   "       nop             ;"      \
                /*4*/   "       and     %0, %1  ;"      \
                /*5*/   "       nop             ;"      \
                /*6*/   "       nop             ;"      \
                /*7*/   "       nop             ;"      \
                /*8*/   "       nop             ;"      \
                        "       jmp     @%1     ;"      \
                        "        nop            ;"      \
                        "       .align 2        ;"      \
                        "1:     .long   ~0x20000000;"   \
                        "2:;"                           \
                        : "=r"(r0), "=r"(pc));          \
        } while (0)

/*
 * If RUN_P1 is the last thing we do in a function we can omit it, b/c
 * we are going to return to a P1 caller anyway, but we still need to
 * ensure there's at least 8 instructions before jump to P1.
 */
#define PAD_P1_SWITCH   __asm volatile ("nop;nop;nop;nop;nop;nop;nop;nop;")

#else  /* __lint__ */
#define RUN_P2          do {} while (/* CONSTCOND */ 0)
#define RUN_P1          do {} while (/* CONSTCOND */ 0)
#define PAD_P1_SWITCH   do {} while (/* CONSTCOND */ 0)
#endif
#endif

#if defined(SH4)
/* SH4 Processor Version Register */
#define SH4_PVR_ADDR    0xff000030      /* P4  address */
#define SH4_PVR         (*(volatile uint32_t *) SH4_PVR_ADDR)
#define SH4_PRR_ADDR    0xff000044      /* P4  address */
#define SH4_PRR         (*(volatile uint32_t *) SH4_PRR_ADDR)

#define SH4_PVR_MASK    0xffffff00
#define SH4_PVR_SH7750  0x04020500      /* SH7750  */
#define SH4_PVR_SH7750S 0x04020600      /* SH7750S */
#define SH4_PVR_SH775xR 0x04050000      /* SH775xR */
#define SH4_PVR_SH7751  0x04110000      /* SH7751  */

#define SH4_PRR_MASK    0xfffffff0
#define SH4_PRR_7750R   0x00000100      /* SH7750R */
#define SH4_PRR_7751R   0x00000110      /* SH7751R */
#endif

/*
 * pull in #defines for kinds of processors
 */
#include <machine/cputypes.h>

#ifdef _KERNEL
void sh_cpu_init(int, int);
void sh_startup(void);
__dead void cpu_reset(void);    /* soft reset */
void _cpu_spin(uint32_t);       /* for delay loop. */
void delay(int);
struct pcb;
void savectx(struct pcb *);
struct fpreg;
void fpu_save(struct fpreg *);
void fpu_restore(struct fpreg *);
u_int cpu_dump(int (*)(dev_t, daddr_t, caddr_t, size_t), daddr_t *);
u_int cpu_dumpsize(void);
void dumpconf(void);
void dumpsys(void);
unsigned int cpu_rnd_messybits(void);

static inline u_long
intr_disable(void)
{
        return (u_long)_cpu_intr_suspend();
}

static inline void
intr_restore(u_long s)
{
        _cpu_intr_resume((int)s);
}
#endif /* _KERNEL */
#endif /* !_SH_CPU_H_ */