#include <sys/param.h>
#include <sys/types.h>
#include <sys/sysmacros.h>
#include <sys/systm.h>
#include <sys/signal.h>
#include <sys/errno.h>
#include <sys/fault.h>
#include <sys/syscall.h>
#include <sys/cpuvar.h>
#include <sys/sysi86.h>
#include <sys/psw.h>
#include <sys/cred.h>
#include <sys/policy.h>
#include <sys/thread.h>
#include <sys/debug.h>
#include <sys/ontrap.h>
#include <sys/privregs.h>
#include <sys/x86_archext.h>
#include <sys/vmem.h>
#include <sys/kmem.h>
#include <sys/mman.h>
#include <sys/archsystm.h>
#include <vm/hat.h>
#include <vm/as.h>
#include <vm/seg.h>
#include <vm/seg_kmem.h>
#include <vm/faultcode.h>
#include <sys/fp.h>
#include <sys/cmn_err.h>
#include <sys/segments.h>
#include <sys/clock.h>
#include <vm/hat_i86.h>
#if defined(__xpv)
#include <sys/hypervisor.h>
#include <sys/note.h>
#endif
static void ldt_alloc(proc_t *, uint_t);
static void ldt_free(proc_t *);
static void ldt_dup(proc_t *, proc_t *);
static void ldt_grow(proc_t *, uint_t);
int
sysi86(short cmd, uintptr_t arg1, uintptr_t arg2, uintptr_t arg3)
{
struct ssd ssd;
int error = 0;
int c;
proc_t *pp = curproc;
switch (cmd) {
case SI86V86:
if (arg1 == V86SC_IOPL) {
struct regs *rp = lwptoregs(ttolwp(curthread));
greg_t oldpl = rp->r_ps & PS_IOPL;
greg_t newpl = arg2 & PS_IOPL;
if (newpl > oldpl && (error =
secpolicy_sys_config(CRED(), B_FALSE)) != 0)
return (set_errno(error));
#if defined(__xpv)
const struct ctxop_template xen_tpl = {
.ct_rev = CTXOP_TPL_REV,
.ct_save = xen_disable_user_iopl,
.ct_restore = xen_enable_user_iopl,
.ct_exit = xen_disable_user_iopl,
};
struct ctxop *ctx;
ctx = ctxop_allocate(&xen_tpl, NULL);
kpreempt_disable();
ctxop_attach(curthread, ctx);
xen_enable_user_iopl(NULL);
kpreempt_enable();
#else
rp->r_ps ^= oldpl ^ newpl;
#endif
} else
error = EINVAL;
break;
case SI86DSCR:
if (curthread != pp->p_agenttp && !holdlwps(SHOLDFORK1)) {
error = EINTR;
break;
}
if (get_udatamodel() == DATAMODEL_LP64) {
error = EINVAL;
break;
}
if (copyin((caddr_t)arg1, &ssd, sizeof (ssd)) < 0) {
error = EFAULT;
break;
}
error = setdscr(&ssd);
mutex_enter(&pp->p_lock);
if (curthread != pp->p_agenttp)
continuelwps(pp);
mutex_exit(&pp->p_lock);
break;
case SI86FPHW:
c = fp_kind & 0xff;
if (suword32((void *)arg1, c) == -1)
error = EFAULT;
break;
case SI86FPSTART:
c = fp_kind & 0xff;
if (suword32((void *)arg1, c) == -1) {
error = EFAULT;
break;
}
fpsetcw((uint16_t)arg2, (uint32_t)arg3);
return ((fp_kind & __FP_SSE) ? 1 : 0);
case WTODC:
if ((error = secpolicy_settime(CRED())) == 0) {
timestruc_t ts;
mutex_enter(&tod_lock);
gethrestime(&ts);
tod_set(ts);
mutex_exit(&tod_lock);
}
break;
#define ONEWEEK (7 * 24 * 60 * 60)
case SGMTL:
if ((error = secpolicy_settime(CRED())) == 0) {
int newlag = (int)arg1;
#ifdef _SYSCALL32_IMPL
if (get_udatamodel() == DATAMODEL_NATIVE &&
(long)newlag != (long)arg1) {
error = EOVERFLOW;
} else
#endif
if (newlag >= -ONEWEEK && newlag <= ONEWEEK)
sgmtl(newlag);
else
error = EOVERFLOW;
}
break;
case GGMTL:
if (get_udatamodel() == DATAMODEL_NATIVE) {
if (sulword((void *)arg1, ggmtl()) == -1)
error = EFAULT;
#ifdef _SYSCALL32_IMPL
} else {
time_t gmtl;
if ((gmtl = ggmtl()) > INT32_MAX) {
error = EOVERFLOW;
} else if (suword32((void *)arg1, (int32_t)gmtl) == -1)
error = EFAULT;
#endif
}
break;
case RTCSYNC:
if ((error = secpolicy_settime(CRED())) == 0)
rtcsync();
break;
default:
error = EINVAL;
break;
}
return (error == 0 ? 0 : set_errno(error));
}
void
usd_to_ssd(user_desc_t *usd, struct ssd *ssd, selector_t sel)
{
ssd->bo = USEGD_GETBASE(usd);
ssd->ls = USEGD_GETLIMIT(usd);
ssd->sel = sel;
ssd->acc1 = usd->usd_type;
ssd->acc1 |= usd->usd_dpl << 5;
ssd->acc1 |= usd->usd_p << (5 + 2);
ssd->acc2 = usd->usd_avl;
ssd->acc2 |= usd->usd_long << 1;
ssd->acc2 |= usd->usd_def32 << (1 + 1);
ssd->acc2 |= usd->usd_gran << (1 + 1 + 1);
}
static void
ssd_to_usd(struct ssd *ssd, user_desc_t *usd)
{
ASSERT(bcmp(usd, &null_udesc, sizeof (*usd)) == 0);
USEGD_SETBASE(usd, ssd->bo);
USEGD_SETLIMIT(usd, ssd->ls);
usd->usd_type = ssd->acc1 | SDT_A;
usd->usd_dpl = ssd->acc1 >> 5;
usd->usd_p = ssd->acc1 >> (5 + 2);
ASSERT(usd->usd_type >= SDT_MEMRO);
ASSERT(usd->usd_dpl == SEL_UPL);
usd->usd_long = 0;
usd->usd_avl = ssd->acc2;
usd->usd_def32 = ssd->acc2 >> (1 + 1);
usd->usd_gran = ssd->acc2 >> (1 + 1 + 1);
}
static void
ldt_load(void)
{
#if defined(__xpv)
xen_set_ldt(curproc->p_ldt, curproc->p_ldtlimit + 1);
#else
size_t len;
system_desc_t desc;
len = (curproc->p_ldtlimit + 1) * sizeof (user_desc_t);
bcopy(curproc->p_ldt, CPU->cpu_m.mcpu_ldt, len);
CPU->cpu_m.mcpu_ldt_len = len;
set_syssegd(&desc, CPU->cpu_m.mcpu_ldt, len - 1, SDT_SYSLDT, SEL_KPL);
*((system_desc_t *)&CPU->cpu_gdt[GDT_LDT]) = desc;
wr_ldtr(ULDT_SEL);
#endif
}
void
ldt_unload(void)
{
#if defined(__xpv)
xen_set_ldt(NULL, 0);
#else
*((system_desc_t *)&CPU->cpu_gdt[GDT_LDT]) = null_sdesc;
wr_ldtr(0);
bzero(CPU->cpu_m.mcpu_ldt, CPU->cpu_m.mcpu_ldt_len);
CPU->cpu_m.mcpu_ldt_len = 0;
#endif
}
static void
ldt_savectx(proc_t *p)
{
ASSERT(p->p_ldt != NULL);
ASSERT(p == curproc);
reset_sregs();
ldt_unload();
cpu_fast_syscall_enable();
}
static void
ldt_restorectx(proc_t *p)
{
ASSERT(p->p_ldt != NULL);
ASSERT(p == curproc);
ldt_load();
cpu_fast_syscall_disable();
}
static void
ldt_freectx(proc_t *p, int isexec)
{
ASSERT(p->p_ldt != NULL);
ASSERT(p == curproc);
kpreempt_disable();
ldt_free(p);
cpu_fast_syscall_enable();
kpreempt_enable();
}
static void
ldt_installctx(proc_t *p, proc_t *cp)
{
proc_t *targ = p;
kthread_t *t;
if (cp != NULL) {
targ = cp;
ldt_dup(p, cp);
}
ASSERT(removepctx(targ, targ, ldt_savectx, ldt_restorectx,
ldt_installctx, ldt_savectx, ldt_freectx) == 0);
installpctx(targ, targ, ldt_savectx, ldt_restorectx,
ldt_installctx, ldt_savectx, ldt_freectx);
mutex_enter(&targ->p_lock);
t = targ->p_tlist;
do {
t->t_post_sys = 1;
} while ((t = t->t_forw) != targ->p_tlist);
mutex_exit(&targ->p_lock);
}
int
setdscr(struct ssd *ssd)
{
ushort_t seli;
user_desc_t *ldp;
user_desc_t ndesc;
proc_t *pp = curproc;
int rc = 0;
if (!SELISLDT(ssd->sel) || !SELISUPL(ssd->sel))
return (EINVAL);
seli = SELTOIDX(ssd->sel);
if (seli >= MAXNLDT || seli < LDT_UDBASE)
return (EINVAL);
ndesc = null_udesc;
mutex_enter(&pp->p_ldtlock);
if (pp->p_ldt == NULL) {
ldt_alloc(pp, seli);
kpreempt_disable();
ldt_installctx(pp, NULL);
cpu_fast_syscall_disable();
ASSERT(curthread->t_post_sys != 0);
kpreempt_enable();
} else if (seli > pp->p_ldtlimit) {
ASSERT(pp->p_pctx != NULL);
ldt_grow(pp, seli);
}
ASSERT(seli <= pp->p_ldtlimit);
ldp = &pp->p_ldt[seli];
if (SI86SSD_PRES(ssd) == 0) {
kthread_t *t;
int bad = 0;
ASSERT(pp->p_lwprcnt == 1);
mutex_enter(&pp->p_lock);
t = pp->p_tlist;
do {
klwp_t *lwp = ttolwp(t);
struct regs *rp = lwp->lwp_regs;
pcb_t *pcb = &lwp->lwp_pcb;
if (ssd->sel == rp->r_cs || ssd->sel == rp->r_ss) {
bad = 1;
break;
}
if (PCB_NEED_UPDATE_SEGS(pcb)) {
if (ssd->sel == pcb->pcb_ds ||
ssd->sel == pcb->pcb_es ||
ssd->sel == pcb->pcb_fs ||
ssd->sel == pcb->pcb_gs) {
bad = 1;
break;
}
} else {
if (ssd->sel == rp->r_ds ||
ssd->sel == rp->r_es ||
ssd->sel == rp->r_fs ||
ssd->sel == rp->r_gs) {
bad = 1;
break;
}
}
} while ((t = t->t_forw) != pp->p_tlist);
mutex_exit(&pp->p_lock);
if (bad) {
mutex_exit(&pp->p_ldtlock);
return (EBUSY);
}
}
if (ssd->acc1 == 0) {
rc = ldt_update_segd(ldp, &null_udesc);
kpreempt_disable();
ldt_load();
kpreempt_enable();
mutex_exit(&pp->p_ldtlock);
return (rc);
}
if (SI86SSD_DPL(ssd) != SEL_UPL) {
mutex_exit(&pp->p_ldtlock);
return (EINVAL);
}
if (SI86SSD_ISUSEG(ssd) && ((SI86SSD_TYPE(ssd) >> 3) & 1) == 1 &&
SI86SSD_ISLONG(ssd)) {
mutex_exit(&pp->p_ldtlock);
return (EINVAL);
}
if (SI86SSD_ISUSEG(ssd)) {
ssd_to_usd(ssd, &ndesc);
rc = ldt_update_segd(ldp, &ndesc);
kpreempt_disable();
ldt_load();
kpreempt_enable();
mutex_exit(&pp->p_ldtlock);
return (rc);
}
mutex_exit(&pp->p_ldtlock);
return (EINVAL);
}
static void
ldt_alloc(proc_t *pp, uint_t seli)
{
user_desc_t *ldt;
size_t ldtsz;
uint_t nsels;
ASSERT(MUTEX_HELD(&pp->p_ldtlock));
ASSERT(pp->p_ldt == NULL);
ASSERT(pp->p_ldtlimit == 0);
ldtsz = P2ROUNDUP((seli + 1) * sizeof (user_desc_t), PAGESIZE);
nsels = ldtsz / sizeof (user_desc_t);
ASSERT(nsels >= MINNLDT && nsels <= MAXNLDT);
ldt = kmem_zalloc(ldtsz, KM_SLEEP);
ASSERT(IS_P2ALIGNED(ldt, PAGESIZE));
#if defined(__xpv)
if (xen_ldt_setprot(ldt, ldtsz, PROT_READ))
panic("ldt_alloc:xen_ldt_setprot(PROT_READ) failed");
#endif
pp->p_ldt = ldt;
pp->p_ldtlimit = nsels - 1;
}
static void
ldt_free(proc_t *pp)
{
user_desc_t *ldt;
size_t ldtsz;
ASSERT(pp->p_ldt != NULL);
mutex_enter(&pp->p_ldtlock);
ldt = pp->p_ldt;
ldtsz = (pp->p_ldtlimit + 1) * sizeof (user_desc_t);
ASSERT(IS_P2ALIGNED(ldtsz, PAGESIZE));
pp->p_ldt = NULL;
pp->p_ldtlimit = 0;
mutex_exit(&pp->p_ldtlock);
if (pp == curproc) {
kpreempt_disable();
ldt_unload();
kpreempt_enable();
}
#if defined(__xpv)
if (xen_ldt_setprot(ldt, ldtsz, PROT_READ | PROT_WRITE))
panic("ldt_free:xen_ldt_setprot(PROT_READ|PROT_WRITE) failed");
#endif
kmem_free(ldt, ldtsz);
}
static void
ldt_dup(proc_t *pp, proc_t *cp)
{
size_t ldtsz;
ASSERT(pp->p_ldt != NULL);
ASSERT(cp != curproc);
mutex_enter(&pp->p_ldtlock);
mutex_enter(&cp->p_ldtlock);
ldtsz = (pp->p_ldtlimit + 1) * sizeof (user_desc_t);
ldt_alloc(cp, pp->p_ldtlimit);
#if defined(__xpv)
if (xen_ldt_setprot(cp->p_ldt, ldtsz, PROT_READ | PROT_WRITE))
panic("ldt_dup:xen_ldt_setprot(PROT_READ|PROT_WRITE) failed");
#endif
bcopy(pp->p_ldt, cp->p_ldt, ldtsz);
#if defined(__xpv)
if (xen_ldt_setprot(cp->p_ldt, ldtsz, PROT_READ))
panic("ldt_dup:xen_ldt_setprot(PROT_READ) failed");
#endif
mutex_exit(&cp->p_ldtlock);
mutex_exit(&pp->p_ldtlock);
}
static void
ldt_grow(proc_t *pp, uint_t seli)
{
user_desc_t *oldt, *nldt;
uint_t nsels;
size_t oldtsz, nldtsz;
ASSERT(MUTEX_HELD(&pp->p_ldtlock));
ASSERT(pp->p_ldt != NULL);
ASSERT(pp->p_ldtlimit != 0);
nldtsz = P2ROUNDUP((seli + 1) * sizeof (user_desc_t), PAGESIZE);
nsels = nldtsz / sizeof (user_desc_t);
ASSERT(nsels >= MINNLDT && nsels <= MAXNLDT);
ASSERT(nsels > pp->p_ldtlimit);
oldt = pp->p_ldt;
oldtsz = (pp->p_ldtlimit + 1) * sizeof (user_desc_t);
nldt = kmem_zalloc(nldtsz, KM_SLEEP);
ASSERT(IS_P2ALIGNED(nldt, PAGESIZE));
bcopy(oldt, nldt, oldtsz);
kpreempt_disable();
ldt_unload();
kpreempt_enable();
#if defined(__xpv)
if (xen_ldt_setprot(oldt, oldtsz, PROT_READ | PROT_WRITE))
panic("ldt_grow:xen_ldt_setprot(PROT_READ|PROT_WRITE) failed");
if (xen_ldt_setprot(nldt, nldtsz, PROT_READ))
panic("ldt_grow:xen_ldt_setprot(PROT_READ) failed");
#endif
pp->p_ldt = nldt;
pp->p_ldtlimit = nsels - 1;
kmem_free(oldt, oldtsz);
}