#include <sys/systm.h>
#include <sys/archsystm.h>
#include <sys/machsystm.h>
#include <sys/cpuvar.h>
#include <sys/intreg.h>
#include <sys/x_call.h>
#include <sys/cmn_err.h>
#include <sys/membar.h>
#include <sys/disp.h>
#include <sys/debug.h>
#include <sys/privregs.h>
#include <sys/xc_impl.h>
#include <sys/ivintr.h>
#include <sys/dmv.h>
#include <sys/sysmacros.h>
#ifdef TRAPTRACE
uint_t x_dstat[NCPU][XC_LOOP_EXIT+1];
uint_t x_rstat[NCPU][4];
#endif
static uint64_t xc_serv_inum;
static uint64_t xc_loop_inum;
kmutex_t xc_sys_mutex;
int xc_spl_enter[NCPU];
static int xc_holder = -1;
static struct xc_mbox {
xcfunc_t *xc_func;
uint64_t xc_arg1;
uint64_t xc_arg2;
cpuset_t xc_cpuset;
volatile uint_t xc_state;
} xc_mbox[NCPU];
uint64_t xc_tick_limit;
uint64_t xc_tick_limit_scale = 1;
uint64_t xc_tick_jump_limit;
uint64_t xc_sync_tick_limit;
uint64_t xc_mondo_time_limit;
uint64_t xc_func_time_limit;
uint64_t xc_scale = 1;
uint64_t xc_mondo_multiplier = 10;
uint_t sendmondo_in_recover;
void init_mondo(xcfunc_t *func, uint64_t arg1, uint64_t arg2);
void send_one_mondo(int cpuid);
void send_mondo_set(cpuset_t set);
static int
xc_func_timeout_adj(cpu_setup_t what, int cpuid)
{
uint64_t freq = cpunodes[cpuid].clock_freq;
switch (what) {
case CPU_ON:
case CPU_INIT:
case CPU_CONFIG:
case CPU_CPUPART_IN:
if (freq * xc_scale > xc_mondo_time_limit) {
xc_mondo_time_limit = freq * xc_scale;
xc_func_time_limit = xc_mondo_time_limit *
xc_mondo_multiplier;
}
break;
case CPU_OFF:
case CPU_UNCONFIG:
case CPU_CPUPART_OUT:
default:
break;
}
return (0);
}
void
xc_init(void)
{
int pix;
uint64_t maxfreq = 0;
mutex_init(&xc_sys_mutex, NULL, MUTEX_SPIN,
(void *)ipltospl(XCALL_PIL));
#ifdef TRAPTRACE
for (pix = 0; pix < NCPU; pix++) {
XC_STAT_INIT(pix);
}
#endif
xc_serv_inum = add_softintr(XCALL_PIL, (softintrfunc)xc_serv, 0,
SOFTINT_MT);
xc_loop_inum = add_softintr(XCALL_PIL, (softintrfunc)xc_loop, 0,
SOFTINT_MT);
xc_tick_limit =
((uint64_t)sys_tick_freq * XC_SEND_MONDO_MSEC) / 1000;
xc_tick_jump_limit = xc_tick_limit / 32;
xc_tick_limit *= xc_tick_limit_scale;
xc_sync_tick_limit = xc_tick_limit;
for (pix = 0; pix < NCPU; pix++) {
maxfreq = MAX(cpunodes[pix].clock_freq, maxfreq);
}
xc_mondo_time_limit = maxfreq * xc_scale;
register_cpu_setup_func((cpu_setup_func_t *)xc_func_timeout_adj, NULL);
xc_func_time_limit = xc_mondo_time_limit * xc_mondo_multiplier;
}
int
spl_xcall(void)
{
return (splr(XCALL_PIL));
}
void
xt_one(int cix, xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
if (!CPU_IN_SET(cpu_ready_set, cix)) {
return;
}
xt_one_unchecked(cix, func, arg1, arg2);
}
void
xt_one_unchecked(int cix, xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
cpuset_t tset;
ASSERT(!DMV_IS_DMV(func));
ASSERT((uintptr_t)func >= KERNELBASE);
CPUSET_ZERO(tset);
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, cix);
if (cix == lcx) {
send_self_xcall(CPU, arg1, arg2, func);
XC_STAT_INC(x_dstat[lcx][XT_ONE_SELF]);
XC_TRACE(XT_ONE_SELF, &tset, func, arg1, arg2);
} else {
XC_TRACE(XT_ONE_OTHER, &tset, func, arg1, arg2);
init_mondo(func, arg1, arg2);
send_one_mondo(cix);
XC_STAT_INC(x_dstat[lcx][XT_ONE_OTHER]);
}
XC_SPL_EXIT(lcx, opl);
}
void
xt_some(cpuset_t cpuset, xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
cpuset_t xc_cpuset, tset;
ASSERT(!DMV_IS_DMV(func));
ASSERT((uintptr_t)func >= KERNELBASE);
CPUSET_ZERO(tset);
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, lcx);
xc_cpuset = cpu_ready_set;
CPUSET_AND(xc_cpuset, cpuset);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_SPL_EXIT(lcx, opl);
return;
}
if (CPU_IN_SET(xc_cpuset, lcx)) {
send_self_xcall(CPU, arg1, arg2, func);
XC_STAT_INC(x_dstat[lcx][XT_SOME_SELF]);
XC_TRACE(XT_SOME_SELF, &tset, func, arg1, arg2);
CPUSET_DEL(xc_cpuset, lcx);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_SPL_EXIT(lcx, opl);
return;
}
}
XC_TRACE(XT_SOME_OTHER, &xc_cpuset, func, arg1, arg2);
init_mondo(func, arg1, arg2);
send_mondo_set(xc_cpuset);
XC_STAT_INC(x_dstat[lcx][XT_SOME_OTHER]);
XC_SPL_EXIT(lcx, opl);
}
void
xt_all(xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
cpuset_t xc_cpuset, tset;
ASSERT(!DMV_IS_DMV(func));
ASSERT((uintptr_t)func >= KERNELBASE);
CPUSET_ZERO(tset);
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, lcx);
if (CPU_IN_SET(cpu_ready_set, lcx))
send_self_xcall(CPU, arg1, arg2, func);
XC_TRACE(XT_ALL_OTHER, &cpu_ready_set, func, arg1, arg2);
xc_cpuset = cpu_ready_set;
CPUSET_DEL(xc_cpuset, lcx);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_STAT_INC(x_dstat[lcx][XT_ALL_SELF]);
XC_TRACE(XT_ALL_SELF, &tset, func, arg1, arg2);
XC_SPL_EXIT(lcx, opl);
return;
}
init_mondo(func, arg1, arg2);
send_mondo_set(xc_cpuset);
XC_STAT_INC(x_dstat[lcx][XT_ALL_OTHER]);
XC_SPL_EXIT(lcx, opl);
}
void
xc_one(int cix, xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
uint64_t loop_cnt = 0;
cpuset_t tset;
int first_time = 1;
if (!CPU_IN_SET(cpu_ready_set, cix))
return;
ASSERT((uintptr_t)func > KERNELBASE);
ASSERT(((uintptr_t)func % PC_ALIGN) == 0);
CPUSET_ZERO(tset);
kpreempt_disable();
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, cix);
if (cix == lcx) {
XC_TRACE(XC_ONE_SELF, &tset, func, arg1, arg2);
(*func)(arg1, arg2);
XC_STAT_INC(x_dstat[lcx][XC_ONE_SELF]);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
if (xc_holder == lcx) {
ASSERT(MUTEX_HELD(&xc_sys_mutex));
ASSERT(CPU_IN_SET(xc_mbox[lcx].xc_cpuset, lcx));
ASSERT(CPU_IN_SET(xc_mbox[cix].xc_cpuset, cix));
ASSERT(xc_mbox[cix].xc_state == XC_WAIT);
XC_TRACE(XC_ONE_OTHER_H, &tset, func, arg1, arg2);
XC_SETUP(cix, func, arg1, arg2);
membar_stld();
while (xc_mbox[cix].xc_state != XC_WAIT) {
if (loop_cnt++ > xc_func_time_limit) {
if (sendmondo_in_recover) {
drv_usecwait(1);
loop_cnt = 0;
continue;
}
cmn_err(CE_PANIC, "xc_one() timeout, "
"xc_state[%d] != XC_WAIT", cix);
}
}
XC_STAT_INC(x_dstat[lcx][XC_ONE_OTHER_H]);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
XC_SPL_EXIT(lcx, opl);
ASSERT(getpil() < XCALL_PIL);
mutex_enter(&xc_sys_mutex);
xc_spl_enter[lcx] = 1;
ASSERT(xc_mbox[cix].xc_state == XC_IDLE);
XC_TRACE(XC_ONE_OTHER, &tset, func, arg1, arg2);
XC_SETUP(cix, func, arg1, arg2);
init_mondo(setsoftint_tl1, xc_serv_inum, 0);
send_one_mondo(cix);
xc_spl_enter[lcx] = 0;
while (xc_mbox[cix].xc_state != XC_IDLE) {
if (loop_cnt++ > xc_func_time_limit) {
if (sendmondo_in_recover) {
drv_usecwait(1);
loop_cnt = 0;
continue;
}
if (first_time) {
XT_SYNC_ONE(cix);
first_time = 0;
loop_cnt = 0;
continue;
}
cmn_err(CE_PANIC, "xc_one() timeout, "
"xc_state[%d] != XC_IDLE", cix);
}
}
XC_STAT_INC(x_dstat[lcx][XC_ONE_OTHER]);
mutex_exit(&xc_sys_mutex);
kpreempt_enable();
}
void
xc_some(cpuset_t cpuset, xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
cpuset_t xc_cpuset, tset;
ASSERT((uintptr_t)func > KERNELBASE);
ASSERT(((uintptr_t)func % PC_ALIGN) == 0);
CPUSET_ZERO(tset);
kpreempt_disable();
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, lcx);
xc_cpuset = cpu_ready_set;
CPUSET_AND(xc_cpuset, cpuset);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
if (CPU_IN_SET(xc_cpuset, lcx)) {
(*func)(arg1, arg2);
CPUSET_DEL(xc_cpuset, lcx);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_STAT_INC(x_dstat[lcx][XC_SOME_SELF]);
XC_TRACE(XC_SOME_SELF, &tset, func, arg1, arg2);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
}
if (xc_holder == lcx) {
cpuset_t mset = xc_mbox[lcx].xc_cpuset;
CPUSET_AND(mset, cpuset);
ASSERT(MUTEX_HELD(&xc_sys_mutex));
ASSERT(CPUSET_ISEQUAL(mset, cpuset));
SEND_MBOX_ONLY(xc_cpuset, func, arg1, arg2, lcx, XC_WAIT);
WAIT_MBOX_DONE(xc_cpuset, lcx, XC_WAIT, 0);
XC_STAT_INC(x_dstat[lcx][XC_SOME_OTHER_H]);
XC_TRACE(XC_SOME_OTHER_H, &xc_cpuset, func, arg1, arg2);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
XC_SPL_EXIT(lcx, opl);
ASSERT(getpil() < XCALL_PIL);
mutex_enter(&xc_sys_mutex);
xc_spl_enter[lcx] = 1;
XC_TRACE(XC_SOME_OTHER, &xc_cpuset, func, arg1, arg2);
init_mondo(setsoftint_tl1, xc_serv_inum, 0);
SEND_MBOX_MONDO(xc_cpuset, func, arg1, arg2, XC_IDLE);
WAIT_MBOX_DONE(xc_cpuset, lcx, XC_IDLE, 1);
xc_spl_enter[lcx] = 0;
XC_STAT_INC(x_dstat[lcx][XC_SOME_OTHER]);
mutex_exit(&xc_sys_mutex);
kpreempt_enable();
}
void
xc_all(xcfunc_t *func, uint64_t arg1, uint64_t arg2)
{
int lcx;
int opl;
cpuset_t xc_cpuset, tset;
ASSERT((uintptr_t)func > KERNELBASE);
ASSERT(((uintptr_t)func % PC_ALIGN) == 0);
CPUSET_ZERO(tset);
kpreempt_disable();
XC_SPL_ENTER(lcx, opl);
CPUSET_ADD(tset, lcx);
(*func)(arg1, arg2);
xc_cpuset = cpu_ready_set;
CPUSET_DEL(xc_cpuset, lcx);
if (CPUSET_ISNULL(xc_cpuset)) {
XC_STAT_INC(x_dstat[lcx][XC_ALL_SELF]);
XC_TRACE(XC_ALL_SELF, &tset, func, arg1, arg2);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
if (xc_holder == lcx) {
cpuset_t mset = xc_mbox[lcx].xc_cpuset;
CPUSET_AND(mset, xc_cpuset);
ASSERT(MUTEX_HELD(&xc_sys_mutex));
ASSERT(CPUSET_ISEQUAL(mset, xc_cpuset));
XC_TRACE(XC_ALL_OTHER_H, &xc_cpuset, func, arg1, arg2);
SEND_MBOX_ONLY(xc_cpuset, func, arg1, arg2, lcx, XC_WAIT);
WAIT_MBOX_DONE(xc_cpuset, lcx, XC_WAIT, 0);
XC_STAT_INC(x_dstat[lcx][XC_ALL_OTHER_H]);
XC_SPL_EXIT(lcx, opl);
kpreempt_enable();
return;
}
XC_SPL_EXIT(lcx, opl);
ASSERT(getpil() < XCALL_PIL);
mutex_enter(&xc_sys_mutex);
xc_spl_enter[lcx] = 1;
XC_TRACE(XC_ALL_OTHER, &xc_cpuset, func, arg1, arg2);
init_mondo(setsoftint_tl1, xc_serv_inum, 0);
SEND_MBOX_MONDO(xc_cpuset, func, arg1, arg2, XC_IDLE);
WAIT_MBOX_DONE(xc_cpuset, lcx, XC_IDLE, 1);
xc_spl_enter[lcx] = 0;
XC_STAT_INC(x_dstat[lcx][XC_ALL_OTHER]);
mutex_exit(&xc_sys_mutex);
kpreempt_enable();
}
void
xc_attention(cpuset_t cpuset)
{
int pix, lcx;
cpuset_t xc_cpuset, tmpset;
cpuset_t recv_cpuset;
uint64_t loop_cnt = 0;
int first_time = 1;
CPUSET_ZERO(recv_cpuset);
ASSERT(getpil() < XCALL_PIL);
mutex_enter(&xc_sys_mutex);
lcx = (int)(CPU->cpu_id);
ASSERT(x_dstat[lcx][XC_ATTENTION] ==
x_dstat[lcx][XC_DISMISSED]);
ASSERT(xc_holder == -1);
xc_mbox[lcx].xc_cpuset = cpuset;
xc_holder = lcx;
xc_cpuset = cpu_ready_set;
CPUSET_AND(xc_cpuset, cpuset);
CPUSET_DEL(xc_cpuset, lcx);
XC_STAT_INC(x_dstat[lcx][XC_ATTENTION]);
XC_TRACE(XC_ATTENTION, &xc_cpuset, NULL, 0, 0);
if (CPUSET_ISNULL(xc_cpuset))
return;
xc_spl_enter[lcx] = 1;
init_mondo(setsoftint_tl1, xc_loop_inum, 0);
SEND_MBOX_MONDO_XC_ENTER(xc_cpuset);
xc_spl_enter[lcx] = 0;
while (!CPUSET_ISEQUAL(recv_cpuset, xc_cpuset)) {
tmpset = xc_cpuset;
for (pix = 0; pix < NCPU; pix++) {
if (CPU_IN_SET(tmpset, pix)) {
if (xc_mbox[pix].xc_state == XC_WAIT) {
CPUSET_ADD(recv_cpuset, pix);
}
CPUSET_DEL(tmpset, pix);
if (CPUSET_ISNULL(tmpset)) {
break;
}
}
}
if (loop_cnt++ > xc_mondo_time_limit) {
if (sendmondo_in_recover) {
drv_usecwait(1);
loop_cnt = 0;
continue;
}
if (first_time) {
XT_SYNC_SOME(xc_cpuset);
first_time = 0;
loop_cnt = 0;
continue;
}
cmn_err(CE_PANIC, "xc_attention() timeout");
}
}
}
void
xc_dismissed(cpuset_t cpuset)
{
int pix;
int lcx = (int)(CPU->cpu_id);
cpuset_t xc_cpuset, tmpset;
cpuset_t recv_cpuset;
uint64_t loop_cnt = 0;
ASSERT(lcx == xc_holder);
ASSERT(CPUSET_ISEQUAL(xc_mbox[lcx].xc_cpuset, cpuset));
ASSERT(getpil() >= XCALL_PIL);
CPUSET_ZERO(xc_mbox[lcx].xc_cpuset);
CPUSET_ZERO(recv_cpuset);
membar_stld();
XC_STAT_INC(x_dstat[lcx][XC_DISMISSED]);
ASSERT(x_dstat[lcx][XC_DISMISSED] == x_dstat[lcx][XC_ATTENTION]);
xc_cpuset = cpu_ready_set;
CPUSET_AND(xc_cpuset, cpuset);
CPUSET_DEL(xc_cpuset, lcx);
XC_TRACE(XC_DISMISSED, &xc_cpuset, NULL, 0, 0);
if (CPUSET_ISNULL(xc_cpuset)) {
xc_holder = -1;
mutex_exit(&xc_sys_mutex);
return;
}
tmpset = xc_cpuset;
for (pix = 0; pix < NCPU; pix++) {
if (CPU_IN_SET(tmpset, pix)) {
xc_mbox[pix].xc_state = XC_EXIT;
membar_stld();
CPUSET_DEL(tmpset, pix);
if (CPUSET_ISNULL(tmpset)) {
break;
}
}
}
while (!CPUSET_ISEQUAL(recv_cpuset, xc_cpuset)) {
tmpset = xc_cpuset;
for (pix = 0; pix < NCPU; pix++) {
if (CPU_IN_SET(tmpset, pix)) {
if (xc_mbox[pix].xc_state == XC_IDLE) {
CPUSET_ADD(recv_cpuset, pix);
}
CPUSET_DEL(tmpset, pix);
if (CPUSET_ISNULL(tmpset)) {
break;
}
}
}
if (loop_cnt++ > xc_func_time_limit) {
if (sendmondo_in_recover) {
drv_usecwait(1);
loop_cnt = 0;
continue;
}
cmn_err(CE_PANIC, "xc_dismissed() timeout");
}
}
xc_holder = -1;
mutex_exit(&xc_sys_mutex);
}
uint_t
xc_serv(void)
{
int lcx = (int)(CPU->cpu_id);
struct xc_mbox *xmp;
xcfunc_t *func;
uint64_t arg1, arg2;
cpuset_t tset;
ASSERT(getpil() == XCALL_PIL);
CPUSET_ZERO(tset);
CPUSET_ADD(tset, lcx);
flush_windows();
xmp = &xc_mbox[lcx];
ASSERT(lcx != xc_holder);
ASSERT(xmp->xc_state == XC_DOIT);
func = xmp->xc_func;
XC_TRACE(XC_SERV, &tset, func, xmp->xc_arg1, xmp->xc_arg2);
if (func != NULL) {
arg1 = xmp->xc_arg1;
arg2 = xmp->xc_arg2;
(*func)(arg1, arg2);
}
XC_STAT_INC(x_rstat[lcx][XC_SERV]);
XC_TRACE(XC_SERV, &tset, func, arg1, arg2);
xmp->xc_state = XC_IDLE;
membar_stld();
return (1);
}
uint_t xc_loop_panic = 0;
uint_t
xc_loop(void)
{
int lcx = (int)(CPU->cpu_id);
struct xc_mbox *xmp;
xcfunc_t *func;
uint64_t arg1, arg2;
uint64_t loop_cnt = 0;
cpuset_t tset;
ASSERT(getpil() == XCALL_PIL);
CPUSET_ZERO(tset);
flush_windows();
xmp = &xc_mbox[lcx];
ASSERT(lcx != xc_holder);
ASSERT(xmp->xc_state == XC_ENTER);
xmp->xc_state = XC_WAIT;
CPUSET_ADD(tset, lcx);
membar_stld();
XC_STAT_INC(x_rstat[lcx][XC_LOOP]);
XC_TRACE(XC_LOOP_ENTER, &tset, NULL, 0, 0);
while (xmp->xc_state != XC_EXIT) {
if (xmp->xc_state == XC_DOIT) {
func = xmp->xc_func;
arg1 = xmp->xc_arg1;
arg2 = xmp->xc_arg2;
XC_TRACE(XC_LOOP_DOIT, &tset, func, arg1, arg2);
if (func != NULL)
(*func)(arg1, arg2);
xmp->xc_state = XC_WAIT;
membar_stld();
loop_cnt = 0;
} else {
loop_cnt++;
}
if (loop_cnt > xc_func_time_limit) {
if (sendmondo_in_recover) {
drv_usecwait(1);
loop_cnt = 0;
continue;
}
cmn_err(xc_loop_panic ? CE_PANIC : CE_WARN,
"xc_loop() timeout");
loop_cnt = 0;
}
}
ASSERT(xmp->xc_state == XC_EXIT);
ASSERT(xc_holder != -1);
XC_TRACE(XC_LOOP_EXIT, &tset, NULL, 0, 0);
xmp->xc_state = XC_IDLE;
membar_stld();
return (1);
}