#include <sys/types.h>
#include <sys/sysmacros.h>
#include <sys/modctl.h>
#include <sys/conf.h>
#include <sys/devops.h>
#include <sys/cmn_err.h>
#include <sys/ksynch.h>
#include <sys/kmem.h>
#include <sys/stat.h>
#include <sys/open.h>
#include <sys/file.h>
#include <sys/ddi.h>
#include <sys/sunddi.h>
#include <sys/param.h>
#include <sys/cpuvar.h>
#include <sys/disp.h>
#include <sys/hsvc.h>
#include <sys/machsystm.h>
#include <sys/hypervisor_api.h>
#include <sys/n2rng.h>
static int n2rng_attach(dev_info_t *, ddi_attach_cmd_t);
static int n2rng_detach(dev_info_t *, ddi_detach_cmd_t);
static int n2rng_suspend(n2rng_t *);
static int n2rng_resume(n2rng_t *);
static uint64_t sticks_per_usec(void);
u_longlong_t gettick(void);
static int n2rng_init_ctl(n2rng_t *);
static void n2rng_uninit_ctl(n2rng_t *);
static int n2rng_config(n2rng_t *);
static void n2rng_config_task(void * targ);
static struct dev_ops devops = {
DEVO_REV,
0,
nodev,
nulldev,
nulldev,
n2rng_attach,
n2rng_detach,
nodev,
NULL,
NULL,
ddi_power,
ddi_quiesce_not_supported,
};
static struct modldrv modldrv = {
&mod_driverops,
"N2 RNG Driver",
&devops,
};
static struct modlinkage modlinkage = {
MODREV_1,
&modldrv,
NULL
};
static void *n2rng_softstate = NULL;
static boolean_t ncs_hsvc_available = B_FALSE;
#define NVERSIONS 2
static hsvc_info_t ncs_hsvc[NVERSIONS] = {
{ HSVC_REV_1, NULL, HSVC_GROUP_RNG, 2, 0, DRIVER },
{ HSVC_REV_1, NULL, HSVC_GROUP_RNG, 1, 0, DRIVER },
};
int ncs_version_index;
int
_init(void)
{
int rv;
rv = ddi_soft_state_init(&n2rng_softstate, sizeof (n2rng_t), 1);
if (rv != 0) {
return (rv);
}
if ((rv = mod_install(&modlinkage)) != 0) {
ddi_soft_state_fini(&n2rng_softstate);
return (rv);
}
return (0);
}
int
_fini(void)
{
int rv;
rv = mod_remove(&modlinkage);
if (rv == 0) {
ddi_soft_state_fini(&n2rng_softstate);
}
return (rv);
}
int
_info(struct modinfo *modinfop)
{
return (mod_info(&modlinkage, modinfop));
}
static int
n2rng_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
{
n2rng_t *n2rng = NULL;
int instance;
int rv;
int version;
uint64_t ncs_minor_ver;
instance = ddi_get_instance(dip);
DBG1(NULL, DENTRY, "n2rng_attach called, instance %d", instance);
if (instance != 0) {
n2rng_diperror(dip, "only one instance (0) allowed");
return (DDI_FAILURE);
}
switch (cmd) {
case DDI_RESUME:
n2rng = (n2rng_t *)ddi_get_soft_state(n2rng_softstate,
instance);
if (n2rng == NULL) {
n2rng_diperror(dip, "no soft state in attach");
return (DDI_FAILURE);
}
return (n2rng_resume(n2rng));
case DDI_ATTACH:
break;
default:
return (DDI_FAILURE);
}
rv = ddi_soft_state_zalloc(n2rng_softstate, instance);
if (rv != DDI_SUCCESS) {
n2rng_diperror(dip, "unable to allocate soft state");
return (DDI_FAILURE);
}
n2rng = (n2rng_t *)ddi_get_soft_state(n2rng_softstate, instance);
ASSERT(n2rng != NULL);
n2rng->n_dip = dip;
mutex_init(&n2rng->n_lock, NULL, MUTEX_DRIVER, NULL);
n2rng->n_flags = 0;
n2rng->n_timeout_id = 0;
n2rng->n_sticks_per_usec = sticks_per_usec();
n2rng->n_binding_name = ddi_binding_name(dip);
if (strncmp(n2rng->n_binding_name, N2RNG_BINDNAME_N2,
strlen(N2RNG_BINDNAME_N2)) == 0) {
n2rng->n_binding = N2RNG_CPU_N2;
} else if (strncmp(n2rng->n_binding_name, N2RNG_BINDNAME_VF,
strlen(N2RNG_BINDNAME_VF)) == 0) {
n2rng->n_binding = N2RNG_CPU_VF;
} else if (strncmp(n2rng->n_binding_name, N2RNG_BINDNAME_KT,
strlen(N2RNG_BINDNAME_KT)) == 0) {
n2rng->n_binding = N2RNG_CPU_KT;
} else {
n2rng_diperror(dip,
"unable to determine n2rng (cpu) binding (%s)",
n2rng->n_binding_name);
goto errorexit;
}
DBG1(n2rng, DCHATTY, "n2rng_attach: n2rng->n_binding_name = %s",
n2rng->n_binding_name);
for (version = 0; version < NVERSIONS; version++) {
rv = hsvc_register(&ncs_hsvc[version], &ncs_minor_ver);
if (rv == 0)
break;
DBG4(n2rng, DCHATTY, "n2rng_attach: grp: 0x%lx, maj: %ld, "
"min: %ld, errno: %d", ncs_hsvc[version].hsvc_group,
ncs_hsvc[version].hsvc_major,
ncs_hsvc[version].hsvc_minor, rv);
}
if (version == NVERSIONS) {
for (version = 0; version < NVERSIONS; version++) {
cmn_err(CE_WARN,
"%s: cannot negotiate hypervisor services "
"group: 0x%lx major: %ld minor: %ld errno: %d",
ncs_hsvc[version].hsvc_modname,
ncs_hsvc[version].hsvc_group,
ncs_hsvc[version].hsvc_major,
ncs_hsvc[version].hsvc_minor, rv);
}
goto errorexit;
}
ncs_version_index = version;
ncs_hsvc_available = B_TRUE;
DBG2(n2rng, DATTACH, "n2rng_attach: ncs api version (%ld.%ld)",
ncs_hsvc[ncs_version_index].hsvc_major, ncs_minor_ver);
n2rng->n_hvapi_major_version = ncs_hsvc[ncs_version_index].hsvc_major;
n2rng->n_hvapi_minor_version = (uint_t)ncs_minor_ver;
if ((n2rng->n_binding != N2RNG_CPU_N2) &&
(n2rng->n_hvapi_major_version < 2)) {
cmn_err(CE_NOTE, "n2rng: Incompatible hyperviser api "
"version %d.%d detected", n2rng->n_hvapi_major_version,
n2rng->n_hvapi_minor_version);
}
if (n2rng_init_ctl(n2rng) != DDI_SUCCESS) {
cmn_err(CE_WARN, "n2rng: unable to initialize rng "
"control structures");
goto errorexit;
}
n2rng->n_taskq = ddi_taskq_create(dip, "n2rng_taskq", 1,
TASKQ_DEFAULTPRI, 0);
if (n2rng->n_taskq == NULL) {
n2rng_diperror(dip, "ddi_taskq_create() failed");
goto errorexit;
}
if (ddi_taskq_dispatch(n2rng->n_taskq, n2rng_config_task,
(void *)n2rng, DDI_SLEEP) != DDI_SUCCESS) {
n2rng_diperror(dip, "ddi_taskq_dispatch() failed");
goto errorexit;
}
return (DDI_SUCCESS);
errorexit:
if (n2rng->n_taskq != NULL) {
ddi_taskq_destroy(n2rng->n_taskq);
n2rng->n_taskq = NULL;
}
n2rng_uninit_ctl(n2rng);
(void) n2rng_uninit(n2rng);
if (ncs_hsvc_available == B_TRUE) {
(void) hsvc_unregister(&ncs_hsvc[ncs_version_index]);
ncs_hsvc_available = B_FALSE;
}
mutex_destroy(&n2rng->n_lock);
ddi_soft_state_free(n2rng_softstate, instance);
return (DDI_FAILURE);
}
static int
n2rng_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
{
int instance;
int rv;
n2rng_t *n2rng;
timeout_id_t tid;
instance = ddi_get_instance(dip);
n2rng = (n2rng_t *)ddi_get_soft_state(n2rng_softstate, instance);
if (n2rng == NULL) {
n2rng_diperror(dip, "no soft state in detach");
return (DDI_FAILURE);
}
switch (cmd) {
case DDI_SUSPEND:
return (n2rng_suspend(n2rng));
case DDI_DETACH:
break;
default:
return (DDI_FAILURE);
}
if (n2rng->n_taskq != NULL) {
ddi_taskq_destroy(n2rng->n_taskq);
n2rng->n_taskq = NULL;
}
mutex_enter(&n2rng->n_lock);
tid = n2rng->n_timeout_id;
n2rng->n_timeout_id = 0;
mutex_exit(&n2rng->n_lock);
if (tid) {
DBG1(n2rng, DCHATTY, "n2rng_detach: untimeout pending retry "
"id = %x", tid);
(void) untimeout(tid);
}
n2rng_uninit_ctl(n2rng);
rv = n2rng_uninit(n2rng) ? DDI_FAILURE : DDI_SUCCESS;
if (ncs_hsvc_available == B_TRUE) {
(void) hsvc_unregister(&ncs_hsvc[ncs_version_index]);
ncs_hsvc_available = B_FALSE;
}
mutex_destroy(&n2rng->n_lock);
ddi_soft_state_free(n2rng_softstate, instance);
return (rv);
}
static int
n2rng_suspend(n2rng_t *n2rng)
{
if (n2rng_uninit(n2rng) != DDI_SUCCESS) {
cmn_err(CE_WARN, "n2rng: unable to unregister from KCF");
return (DDI_FAILURE);
}
return (DDI_SUCCESS);
}
static int
n2rng_resume(n2rng_t *n2rng)
{
return (n2rng_config(n2rng));
}
int
n2rng_herr2kerr(uint64_t hv_errcode)
{
int s_errcode;
switch (hv_errcode) {
case H_EWOULDBLOCK:
s_errcode = EWOULDBLOCK;
break;
case H_EIO:
s_errcode = EIO;
break;
case H_EBUSY:
s_errcode = EBUSY;
break;
case H_EOK:
s_errcode = 0;
break;
case H_ENOACCESS:
s_errcode = EPERM;
break;
case H_ENORADDR:
case H_EBADALIGN:
default:
s_errcode = EINVAL;
break;
}
return (s_errcode);
}
void
cyclesleep(n2rng_t *n2rng, uint64_t delay_sticks)
{
uint64_t end_stick = gettick() + delay_sticks;
int64_t sticks_to_wait;
clock_t sys_ticks_to_wait;
clock_t usecs_to_wait;
while (1) {
sticks_to_wait = end_stick - gettick();
if (sticks_to_wait <= 0) {
return;
}
usecs_to_wait = sticks_to_wait / n2rng->n_sticks_per_usec;
sys_ticks_to_wait = drv_usectohz(usecs_to_wait);
if (sys_ticks_to_wait > 0) {
delay(sys_ticks_to_wait);
} else if (usecs_to_wait > 0) {
drv_usecwait(usecs_to_wait);
}
}
}
static void
log_internal_errors(uint64_t hverr, char *fname)
{
switch (hverr) {
case H_EBADALIGN:
cmn_err(CE_WARN,
"n2rng: internal alignment "
"problem");
break;
case H_ENORADDR:
cmn_err(CE_WARN, "n2rng: internal "
"invalid address");
break;
case H_ENOACCESS:
cmn_err(CE_WARN, "n2rng: access failure");
break;
case H_EWOULDBLOCK:
cmn_err(CE_WARN, "n2rng: hardware busy");
break;
default:
cmn_err(CE_NOTE,
"n2rng: %s "
"unexpectedly "
"returned hverr %ld", fname, hverr);
break;
}
}
int
n2rng_collect_diag_bits(n2rng_t *n2rng, int rngid,
n2rng_setup_t *collect_setupp, void *buffer, int numbytes,
n2rng_setup_t *exit_setupp, uint64_t exitstate)
{
int rv;
int override_rv = 0;
uint64_t hverr;
int i;
uint64_t tdelta;
n2rng_setup_t setupbuffer[2];
n2rng_setup_t *setupcontigp;
uint64_t setupphys;
int numchunks;
boolean_t rnglooping;
int busycount = 0;
int blockcount = 0;
if (numbytes % sizeof (uint64_t)) {
return (EINVAL);
}
if ((uint64_t)buffer % sizeof (uint64_t) != 0) {
return (EINVAL);
}
numchunks = ((numbytes / sizeof (uint64_t)) + RNG_DIAG_CHUNK_SIZE - 1)
/ RNG_DIAG_CHUNK_SIZE;
setupcontigp = &setupbuffer[
CONTIGUOUS(&setupbuffer[0], n2rng_setup_t) ? 0 : 1];
setupphys = va_to_pa(setupcontigp);
if (collect_setupp != NULL) {
*setupcontigp = *collect_setupp;
rnglooping = B_TRUE;
while (rnglooping) {
hverr = n2rng_ctl_write(n2rng, rngid, setupphys,
CTL_STATE_HEALTHCHECK,
n2rng->n_ctl_data->n_watchdog_cycles, &tdelta);
rv = n2rng_herr2kerr(hverr);
switch (hverr) {
case H_EOK:
rnglooping = B_FALSE;
break;
case H_EIO:
case H_ENOACCESS:
return (rv);
case H_EWOULDBLOCK:
if (++blockcount > RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_collect_diag_bits(1) : "
"exceeded block count of %d",
RNG_MAX_BLOCK_ATTEMPTS);
return (rv);
} else {
cyclesleep(n2rng, tdelta);
}
break;
case H_EBUSY:
if (++busycount > RNG_MAX_BUSY_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_collect_diag_bits(1): "
"exceeded busy count of %d",
RNG_MAX_BUSY_ATTEMPTS);
return (rv);
} else {
delay(RNG_RETRY_BUSY_DELAY);
}
break;
default:
log_internal_errors(hverr, "hv_rng_ctl_write");
override_rv = rv;
goto restore_state;
}
}
}
if (numbytes > 0) {
for (i = 0; i < numchunks; i++) {
size_t thisnumbytes = (i == numchunks - 1) ?
numbytes - i * (RNG_DIAG_CHUNK_SIZE *
sizeof (uint64_t)) :
RNG_DIAG_CHUNK_SIZE * sizeof (uint64_t);
rnglooping = B_TRUE;
busycount = 0;
blockcount = 0;
while (rnglooping) {
hverr = n2rng_data_read_diag(n2rng, rngid,
va_to_pa((uint64_t *)buffer +
RNG_DIAG_CHUNK_SIZE * i),
thisnumbytes, &tdelta);
rv = n2rng_herr2kerr(hverr);
switch (hverr) {
case H_EOK:
rnglooping = B_FALSE;
break;
case H_EIO:
case H_ENOACCESS:
return (rv);
case H_EWOULDBLOCK:
if (++blockcount >
RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_collect_diag_bits"
"(2): exceeded block count"
" of %d",
RNG_MAX_BLOCK_ATTEMPTS);
return (rv);
} else {
cyclesleep(n2rng, tdelta);
}
break;
default:
log_internal_errors(hverr,
"hv_rng_data_read_diag");
override_rv = rv;
goto restore_state;
}
}
}
}
restore_state:
if (exit_setupp != NULL) {
*setupcontigp = *exit_setupp;
rnglooping = B_TRUE;
busycount = 0;
blockcount = 0;
while (rnglooping) {
hverr = n2rng_ctl_write(n2rng, rngid, setupphys,
exitstate, n2rng->n_ctl_data->n_watchdog_cycles,
&tdelta);
rv = n2rng_herr2kerr(hverr);
switch (hverr) {
case H_EOK:
case H_EIO:
case H_EINVAL:
case H_ENOACCESS:
rnglooping = B_FALSE;
break;
case H_EWOULDBLOCK:
if (++blockcount > RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_collect_diag_bits(3): "
"exceeded block count of %d",
RNG_MAX_BLOCK_ATTEMPTS);
return (rv);
} else {
cyclesleep(n2rng, tdelta);
}
break;
case H_EBUSY:
if (++busycount > RNG_MAX_BUSY_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_collect_diag_bits(3): "
"exceeded busy count of %d",
RNG_MAX_BUSY_ATTEMPTS);
return (rv);
} else {
delay(RNG_RETRY_BUSY_DELAY);
}
break;
default:
rnglooping = B_FALSE;
log_internal_errors(hverr, "hv_rng_ctl_write");
break;
}
}
}
return (override_rv ? override_rv : rv);
}
int
n2rng_getentropy(n2rng_t *n2rng, void *buffer, size_t size)
{
int i, rv = 0;
uint64_t hverr;
uint64_t *buffer_w = (uint64_t *)buffer;
int num_w = size / sizeof (uint64_t);
uint64_t randval;
uint64_t randvalphys = va_to_pa(&randval);
uint64_t tdelta;
int failcount = 0;
int blockcount = 0;
boolean_t rnglooping;
for (i = 0; i < num_w; i++) {
rnglooping = B_TRUE;
while (rnglooping) {
hverr = hv_rng_data_read(randvalphys, &tdelta);
rv = n2rng_herr2kerr(hverr);
switch (hverr) {
case H_EOK:
buffer_w[i] = randval;
failcount = 0;
rnglooping = B_FALSE;
break;
case H_EIO:
if (n2rng->n_hvapi_major_version < 2) {
if (++failcount >
RNG_MAX_DATA_READ_ATTEMPTS) {
DBG2(n2rng, DHEALTH,
"n2rng_getentropy: exceeded"
"EIO count of %d on cpu %d",
RNG_MAX_DATA_READ_ATTEMPTS,
CPU->cpu_id);
goto exitpoint;
} else {
delay(drv_usectohz
(RNG_RETRY_HLCHK_USECS));
}
} else {
goto exitpoint;
}
break;
case H_EWOULDBLOCK:
if (++blockcount > RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH,
"n2rng_getentropy: "
"exceeded block count of %d",
RNG_MAX_BLOCK_ATTEMPTS);
goto exitpoint;
} else {
cyclesleep(n2rng, tdelta);
}
break;
default:
log_internal_errors(hverr, "hv_rng_data_read");
goto exitpoint;
}
}
}
exitpoint:
return (rv);
}
uint64_t
n2rng_ctl_read(n2rng_t *n2rng, int rngid, uint64_t ctlregs_pa, uint64_t *state,
uint64_t *tdelta, uint64_t *wdelta)
{
uint64_t rv;
uint64_t wstatus;
if (n2rng->n_hvapi_major_version == 2) {
rv = hv_rng_ctl_read_v2(ctlregs_pa, (uint64_t)rngid, state,
tdelta, wdelta, &wstatus);
if (rv == 0) {
rv = wstatus;
}
} else {
rv = hv_rng_ctl_read(ctlregs_pa, state, tdelta);
*wdelta = 0;
}
return (rv);
}
uint64_t
n2rng_ctl_wait(n2rng_t *n2rng, int rngid)
{
uint64_t state;
uint64_t tdelta;
uint64_t wdelta;
uint64_t wstatus;
boolean_t rnglooping = B_TRUE;
uint64_t rv;
n2rng_setup_t setupbuffer[2];
n2rng_setup_t *setupcontigp;
uint64_t setupphys;
int busycount = 0;
int blockcount = 0;
setupcontigp = &setupbuffer[
CONTIGUOUS(&setupbuffer[0], n2rng_setup_t) ? 0 : 1];
setupphys = va_to_pa(setupcontigp);
while (rnglooping) {
rv = hv_rng_ctl_read_v2(setupphys, (uint64_t)rngid, &state,
&tdelta, &wdelta, &wstatus);
switch (rv) {
case H_EOK:
rv = wstatus;
rnglooping = B_FALSE;
break;
case H_EWOULDBLOCK:
if (++blockcount > RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH, "n2rng_ctl_wait: "
"exceeded block count of %d",
RNG_MAX_BLOCK_ATTEMPTS);
return (rv);
} else {
cyclesleep(n2rng, tdelta);
}
break;
case H_EBUSY:
if (++busycount > RNG_MAX_BUSY_ATTEMPTS) {
DBG1(n2rng, DHEALTH, "n2rng_ctl_wait: "
"exceeded busy count of %d",
RNG_MAX_BUSY_ATTEMPTS);
return (rv);
} else {
delay(RNG_RETRY_BUSY_DELAY);
}
break;
default:
log_internal_errors(rv, "n2rng_ctl_wait");
rnglooping = B_FALSE;
}
}
return (rv);
}
uint64_t
n2rng_ctl_write(n2rng_t *n2rng, int rngid, uint64_t ctlregs_pa,
uint64_t newstate, uint64_t wtimeout, uint64_t *tdelta)
{
uint64_t rv;
if (n2rng->n_hvapi_major_version == 2) {
rv = hv_rng_ctl_write_v2(ctlregs_pa, newstate, wtimeout,
(uint64_t)rngid);
if (rv == H_EOK) {
rv = n2rng_ctl_wait(n2rng, rngid);
}
*tdelta = RNG_DEFAULT_ACCUMULATE_CYCLES;
} else {
rv = hv_rng_ctl_write(ctlregs_pa, newstate, wtimeout, tdelta);
}
return (rv);
}
uint64_t
n2rng_data_read_diag(n2rng_t *n2rng, int rngid, uint64_t data_pa,
size_t datalen, uint64_t *tdelta)
{
uint64_t rv;
if (n2rng->n_hvapi_major_version == 2) {
rv = hv_rng_data_read_diag_v2(data_pa, datalen,
(uint64_t)rngid, tdelta);
if (*tdelta == 0) {
*tdelta = RNG_DEFAULT_ACCUMULATE_CYCLES;
}
} else {
rv = hv_rng_data_read_diag(data_pa, datalen, tdelta);
}
return (rv);
}
uint64_t
n2rng_check_ctl_access(n2rng_t *n2rng)
{
uint64_t rv;
uint64_t unused_64;
if (n2rng->n_hvapi_major_version == 2) {
rv = hv_rng_ctl_read_v2(0, N2RNG_INVALID_ID,
&unused_64, &unused_64, &unused_64, &unused_64);
} else {
rv = hv_rng_get_diag_control();
}
return (rv);
}
void
n2rng_config_retry(n2rng_t *n2rng, clock_t seconds)
{
mutex_enter(&n2rng->n_lock);
if (n2rng->n_timeout_id) {
DBG1(n2rng, DCFG, "n2rng_config_retry: retry pending "
"id = %x", n2rng->n_timeout_id);
} else {
n2rng->n_timeout_id = timeout(n2rng_config_task,
(void *)n2rng, drv_usectohz(seconds * SECOND));
DBG2(n2rng, DCFG, "n2rng_config_retry: retry scheduled in "
"%d seconds, id = %x", seconds, n2rng->n_timeout_id);
}
mutex_exit(&n2rng->n_lock);
}
static uint64_t
sticks_per_usec(void)
{
uint64_t starttick = gettick();
hrtime_t starttime = gethrtime();
uint64_t endtick;
hrtime_t endtime;
delay(2);
endtick = gettick();
endtime = gethrtime();
return ((1000 * (endtick - starttick)) / (endtime - starttime));
}
static int
n2rng_init_ctl(n2rng_t *n2rng)
{
int rv;
int hverr;
rng_entry_t *rng;
int rngid;
int blockcount = 0;
n2rng->n_ctl_data = NULL;
do {
hverr = n2rng_check_ctl_access(n2rng);
rv = n2rng_herr2kerr(hverr);
if ((hverr == H_EWOULDBLOCK) &&
(++blockcount > RNG_MAX_BUSY_ATTEMPTS)) {
DBG1(n2rng, DHEALTH, "n2rng_int_ctl: exceeded busy "
"count of %d", RNG_MAX_BUSY_ATTEMPTS);
return (rv);
} else {
delay(RNG_RETRY_BUSY_DELAY);
}
} while (hverr == H_EWOULDBLOCK);
if (rv == EPERM) {
DBG0(n2rng, DATTACH,
"n2rng_init_ctl: Running in guest domain");
return (DDI_SUCCESS);
}
n2rng->n_ctl_data = kmem_alloc(sizeof (rng_ctl_data_t), KM_SLEEP);
n2rng->n_ctl_data->n_num_rngs_online = 0;
if (n2rng->n_hvapi_major_version < 2) {
n2rng->n_ctl_data->n_num_rngs = 1;
} else {
n2rng->n_ctl_data->n_num_rngs =
ddi_getprop(DDI_DEV_T_ANY, n2rng->n_dip,
DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS,
N2RNG_PROP_NUM_UNITS, 0);
if (n2rng->n_ctl_data->n_num_rngs == 0) {
cmn_err(CE_WARN, "n2rng: %s property not found",
N2RNG_PROP_NUM_UNITS);
return (DDI_FAILURE);
}
}
n2rng->n_ctl_data->n_rngs =
kmem_zalloc(n2rng->n_ctl_data->n_num_rngs *
sizeof (rng_entry_t), KM_SLEEP);
n2rng->n_ctl_data->n_accumulate_cycles =
ddi_getprop(DDI_DEV_T_ANY, n2rng->n_dip,
DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "acc_cycles",
RNG_DEFAULT_ACCUMULATE_CYCLES);
n2rng->n_ctl_data->n_hc_secs = ddi_getprop(DDI_DEV_T_ANY, n2rng->n_dip,
DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "hc_seconds",
RNG_DEFAULT_HC_SECS);
if ((n2rng->n_hvapi_major_version < 2) &&
(n2rng->n_ctl_data->n_hc_secs > 0)) {
cmn_err(CE_WARN, "n2rng: Hyperviser api "
"version %d.%d does not support health checks",
n2rng->n_hvapi_major_version,
n2rng->n_hvapi_minor_version);
n2rng->n_ctl_data->n_hc_secs = 0;
}
if (n2rng->n_ctl_data->n_hc_secs <= 0) {
n2rng->n_ctl_data->n_watchdog_cycles = 0;
} else {
n2rng->n_ctl_data->n_watchdog_cycles =
((uint64_t)(RNG_EXTRA_WATCHDOG_SECS) +
n2rng->n_ctl_data->n_hc_secs) *
n2rng->n_sticks_per_usec * 1000000;
}
for (rngid = 0; rngid < n2rng->n_ctl_data->n_num_rngs; rngid++) {
rng = &n2rng->n_ctl_data->n_rngs[rngid];
rng->n_preferred_config.ctlwds[0].word = 0;
rng->n_preferred_config.ctlwds[0].fields.rnc_anlg_sel =
N2RNG_NOANALOGOUT;
rng->n_preferred_config.ctlwds[0].fields.rnc_cnt =
RNG_DEFAULT_ACCUMULATE_CYCLES;
rng->n_preferred_config.ctlwds[0].fields.rnc_mode =
RNG_MODE_NORMAL;
rng->n_preferred_config.ctlwds[1].word =
rng->n_preferred_config.ctlwds[0].word;
rng->n_preferred_config.ctlwds[2].word =
rng->n_preferred_config.ctlwds[0].word;
rng->n_preferred_config.ctlwds[3].word =
rng->n_preferred_config.ctlwds[0].word;
rng->n_preferred_config.ctlwds[0].fields.rnc_vcoctl = 1;
rng->n_preferred_config.ctlwds[0].fields.rnc_selbits = 1;
rng->n_preferred_config.ctlwds[1].fields.rnc_vcoctl = 2;
rng->n_preferred_config.ctlwds[1].fields.rnc_selbits = 2;
rng->n_preferred_config.ctlwds[2].fields.rnc_vcoctl = 3;
rng->n_preferred_config.ctlwds[2].fields.rnc_selbits = 4;
rng->n_preferred_config.ctlwds[3].fields.rnc_vcoctl = 0;
rng->n_preferred_config.ctlwds[3].fields.rnc_selbits = 7;
}
n2rng_setcontrol(n2rng);
DBG2(n2rng, DATTACH,
"n2rng_init_ctl: Running in control domain with %d rng device%s",
n2rng->n_ctl_data->n_num_rngs,
(n2rng->n_ctl_data->n_num_rngs == 1) ? "" : "s");
DBG2(n2rng, DCFG,
"n2rng_init_ctl: n_sticks_per_usec = %ld, n_hc_secs = %d",
n2rng->n_sticks_per_usec,
n2rng->n_ctl_data->n_hc_secs);
DBG2(n2rng, DCFG,
"n2rng_init_ctl: n_watchdog_cycles = %ld, "
"n_accumulate_cycles = %ld", n2rng->n_ctl_data->n_watchdog_cycles,
n2rng->n_ctl_data->n_accumulate_cycles);
return (DDI_SUCCESS);
}
static void
n2rng_uninit_ctl(n2rng_t *n2rng)
{
if (n2rng->n_ctl_data) {
if (n2rng->n_ctl_data->n_num_rngs) {
kmem_free(n2rng->n_ctl_data->n_rngs,
n2rng->n_ctl_data->n_num_rngs *
sizeof (rng_entry_t));
n2rng->n_ctl_data->n_rngs = NULL;
n2rng->n_ctl_data->n_num_rngs = 0;
}
kmem_free(n2rng->n_ctl_data, sizeof (rng_ctl_data_t));
n2rng->n_ctl_data = NULL;
}
}
int
n2rng_config_test(n2rng_t *n2rng)
{
int rv = 0;
uint64_t hverr;
uint64_t randval = 0;
uint64_t randvalphys = va_to_pa(&randval);
uint64_t tdelta;
int failcount = 0;
int blockcount = 0;
boolean_t rnglooping = B_TRUE;
while (rnglooping) {
hverr = hv_rng_data_read(randvalphys, &tdelta);
rv = n2rng_herr2kerr(hverr);
switch (hverr) {
case H_EOK:
failcount = 0;
rnglooping = B_FALSE;
break;
case H_EIO:
if (++failcount > RNG_MAX_DATA_READ_ATTEMPTS) {
goto exitpoint;
} else {
delay(drv_usectohz(RNG_RETRY_HLCHK_USECS));
}
break;
case H_EWOULDBLOCK:
if (++blockcount > RNG_MAX_BLOCK_ATTEMPTS) {
DBG1(n2rng, DHEALTH, "n2rng_config_test: "
"exceeded block count of %d",
RNG_MAX_BLOCK_ATTEMPTS);
goto exitpoint;
} else {
cyclesleep(n2rng, tdelta);
}
break;
case H_ENOACCESS:
goto exitpoint;
default:
log_internal_errors(hverr, "hv_rng_data_read");
goto exitpoint;
}
}
exitpoint:
return (rv);
}
static int
n2rng_config(n2rng_t *n2rng)
{
int rv;
rng_entry_t *rng;
int rngid;
if (n2rng_iscontrol(n2rng)) {
for (rngid = 0; rngid < n2rng->n_ctl_data->n_num_rngs;
rngid++) {
rng = &n2rng->n_ctl_data->n_rngs[rngid];
if (rng->n_rng_state == CTL_STATE_ERROR) {
continue;
}
if ((n2rng->n_binding == N2RNG_CPU_VF) &&
(n2rng->n_hvapi_major_version < 2)) {
thread_affinity_set(curthread, CPU_CURRENT);
DBG1(n2rng, DCFG, "n2rng_config: "
"Configuring single rng from cpu %d",
CPU->cpu_id);
rv = n2rng_do_health_check(n2rng, rngid);
thread_affinity_clear(curthread);
} else {
rv = n2rng_do_health_check(n2rng, rngid);
}
switch (rv) {
case 0:
DBG1(n2rng, DCFG, "n2rng_config: rng(%d) "
"passed health checks", rngid);
if (rng->n_rng_state != CTL_STATE_CONFIGURED) {
rng->n_rng_state =
CTL_STATE_CONFIGURED;
n2rng->n_ctl_data->n_num_rngs_online++;
}
break;
default:
cmn_err(CE_WARN, "n2rng: rng(%d) "
"failed health checks", rngid);
if (rng->n_rng_state == CTL_STATE_CONFIGURED) {
n2rng->n_ctl_data->n_num_rngs_online--;
}
rng->n_rng_state = CTL_STATE_ERROR;
break;
}
}
DBG2(n2rng, DCFG, "n2rng_config: %d rng%s online",
n2rng->n_ctl_data->n_num_rngs_online,
(n2rng->n_ctl_data->n_num_rngs_online == 1) ? "" : "s");
if (n2rng->n_ctl_data->n_num_rngs_online == 0) {
cmn_err(CE_WARN, "n2rng: %d RNG device%s failed",
n2rng->n_ctl_data->n_num_rngs,
(n2rng->n_ctl_data->n_num_rngs == 1) ? "" : "s");
goto errorexit;
} else {
n2rng_setconfigured(n2rng);
}
} else {
rv = n2rng_config_test(n2rng);
switch (rv) {
case 0:
n2rng_setconfigured(n2rng);
break;
case EIO:
break;
default:
goto errorexit;
}
}
if (n2rng_isconfigured(n2rng)) {
if (n2rng_init(n2rng) != DDI_SUCCESS) {
cmn_err(CE_WARN, "n2rng: unable to register with KCF");
goto errorexit;
}
if (n2rng_iscontrol(n2rng) &&
(n2rng->n_ctl_data->n_hc_secs > 0)) {
n2rng_config_retry(n2rng,
n2rng->n_ctl_data->n_hc_secs);
}
} else if (!n2rng_isfailed(n2rng)) {
n2rng_config_retry(n2rng, RNG_CFG_RETRY_SECS);
}
return (DDI_SUCCESS);
errorexit:
(void) n2rng_unregister_provider(n2rng);
n2rng_setfailed(n2rng);
cmn_err(CE_WARN, "n2rng: hardware failure detected");
return (DDI_FAILURE);
}
static void
n2rng_config_task(void *targ)
{
n2rng_t *n2rng = (n2rng_t *)targ;
mutex_enter(&n2rng->n_lock);
n2rng->n_timeout_id = 0;
mutex_exit(&n2rng->n_lock);
(void) n2rng_config(n2rng);
}