vml
if (vml(pmtab->p_modules) != 0)
extern int vml(char *);
if (vml(gptr->p_modules) != 0) {
if (mflag && (vml(ptr->p_modules) != 0))
vdev_t *rvd, **vml = NULL; /* vdev modify list */
vml = kmem_zalloc(children * sizeof (vdev_t *), KM_SLEEP);
vml[c] = spa_lookup_by_guid(spa, glist[c], B_FALSE);
if (vml[c] == NULL) {
if (vml[c]->vdev_parent->vdev_ops != &vdev_mirror_ops ||
vml[c]->vdev_islog ||
!vdev_is_concrete(vml[c]) ||
vml[c]->vdev_isspare ||
vml[c]->vdev_isl2cache ||
!vdev_writeable(vml[c]) ||
vml[c]->vdev_children != 0 ||
vml[c]->vdev_state != VDEV_STATE_HEALTHY ||
if (vdev_dtl_required(vml[c])) {
vml[c]->vdev_top->vdev_ms_array) == 0);
vml[c]->vdev_top->vdev_ms_shift) == 0);
vml[c]->vdev_top->vdev_asize) == 0);
vml[c]->vdev_top->vdev_ashift) == 0);
ASSERT3U(vml[c]->vdev_leaf_zap, !=, 0);
ZPOOL_CONFIG_VDEV_LEAF_ZAP, vml[c]->vdev_leaf_zap));
ASSERT3U(vml[c]->vdev_top->vdev_top_zap, !=, 0);
vml[c]->vdev_parent->vdev_top_zap));
kmem_free(vml, children * sizeof (vdev_t *));
if (vml[c] != NULL)
vml[c]->vdev_offline = B_TRUE;
if (vml[c] != NULL) {
mutex_enter(&vml[c]->vdev_initialize_lock);
vdev_initialize_stop(vml[c],
mutex_exit(&vml[c]->vdev_initialize_lock);
mutex_enter(&vml[c]->vdev_trim_lock);
vdev_trim_stop(vml[c], VDEV_TRIM_ACTIVE, &vd_trim_list);
mutex_exit(&vml[c]->vdev_trim_lock);
if (vml[c] != NULL) {
vdev_split(vml[c]);
"vdev=%s", vml[c]->vdev_path);
vdev_free(vml[c]);
kmem_free(vml, children * sizeof (vdev_t *));
if (vml[c] != NULL)
vml[c]->vdev_offline = B_FALSE;
kmem_free(vml, children * sizeof (vdev_t *));
struct vmap vml[MINVMAPS];
fetchpage, vml, MINVMAPS, &npages, rw);
(void) kzero(vml[0].vs_addr, (uint_t)pon);
vml[i].vs_addr != NULL; i++) {
error = uiomove(vml[i].vs_addr + pon,
(void) kzero(vml[i].vs_addr + pon, (uint_t)nzero);
vpm_unmap_pages(vml, rw);
vmap_t vml[],
vmap_t vml[],
vpm_unmap_pages(vmap_t vml[], enum seg_rw rw)
vmap_t vml[],
vml[i].vs_data = (void *)&vpm->vpm_pp;
vml[i].vs_data = (void *)pp;
vml[i].vs_addr = hat_kpm_mapin(pp, 0);
vml[i].vs_len = PAGESIZE;
vml[i].vs_data = NULL;
vml[i].vs_addr = (caddr_t)NULL;
vmap_t *vml,
vml[0].vs_data = NULL;
vml[0].vs_addr = (caddr_t)NULL;
return (vpm_pagecreate(vp, baseoff, len, vml, nseg, newpage));
vml[0].vs_addr = NULL;
vml[0].vs_data = NULL;
vml[i].vs_data = (void *)&(vpm->vpm_pp);
vml[i].vs_data = (void *)pplist[i];
vml[i].vs_addr = hat_kpm_mapin(pplist[i], 0);
vml[i].vs_len = PAGESIZE;
vml[i].vs_data = NULL;
vml[i].vs_addr = (caddr_t)NULL;
vpm_unmap_pages(vmap_t vml[], enum seg_rw rw)
for (i = 0; vml[i].vs_data != NULL; i++) {
ASSERT(IS_KPM_ADDR(vml[i].vs_addr));
pp = *(((page_t **)vml[i].vs_data));
pp = (page_t *)vml[i].vs_data;
vpm = (struct vpmap *)((char *)vml[i].vs_data
hat_kpm_mapout(pp, 0, vml[i].vs_addr);
hat_kpm_mapout(pp, 0, vml[i].vs_addr);
vml[i].vs_data = NULL;
vml[i].vs_addr = NULL;
list_t *vml = &vmm_list;
for (sc = list_head(vml); sc != NULL; sc = list_next(vml, sc)) {