TTE8K
cb_mapin(vaddr, src_ppn, TTE8K, 0, dtlb_index);
cb_mapin(vaddr, dst_ppn, TTE8K, TTE_HWWR_INT, dtlb_index);
TTE8K, TTE_HWWR_INT, cb_dents - 1);
(TTE_PAGE_SHIFT(TTE8K) + 2);
if (get_hblk_ttesz(hmeblkp) == TTE8K) { \
sfmmup->sfmmu_ttecnt[TTE8K] <= tsb_rss_factor)
tte8k_cnt = ttecnt[TTE8K] +
for (i = TTE8K; i < max_mmu_page_sizes; i++) {
ASSERT(sfmmup->sfmmu_ttecnt[TTE8K] == 0);
if ((*hmeblkpp = sfmmu_check_pending_hblks(TTE8K)) != NULL) {
hmebp = sfmmu_tteload_acquire_hashbucket(ksfmmup, base, TTE8K,
for (vaddr = base; vaddr < endaddr; vaddr += TTEBYTES(TTE8K)) {
hblktag.htag_bspage = HME_HASH_BSPAGE(base, HME_HASH_SHIFT(TTE8K));
hblktag.htag_rehash = HME_HASH_REHASH(TTE8K);
if (size == TTE8K) {
if (owner && size == TTE8K) {
sfmmu_cache = ((size == TTE8K) ? sfmmu8_cache : sfmmu1_cache);
if (size == TTE8K && sfmmup != KHATID) {
if (size == TTE8K &&
if (size != TTE8K ||
if (size == TTE8K) {
uchar_t tteflag_mask = (1 << TTE64K) | (1 << TTE8K);
sfmmu_unload_tsb_range(sfmmup, addr, eaddr, TTE8K);
PROT_WRITE|PROT_READ, TTE8K);
ASSERT(ttesz > TTE8K);
if (r_pgszc > TTE8K) {
r_size >> (TTE_PAGE_SHIFT(TTE8K) + 2);
i = TTE8K;
rttecnt = r_size >> (TTE_PAGE_SHIFT(TTE8K) + 2);
rgnp->rgn_size >> (TTE_PAGE_SHIFT(TTE8K) + 2);
TTE_PAGE_SHIFT(TTE8K);
(TTE_PAGE_SHIFT(TTE8K) + 2);
sfmmup->sfmmu_pgsz[i] = TTE8K;
ASSERT(sfmmup->sfmmu_ttecnt[TTE8K] == 0);
sfmmu_memtte(&tte, pp->p_pagenum, attr, TTE8K);
sfmmu_memtte(&tte, pfn, attr, TTE8K);
sfmmu_memtte(&tte, pfn, attr, TTE8K);
for (ttesz = pp->p_szc; ttesz != TTE8K; ttesz--) {
if (ttesz == TTE8K) {
hmebp = sfmmu_tteload_acquire_hashbucket(hat, vaddr, TTE8K,
TTE8K, flags, rid);
sfmmu_memtte(&tte, pp->p_pagenum, attr, TTE8K);
case TTE8K:
if (size == TTE8K) {
if (size > TTE8K && (flags & HAT_LOAD_SHARE) == 0 &&
if (size == TTE8K || size == TTE4M) {
ASSERT(ttesz > TTE8K);
if (ttesz == TTE8K) {
if (get_hblk_ttesz(hmeblkp) != TTE8K &&
if (get_hblk_ttesz(hmeblkp) != TTE8K &&
if (sz != TTE8K) {
cons = TTE8K;
cons = TTE8K;
if (cons == TTE8K) {
cons = TTE8K;
cons = TTE8K;
cons = TTE8K;
cons = TTE8K;
get_hblk_ttesz(hmeblkp) > TTE8K) {
} else if (get_hblk_ttesz(hmeblkp) > TTE8K ||
int index, sz = TTE8K;
if (ttesz == TTE8K || ttesz == TTE4M) { \
(void) sfmmu_pageunload(pp, sfhmep, TTE8K);
cursz = TTE8K;
cursz = TTE8K;
if (hme_size(hme) > TTE8K) { \
#define get_hblk_cache(hmeblkp) (((hmeblkp)->hblk_ttesz == TTE8K) ? \
((ttesz == TTE8K)? (TTEBYTES(ttesz) * NHMENTS) : TTEBYTES(ttesz))
((ttesz == TTE8K)? HBLK_RANGE_SHIFT : TTE_PAGE_SHIFT(ttesz))
#define TSB8K (1 << TTE8K)
tte.tte_inthi = TTE_VALID_INT | TTE_SZ_INT(TTE8K) |
pages_per_hblk = btop(HMEBLK_SPAN(TTE8K));
TTE8K,
ASSERT(mmu_exported_pagesize_mask & (1 << TTE8K));
for (i = max_bootlp_tteszc; i >= TTE8K; i--) {
if (i == TTE8K) {
sfmmu_memtte(ttep, pfn, attr, TTE8K);
if (pszc > TTE8K) {
max_bootlp_tteszc = TTE8K;
(HMENT_HASHAVELEN * (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT));
hme_buckets = npages / (HMEBLK_SPAN(TTE8K) >> MMU_PAGESHIFT);
static uchar_t shctx_cext = TAGACCEXT_MKSZPAIR(DEFAULT_ISM_PAGESZC, TTE8K);
tte = TTE8K;
tte = TTE8K;
new_cext_nucleus = TAGACCEXT_MKSZPAIR(tte, TTE8K);
new_cext_primary = TAGACCEXT_MKSZPAIR(TTE8K, tte);
for (tte = TTE8K; tte <= TTE4M; tte++) {
return (TTEBYTES(TTE8K));
mmu_exported_pagesize_mask = (1 << TTE8K) |
shctx_cext = TAGACCEXT_MKSZPAIR(TTE4M, TTE8K);
shctx_cext = TAGACCEXT_MKSZPAIR(TTE32M, TTE8K);
shctx_cext = TAGACCEXT_MKSZPAIR(TTE256M, TTE8K);
for (tte = TTE8K; tte <= TTE4M; tte++) {
return (TTEBYTES(TTE8K));
max_bootlp_tteszc = TTE8K;
mmu_exported_pagesize_mask = (1 << TTE8K) |
{CHEETAH_IMPL, TTE8K, 0}, /* default */
{ CHEETAH_PLUS_IMPL, TTE8K, 0 },
{ JALAPENO_IMPL, TTE8K, 0 },
{ JAGUAR_IMPL, TTE8K, 0 },
{ SERRANO_IMPL, TTE8K, 0 },
{ PANTHER_IMPL, TTE8K, 0 }
tte = TTE8K;
new_cext_nucleus = TAGACCEXT_MKSZPAIR(tte, TTE8K);
new_cext_primary = TAGACCEXT_MKSZPAIR(TTE8K, tte);
uint_t tte = TTE8K;
tte = TTE8K;
sfmmu_memtte(&tte, va_to_pfn(va), PROC_DATA|HAT_NOSYNC, TTE8K);
sfmmu_memtte(&tte, pa >> PAGESHIFT, PROC_DATA|HAT_NOSYNC, TTE8K);
tte.tte_inthi = TTE_VALID_INT | TTE_SZ_INT(TTE8K) |
tte.tte_inthi = TTE_VALID_INT | TTE_SZ_INT(TTE8K) |
tte.tte_inthi = TTE_VALID_INT | TTE_SZ_INT(TTE8K) |
#define TTE_IS_8K(ttep) (TTE_CSZ(ttep) == TTE8K)
conv_tnc(pp, TTE8K);
conv_tnc(pp, TTE8K);
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VUNCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VUNCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
(void) sfmmu_pageunload(pp, sfhmep, TTE8K);
uint_t szc = kpm_smallpages ? TTE8K : TTE4M;
uint_t szc = kpm_smallpages ? TTE8K : TTE4M;
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VUNCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VUNCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VCACHED(tte.ll, pfn, TTE8K);
KPM_TTE_VUNCACHED(tte.ll, pfn, TTE8K);
conv_tnc(pp, TTE8K);
tte.tte_inthi = TTE_VALID_INT | TTE_SZ_INT(TTE8K) | TTE_PFN_INTHI(pfn);
int mmu_exported_pagesize_mask = (1 << TTE8K) | (1 << TTE64K) |
#define NI_MMU_PAGESIZE_MASK ((1 << TTE8K) | (1 << TTE64K) | (1 << TTE4M) \
pfn_ceq_mask = (szc == TTE8K) ? 0 :
npfn = (szc == TTE8K) ? ((pfn >> 15) << 15) :
tmpmask = (szc == TTE8K) ? 0 : (ceq_mask & 0x1c) << 13;
PROT_READ | PROT_WRITE | HAT_NOSYNC, TTE8K);
#define DEFAULT_SUN4V_MMU_PAGESIZE_MASK ((1 << TTE8K) | (1 << TTE64K) \
#define TTE_IS_8K(ttep) (TTE_CSZ(ttep) == TTE8K)
sfmmu_memtte(&tte, pfn, (PROC_TEXT | HAT_NOSYNC), TTE8K);
for (tte = TTE8K; tte <= TTE256M; tte++) {
lpsize = TTEBYTES(TTE8K);