paddr_t
static paddr_t
return (-(paddr_t)1);
return (mmu_ptob((paddr_t)pfn) | (ma & (MMU_PAGESIZE - 1)));
paddr_t paddr = mmu_ptob((paddr_t)htable->ht_pfn);
extern caddr_t psm_map_phys_new(paddr_t, size_t, int);
extern gfx_maddr_t gfxp_convert_addr(paddr_t paddr);
extern paddr_t dboot_multiboot2_highest_addr(multiboot2_info_header_t *);
paddr_t dma_addr;
blk_off = (uint_t)((paddr_t)XB_SEGOFFSET & dma_addr);
paddr_t xnf_tx_ring_phys_addr;
paddr_t xnf_rx_ring_phys_addr;
paddr_t buf_phys;
paddr_t rpaddr;
paddr_t xr_paddr;
paddr_t top_page_table;
find_pte(uint64_t va, paddr_t *pa, uint_t level, uint_t probe_only)
paddr_t table;
paddr_t new_table;
paddr_t src;
paddr_t dst;
paddr_t
return ((paddr_t)(uintptr_t)mbi + mbi->mbi_total_size);
max_mem = (paddr_t)xen_info->nr_pages << MMU_PAGESHIFT;
paddr_t ktext_phys;
static paddr_t next_avail_addr = 0;
static paddr_t scratch_end = 0; /* we can't write all of mem here */
static paddr_t mfn_base; /* addr corresponding to mfn_list[0] */
check_higher((paddr_t)(uintptr_t)&_end);
static paddr_t max_mem; /* maximum memory address */
static paddr_t
paddr_t addr = (paddr_t)(uintptr_t)NULL;
addr = ((paddr_t)((uintptr_t)cmdl + strlen(cmdl) + 1));
((paddr_t)(mb_info->mmap_addr + mb_info->mmap_length)));
paddr_t addr;
if (addr != (paddr_t)(uintptr_t)NULL)
if (addr != (paddr_t)(uintptr_t)NULL)
top_page_table = (paddr_t)(uintptr_t)xen_info->pt_base;
top_page_table = (paddr_t)(uintptr_t)mem_alloc(MMU_PAGESIZE);
paddr_t
paddr_t pa;
return (-(paddr_t)1);
pa = mfn_base + mmu_ptob((paddr_t)pfn) + pgoff;
pa_to_ma(paddr_t pa)
get_pteval(paddr_t table, uint_t index)
set_pteval(paddr_t table, uint_t index, uint_t level, x86pte_t pteval)
paddr_t
paddr_t new_table = (paddr_t)(uintptr_t)mem_alloc(MMU_PAGESIZE);
map_pte(paddr_t table, uint_t index)
#define maddr_t paddr_t
map_pa_at_va(paddr_t pa, native_ptr_t va, uint_t level)
check_higher(paddr_t a)
paddr_t scratch_start;
scratch_start = RNDUP((paddr_t)(uintptr_t)&local, MMU_PAGESIZE);
scratch_end = RNDUP((paddr_t)scratch_start + 512 * 1024, TWO_MEG);
extern paddr_t ktext_phys;
gfxp_convert_addr(paddr_t paddr)
PDTE_check(immu_t *immu, hw_pdte_t pdte, pgtable_t *next, paddr_t paddr,
PTE_set_one(immu_t *immu, hw_pdte_t *hwp, paddr_t paddr,
paddr_t paddr;
paddr_t qinv_mem_paddr;
paddr_t pbase;
paddr_t pbase;
typedef paddr_t rootnex_addr_t;
paddr_t paddr;
paddr_t paddr;
paddr_t maxphysaddr;
maxphysaddr = mmu_ptob((paddr_t)physmax) + MMU_PAGEOFFSET;
maxphysaddr = mmu_ptob((paddr_t)HYPERVISOR_memory_op(
i_ddi_paddr_to_pfn(paddr_t paddr)
static paddr_t next_phys; /* next available physical address from dboot */
static paddr_t high_phys = -(paddr_t)1; /* last used physical address */
paddr_t
paddr_t pa = 0;
paddr_t start;
paddr_t end;
vmap_phys(size_t length, paddr_t pa)
paddr_t start, end;
scan_rsdp(paddr_t *paddrp, size_t len)
paddr_t paddr = *paddrp;
paddr_t paddr = 0;
map_fw_table(paddr_t table_addr)
paddr_t xsdt_addr;
paddr_t table_addr;
alloc_vaddr(size_t size, paddr_t align)
paddr_t a = align; /* same type as pa for masking */
paddr_t pa;
paddr_t, size_t, int);
fastboot_map_with_size(fastboot_info_t *nk, uintptr_t vstart, paddr_t pstart,
paddr_t paddr;
((uintptr_t)(((paddr_t)table[index] & MMU_PAGEMASK)
nk->fi_new_mbi_pa = (paddr_t)start_addr_pa;
va = (void *)psm_map((paddr_t)(start&0xffffffff),
(paddr_t)((start+\
va = psm_map((paddr_t)(start&0xffffffff),
psm_map_phys_new(paddr_t addr, size_t len, int prot)
paddr_t base;
psm_map_new(paddr_t addr, size_t len, int prot)
return (psm_map_phys_new((paddr_t)(addr & 0xffffffff), len, prot));
return (psm_map_new((paddr_t)(addr & 0xffffffff), len, prot));
paddr_t paddr;
paddr = (paddr_t)req->bdf << 12;
extern uintptr_t alloc_vaddr(size_t, paddr_t);
pci_cfgacc_map(paddr_t phys_addr)
paddr_t pa_base = P2ALIGN(phys_addr, MMU_PAGESIZE);
paddr_t fb_pte_list_pa; /* PA for PTE list */
paddr_t fi_pagetable_pa;
paddr_t fi_last_table_pa;
paddr_t fi_new_mbi_pa; /* new multiboot info PA */
paddr_t fi_next_table_pa;
paddr_t hwpg_paddr; /* HW pgtable PA */
paddr_t intrmap_paddr;
paddr_t dck_paddr;
x86pte_t get_pteval(paddr_t, uint_t);
void set_pteval(paddr_t, uint_t, uint_t, x86pte_t);
paddr_t make_ptable(x86pte_t *, uint_t);
x86pte_t *find_pte(uint64_t, paddr_t *, uint_t, uint_t);
x86pte_t *map_pte(paddr_t, uint_t);
extern paddr_t top_page_table;
extern caddr_t psm_map_new(paddr_t, size_t, int);
extern caddr_t psm_map_phys_new(paddr_t, size_t, int);
typedef paddr_t hat_mempte_t; /* phys addr of PTE */
paddr_t pwin_pte_pa;
((paddr_t)(p) + ((x) << mmu.pte_size_shift))
#define pfn_to_pa(pfn) (mmu_ptob((paddr_t)(pfn)))
paddr_t pa = mmu_ptob((paddr_t)pfn);
xen_kpm_create(paddr_t paddr, level_t lvl)
kbm_map((uintptr_t)kpm_vbase + paddr, (paddr_t)0, 0, 1);
kbm_remap_window(paddr_t physaddr, int writeable)
kbm_map(uintptr_t va, paddr_t pa, uint_t level, uint_t is_kernel)
paddr_t pte_physaddr;
paddr_t pte_physaddr;
paddr_t pte_physaddr;
kbm_read_only(uintptr_t va, paddr_t pa)
kbm_push(paddr_t pa)
get_pteval(paddr_t table, uint_t index)
set_pteval(paddr_t table, uint_t index, uint_t level, x86pte_t pteval)
paddr_t
paddr_t new_table;
map_pte(paddr_t table, uint_t index)
extern void *kbm_remap_window(paddr_t physaddr, int writeable);
extern void kbm_map(uintptr_t va, paddr_t pa, uint_t level, uint_t is_kernel);
extern void kbm_read_only(uintptr_t va, paddr_t pa);
extern void *kbm_push(paddr_t pa);
paddr_t metasz;
set_pteval(paddr_t table, uint_t index, uint_t level, x86pte_t pteval)
pa_to_ma(paddr_t pa)
paddr_t
paddr_t ma_to_pa(maddr_t);
maddr_t pa_to_ma(paddr_t);
return (psm_map_new((paddr_t)PhysicalAddress,
ptr = psm_map_new((paddr_t)Address, maplen,
paddr_t ac_sg_paddr; /* phys addr of S/G list */
extern paddr_t do_bop_phys_alloc(uint64_t, uint64_t);
pfn_t i_ddi_paddr_to_pfn(paddr_t paddr);