#ifndef _AMD_IOMMU_PAGE_TABLES_H
#define _AMD_IOMMU_PAGE_TABLES_H
#ifdef __cplusplus
extern "C" {
#endif
#ifdef _KERNEL
#define AMD_IOMMU_PTDE_IW (62 << 16 | 62)
#define AMD_IOMMU_PTDE_IR (61 << 16 | 61)
#define AMD_IOMMU_PTDE_ADDR (51 << 16 | 12)
#define AMD_IOMMU_PTDE_NXT_LVL (11 << 16 | 9)
#define AMD_IOMMU_PTDE_PR (0 << 16 | 0)
#define AMD_IOMMU_PTE_FC (60 << 16 | 60)
#define AMD_IOMMU_PTE_U (59 << 16 | 59)
#define AMD_IOMMU_VA_NBITS(l) ((l) == 6 ? 7 : 9)
#define AMD_IOMMU_VA_BITMASK(l) ((1 << AMD_IOMMU_VA_NBITS(l)) - 1)
#define AMD_IOMMU_VA_SHIFT(v, l) \
((v) >> (MMU_PAGESHIFT + (AMD_IOMMU_VA_NBITS(l - 1) * (l - 1))))
#define AMD_IOMMU_VA_BITS(v, l) \
(AMD_IOMMU_VA_SHIFT(v, l) & AMD_IOMMU_VA_BITMASK(l))
#define AMD_IOMMU_VA_TOTBITS(l) \
(((l) == 6 ? 7 + (l - 1) * 9: l*9) + MMU_PAGESHIFT)
#define AMD_IOMMU_VA_TOTMASK(l) ((1 << AMD_IOMMU_VA_TOTBITS(l)) - 1)
#define AMD_IOMMU_VA_INVAL_SETMASK(l) \
(((1 << AMD_IOMMU_VA_TOTBITS(l)) - 1) >> 1)
#define AMD_IOMMU_VA_INVAL_CLRMASK(l) \
(~(1 << (AMD_IOMMU_VA_TOTBITS(l) - 1)))
#define AMD_IOMMU_VA_INVAL(v, l) \
(((v) & AMD_IOMMU_VA_INVAL_CLRMASK(l)) | AMD_IOMMU_VA_INVAL_SETMASK(l))
#define AMD_IOMMU_PGTABLE_SZ (4096)
#define AMD_IOMMU_PGTABLE_MAXLEVEL (6)
#define AMD_IOMMU_PGTABLE_HASH_SZ (256)
#define AMD_IOMMU_PGTABLE_ALIGN ((1ULL << 12) - 1)
#define AMD_IOMMU_PGTABLE_SIZE (1ULL << 12)
#define AMD_IOMMU_MAX_PDTE (1ULL << AMD_IOMMU_VA_NBITS(1))
#define PT_REF_VALID(p) ((p)->pt_ref >= 0 && \
(p)->pt_ref <= AMD_IOMMU_MAX_PDTE)
#define AMD_IOMMU_DOMAIN_HASH_SZ (256)
#define AMD_IOMMU_PGTABLE_FREELIST_MAX (256)
#define AMD_IOMMU_PA2VA_HASH_SZ (256)
#define AMD_IOMMU_SIZE_4G ((uint64_t)1 << 32)
#define AMD_IOMMU_VMEM_NAMELEN (30)
typedef enum {
AMD_IOMMU_INVALID_DOMAIN = 0,
AMD_IOMMU_IDENTITY_DOMAIN = 0xFFFD,
AMD_IOMMU_PASSTHRU_DOMAIN = 0xFFFE,
AMD_IOMMU_SYS_DOMAIN = 0xFFFF
} domain_id_t;
typedef enum {
AMD_IOMMU_INVALID_MAP = 0,
AMD_IOMMU_UNITY_MAP,
AMD_IOMMU_VMEM_MAP
} map_type_t;
typedef struct amd_iommu_page_table {
domain_id_t pt_domainid;
int pt_level;
ddi_dma_handle_t pt_dma_hdl;
ddi_acc_handle_t pt_mem_hdl;
uint64_t pt_mem_reqsz;
uint64_t pt_mem_realsz;
uint64_t *pt_pgtblva;
uint64_t pt_pte_ref[AMD_IOMMU_MAX_PDTE];
uint16_t pt_index;
int pt_ref;
ddi_dma_cookie_t pt_cookie;
struct amd_iommu_page_table *pt_next;
struct amd_iommu_page_table *pt_prev;
struct amd_iommu_page_table *pt_parent;
} amd_iommu_page_table_t;
typedef struct amd_iommu_domain {
domain_id_t d_domainid;
uint64_t d_pgtable_root_4K;
int64_t d_ref;
vmem_t *d_vmem;
struct amd_iommu_domain *d_prev;
struct amd_iommu_domain *d_next;
} amd_iommu_domain_t;
int amd_iommu_map_pa2va(amd_iommu_t *iommu, dev_info_t *rdip,
ddi_dma_attr_t *attrp, struct ddi_dma_req *dmareq,
uint64_t pa, uint64_t pa_sz, map_type_t type,
uint64_t *start_vap, int km_flags);
int amd_iommu_unmap_va(amd_iommu_t *iommu, dev_info_t *rdip,
uint64_t va, uint64_t va_sz, map_type_t type);
void amd_iommu_init_page_tables(amd_iommu_t *iommu);
void amd_iommu_fini_page_tables(amd_iommu_t *iommu);
void amd_iommu_set_passthru(amd_iommu_t *iommu, dev_info_t *rdip);
#endif
#ifdef __cplusplus
}
#endif
#endif