Symbol: kvm_mmu_page
arch/x86/include/asm/kvm_host.h
455
struct kvm_mmu_page;
arch/x86/include/asm/kvm_host.h
473
struct kvm_mmu_page *sp, int i);
arch/x86/kvm/mmu/mmu.c
1188
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
1292
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
1684
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
1793
static void kvm_mmu_check_sptes_at_free(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1807
static void kvm_account_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1813
static void kvm_unaccount_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1819
static void kvm_mmu_free_shadow_page(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1837
struct kvm_mmu_page *sp, u64 *parent_pte)
arch/x86/kvm/mmu/mmu.c
1845
static void mmu_page_remove_parent_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1851
static void drop_parent_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1859
static void kvm_mmu_mark_parents_unsync(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1871
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
1885
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
1891
static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1907
static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx)
arch/x86/kvm/mmu/mmu.c
1914
static int __mmu_unsync_walk(struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1920
struct kvm_mmu_page *child;
arch/x86/kvm/mmu/mmu.c
1955
static int mmu_unsync_walk(struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1966
static void kvm_unlink_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
1974
static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
1979
static bool sp_has_gptes(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2021
static bool kvm_sync_page_check(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2054
static int kvm_sync_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i)
arch/x86/kvm/mmu/mmu.c
2063
static int __kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2091
static int kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2115
static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2126
struct kvm_mmu_page *parent[PT64_ROOT_MAX_LEVEL];
arch/x86/kvm/mmu/mmu.c
2142
struct kvm_mmu_page *sp = pvec->page[n].sp;
arch/x86/kvm/mmu/mmu.c
2159
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2182
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2198
struct kvm_mmu_page *parent, bool can_yield)
arch/x86/kvm/mmu/mmu.c
2201
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2239
static void __clear_sp_write_flooding_count(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2255
static struct kvm_mmu_page *kvm_mmu_find_shadow_page(struct kvm *kvm,
arch/x86/kvm/mmu/mmu.c
2261
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2340
static struct kvm_mmu_page *kvm_mmu_alloc_shadow_page(struct kvm *kvm,
arch/x86/kvm/mmu/mmu.c
2346
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2376
static struct kvm_mmu_page *__kvm_mmu_get_shadow_page(struct kvm *kvm,
arch/x86/kvm/mmu/mmu.c
2383
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2404
static struct kvm_mmu_page *kvm_mmu_get_shadow_page(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2420
struct kvm_mmu_page *parent_sp = sptep_to_sp(sptep);
arch/x86/kvm/mmu/mmu.c
2463
static struct kvm_mmu_page *kvm_mmu_get_child_sp(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/mmu.c
2541
struct kvm_mmu_page *sp, bool flush)
arch/x86/kvm/mmu/mmu.c
2575
struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2584
struct kvm_mmu_page *child;
arch/x86/kvm/mmu/mmu.c
2603
static int mmu_page_zap_pte(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2607
struct kvm_mmu_page *child;
arch/x86/kvm/mmu/mmu.c
2635
struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2647
static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2657
struct kvm_mmu_page *parent,
arch/x86/kvm/mmu/mmu.c
2668
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2681
struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2745
static bool kvm_mmu_prepare_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
2757
struct kvm_mmu_page *sp, *nsp;
arch/x86/kvm/mmu/mmu.c
2783
struct kvm_mmu_page *sp, *tmp;
arch/x86/kvm/mmu/mmu.c
281
static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index);
arch/x86/kvm/mmu/mmu.c
286
struct kvm_mmu_page *sp = sptep_to_sp(sptep);
arch/x86/kvm/mmu/mmu.c
2873
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
2914
static void kvm_unsync_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
2932
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
3034
struct kvm_mmu_page *sp = sptep_to_sp(sptep);
arch/x86/kvm/mmu/mmu.c
3057
struct kvm_mmu_page *child;
arch/x86/kvm/mmu/mmu.c
3144
struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
3154
struct kvm_mmu_page *sp, u64 *sptep)
arch/x86/kvm/mmu/mmu.c
3181
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
3439
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
3662
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
370
struct kvm_mmu_page *sp = sptep_to_sp(sptep);
arch/x86/kvm/mmu/mmu.c
3786
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
3876
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
3904
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
4249
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
4285
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
454
struct kvm_mmu_page *sp = sptep_to_sp(sptep);
arch/x86/kvm/mmu/mmu.c
4768
struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/mmu.c
5113
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
5245
struct kvm_mmu_page *sp = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/mmu.c
6096
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
6175
static bool detect_write_flooding(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
6192
static bool detect_write_misaligned(struct kvm_mmu_page *sp, gpa_t gpa,
arch/x86/kvm/mmu/mmu.c
6213
static u64 *get_written_sptes(struct kvm_mmu_page *sp, gpa_t gpa, int *nspte)
arch/x86/kvm/mmu/mmu.c
6248
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
637
static bool sp_has_gptes(struct kvm_mmu_page *sp);
arch/x86/kvm/mmu/mmu.c
639
static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index)
arch/x86/kvm/mmu/mmu.c
6507
struct kvm_mmu_page *sp = sptep_to_sp(iterator.sptep);
arch/x86/kvm/mmu/mmu.c
656
static u32 kvm_mmu_page_get_access(struct kvm_mmu_page *sp, int index)
arch/x86/kvm/mmu/mmu.c
6721
struct kvm_mmu_page *sp, *node;
arch/x86/kvm/mmu/mmu.c
676
static void kvm_mmu_page_set_translation(struct kvm_mmu_page *sp, int index,
arch/x86/kvm/mmu/mmu.c
695
static void kvm_mmu_page_set_access(struct kvm_mmu_page *sp, int index,
arch/x86/kvm/mmu/mmu.c
7027
static struct kvm_mmu_page *shadow_mmu_get_sp_for_split(struct kvm *kvm, u64 *huge_sptep)
arch/x86/kvm/mmu/mmu.c
7029
struct kvm_mmu_page *huge_sp = sptep_to_sp(huge_sptep);
arch/x86/kvm/mmu/mmu.c
7061
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
7104
struct kvm_mmu_page *huge_sp = sptep_to_sp(huge_sptep);
arch/x86/kvm/mmu/mmu.c
7144
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
7261
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
7349
struct kvm_mmu_page *sp, *node;
arch/x86/kvm/mmu/mmu.c
7394
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
749
static void account_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
7606
sizeof(struct kvm_mmu_page),
arch/x86/kvm/mmu/mmu.c
7704
struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
7740
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/mmu.c
7778
struct kvm_mmu_page,
arch/x86/kvm/mmu/mmu.c
779
void track_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
799
static void account_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
808
static void unaccount_shadowed(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu.c
824
void untrack_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu.c
835
static void unaccount_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
150
static inline int kvm_mmu_page_as_id(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
155
static inline bool is_mirror_sp(const struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
160
static inline void kvm_mmu_alloc_external_spt(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
171
static inline gfn_t kvm_gfn_root_bits(const struct kvm *kvm, const struct kvm_mmu_page *root)
arch/x86/kvm/mmu/mmu_internal.h
184
struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/mmu_internal.h
409
void track_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmu_internal.h
411
void untrack_possible_nx_huge_page(struct kvm *kvm, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/mmutrace.h
163
TP_PROTO(struct kvm_mmu_page *sp, bool created),
arch/x86/kvm/mmu/mmutrace.h
182
TP_PROTO(struct kvm_mmu_page *sp),
arch/x86/kvm/mmu/mmutrace.h
197
TP_PROTO(struct kvm_mmu_page *sp),
arch/x86/kvm/mmu/mmutrace.h
203
TP_PROTO(struct kvm_mmu_page *sp),
arch/x86/kvm/mmu/mmutrace.h
209
TP_PROTO(struct kvm_mmu_page *sp),
arch/x86/kvm/mmu/paging_tmpl.h
152
struct kvm_mmu_page *sp, u64 *spte,
arch/x86/kvm/mmu/paging_tmpl.h
532
FNAME(prefetch_gpte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/paging_tmpl.h
574
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/paging_tmpl.h
617
struct kvm_mmu_page *sp = NULL;
arch/x86/kvm/mmu/paging_tmpl.h
847
static gpa_t FNAME(get_level1_sp_gpa)(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/paging_tmpl.h
895
static int FNAME(sync_spte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp, int i)
arch/x86/kvm/mmu/spte.c
143
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/spte.c
186
bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/spte.h
259
static inline struct kvm_mmu_page *to_shadow_page(hpa_t shadow_page)
arch/x86/kvm/mmu/spte.h
263
return (struct kvm_mmu_page *)page_private(page);
arch/x86/kvm/mmu/spte.h
266
static inline struct kvm_mmu_page *spte_to_child_sp(u64 spte)
arch/x86/kvm/mmu/spte.h
271
static inline struct kvm_mmu_page *sptep_to_sp(u64 *sptep)
arch/x86/kvm/mmu/spte.h
276
static inline struct kvm_mmu_page *root_to_sp(hpa_t root)
arch/x86/kvm/mmu/spte.h
295
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/spte.h
321
static inline bool sp_ad_disabled(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/spte.h
543
bool make_spte(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
arch/x86/kvm/mmu/tdp_iter.c
39
void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_iter.h
138
void tdp_iter_start(struct tdp_iter *iter, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1028
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1039
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1066
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1118
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1172
struct kvm_mmu_page *sp = sptep_to_sp(rcu_dereference(iter->sptep));
arch/x86/kvm/mmu/tdp_mmu.c
121
static struct kvm_mmu_page *tdp_mmu_next_root(struct kvm *kvm,
arch/x86/kvm/mmu/tdp_mmu.c
122
struct kvm_mmu_page *prev_root,
arch/x86/kvm/mmu/tdp_mmu.c
1238
struct kvm_mmu_page *sp, bool shared)
arch/x86/kvm/mmu/tdp_mmu.c
125
struct kvm_mmu_page *next_root;
arch/x86/kvm/mmu/tdp_mmu.c
1257
struct kvm_mmu_page *sp, bool shared);
arch/x86/kvm/mmu/tdp_mmu.c
1265
struct kvm_mmu_page *root = tdp_mmu_get_root_for_fault(vcpu, fault);
arch/x86/kvm/mmu/tdp_mmu.c
1268
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/tdp_mmu.c
1357
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1403
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1449
static bool wrprot_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1490
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1502
static struct kvm_mmu_page *tdp_mmu_alloc_sp_for_split(void)
arch/x86/kvm/mmu/tdp_mmu.c
1504
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/tdp_mmu.c
1521
struct kvm_mmu_page *sp, bool shared)
arch/x86/kvm/mmu/tdp_mmu.c
1559
struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1563
struct kvm_mmu_page *sp = NULL;
arch/x86/kvm/mmu/tdp_mmu.c
1645
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1658
static bool tdp_mmu_need_write_protect(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
1668
static void clear_dirty_gfn_range(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1706
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1714
static void clear_dirty_pt_masked(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1764
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1774
struct kvm_mmu_page *root = spte_to_child_sp(parent->old_spte);
arch/x86/kvm/mmu/tdp_mmu.c
1797
struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1871
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1883
static bool write_protect_gfn(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
1923
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
1942
struct kvm_mmu_page *root = root_to_sp(vcpu->arch.mmu->root.hpa);
arch/x86/kvm/mmu/tdp_mmu.c
1972
struct kvm_mmu_page *root = tdp_mmu_get_root(vcpu, KVM_DIRECT_ROOTS);
arch/x86/kvm/mmu/tdp_mmu.c
214
static struct kvm_mmu_page *tdp_mmu_alloc_sp(struct kvm_vcpu *vcpu)
arch/x86/kvm/mmu/tdp_mmu.c
216
struct kvm_mmu_page *sp;
arch/x86/kvm/mmu/tdp_mmu.c
224
static void tdp_mmu_init_sp(struct kvm_mmu_page *sp, tdp_ptep_t sptep,
arch/x86/kvm/mmu/tdp_mmu.c
239
static void tdp_mmu_init_child_sp(struct kvm_mmu_page *child_sp,
arch/x86/kvm/mmu/tdp_mmu.c
242
struct kvm_mmu_page *parent_sp;
arch/x86/kvm/mmu/tdp_mmu.c
259
struct kvm_mmu_page *root;
arch/x86/kvm/mmu/tdp_mmu.c
327
static void tdp_account_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
335
static void tdp_unaccount_mmu_page(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
349
static void tdp_mmu_unlink_sp(struct kvm *kvm, struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
400
struct kvm_mmu_page *sp = sptep_to_sp(rcu_dereference(pt));
arch/x86/kvm/mmu/tdp_mmu.c
501
struct kvm_mmu_page *sp = spte_to_child_sp(new_spte);
arch/x86/kvm/mmu/tdp_mmu.c
56
static void tdp_mmu_free_sp(struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
73
struct kvm_mmu_page *sp = container_of(head, struct kvm_mmu_page,
arch/x86/kvm/mmu/tdp_mmu.c
79
void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root)
arch/x86/kvm/mmu/tdp_mmu.c
853
static void __tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
876
static void tdp_mmu_zap_root(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
924
struct kvm_mmu_page *sp)
arch/x86/kvm/mmu/tdp_mmu.c
97
static bool tdp_mmu_root_match(struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.c
980
static bool tdp_mmu_zap_leafs(struct kvm *kvm, struct kvm_mmu_page *root,
arch/x86/kvm/mmu/tdp_mmu.h
117
static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return sp->tdp_mmu_page; }
arch/x86/kvm/mmu/tdp_mmu.h
119
static inline bool is_tdp_mmu_page(struct kvm_mmu_page *sp) { return false; }
arch/x86/kvm/mmu/tdp_mmu.h
15
__must_check static inline bool kvm_tdp_mmu_get_root(struct kvm_mmu_page *root)
arch/x86/kvm/mmu/tdp_mmu.h
20
void kvm_tdp_mmu_put_root(struct kvm *kvm, struct kvm_mmu_page *root);
arch/x86/kvm/mmu/tdp_mmu.h
48
static inline struct kvm_mmu_page *tdp_mmu_get_root_for_fault(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/tdp_mmu.h
57
static inline struct kvm_mmu_page *tdp_mmu_get_root(struct kvm_vcpu *vcpu,
arch/x86/kvm/mmu/tdp_mmu.h
68
struct kvm_mmu_page *sp);
arch/x86/kvm/vmx/vmx.c
3391
struct kvm_mmu_page *root;