|  | #ifndef __KVM_X86_MMU_H | 
|  | #define __KVM_X86_MMU_H | 
|  |  | 
|  | #include <linux/kvm_host.h> | 
|  |  | 
|  | static inline void kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | if (unlikely(vcpu->kvm->arch.n_free_mmu_pages < KVM_MIN_FREE_MMU_PAGES)) | 
|  | __kvm_mmu_free_some_pages(vcpu); | 
|  | } | 
|  |  | 
|  | static inline int kvm_mmu_reload(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | if (likely(vcpu->arch.mmu.root_hpa != INVALID_PAGE)) | 
|  | return 0; | 
|  |  | 
|  | return kvm_mmu_load(vcpu); | 
|  | } | 
|  |  | 
|  | static inline int is_long_mode(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | #ifdef CONFIG_X86_64 | 
|  | return vcpu->arch.shadow_efer & EFER_LME; | 
|  | #else | 
|  | return 0; | 
|  | #endif | 
|  | } | 
|  |  | 
|  | static inline int is_pae(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | return vcpu->arch.cr4 & X86_CR4_PAE; | 
|  | } | 
|  |  | 
|  | static inline int is_pse(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | return vcpu->arch.cr4 & X86_CR4_PSE; | 
|  | } | 
|  |  | 
|  | static inline int is_paging(struct kvm_vcpu *vcpu) | 
|  | { | 
|  | return vcpu->arch.cr0 & X86_CR0_PG; | 
|  | } | 
|  |  | 
|  | #endif |