KVM: arm64: Remove unused page-table code
authorWill Deacon <will@kernel.org>
Fri, 11 Sep 2020 13:25:27 +0000 (14:25 +0100)
committerMarc Zyngier <maz@kernel.org>
Fri, 11 Sep 2020 14:51:15 +0000 (15:51 +0100)
Now that KVM is using the generic page-table code to manage the guest
stage-2 page-tables, we can remove a bunch of unused macros, #defines
and static inline functions from the old implementation.

Signed-off-by: Will Deacon <will@kernel.org>
Signed-off-by: Marc Zyngier <maz@kernel.org>
Reviewed-by: Gavin Shan <gshan@redhat.com>
Cc: Marc Zyngier <maz@kernel.org>
Cc: Quentin Perret <qperret@google.com>
Link: https://lore.kernel.org/r/20200911132529.19844-20-will@kernel.org
arch/arm64/include/asm/kvm_mmu.h
arch/arm64/include/asm/pgtable-hwdef.h
arch/arm64/include/asm/pgtable-prot.h
arch/arm64/include/asm/stage2_pgtable.h
arch/arm64/kvm/mmu.c

index 42fb50c..c490fe8 100644 (file)
@@ -107,12 +107,6 @@ static __always_inline unsigned long __kern_hyp_va(unsigned long v)
 #define kvm_phys_size(kvm)             (_AC(1, ULL) << kvm_phys_shift(kvm))
 #define kvm_phys_mask(kvm)             (kvm_phys_size(kvm) - _AC(1, ULL))
 
-static inline bool kvm_page_empty(void *ptr)
-{
-       struct page *ptr_page = virt_to_page(ptr);
-       return page_count(ptr_page) == 1;
-}
-
 #include <asm/kvm_pgtable.h>
 #include <asm/stage2_pgtable.h>
 
@@ -135,123 +129,6 @@ int kvm_handle_guest_abort(struct kvm_vcpu *vcpu);
 phys_addr_t kvm_mmu_get_httbr(void);
 phys_addr_t kvm_get_idmap_vector(void);
 int kvm_mmu_init(void);
-#define kvm_mk_pmd(ptep)                                       \
-       __pmd(__phys_to_pmd_val(__pa(ptep)) | PMD_TYPE_TABLE)
-#define kvm_mk_pud(pmdp)                                       \
-       __pud(__phys_to_pud_val(__pa(pmdp)) | PMD_TYPE_TABLE)
-#define kvm_mk_p4d(pmdp)                                       \
-       __p4d(__phys_to_p4d_val(__pa(pmdp)) | PUD_TYPE_TABLE)
-
-#define kvm_set_pud(pudp, pud)         set_pud(pudp, pud)
-
-#define kvm_pfn_pte(pfn, prot)         pfn_pte(pfn, prot)
-#define kvm_pfn_pmd(pfn, prot)         pfn_pmd(pfn, prot)
-#define kvm_pfn_pud(pfn, prot)         pfn_pud(pfn, prot)
-
-#define kvm_pud_pfn(pud)               pud_pfn(pud)
-
-#define kvm_pmd_mkhuge(pmd)            pmd_mkhuge(pmd)
-#define kvm_pud_mkhuge(pud)            pud_mkhuge(pud)
-
-static inline pte_t kvm_s2pte_mkwrite(pte_t pte)
-{
-       pte_val(pte) |= PTE_S2_RDWR;
-       return pte;
-}
-
-static inline pmd_t kvm_s2pmd_mkwrite(pmd_t pmd)
-{
-       pmd_val(pmd) |= PMD_S2_RDWR;
-       return pmd;
-}
-
-static inline pud_t kvm_s2pud_mkwrite(pud_t pud)
-{
-       pud_val(pud) |= PUD_S2_RDWR;
-       return pud;
-}
-
-static inline pte_t kvm_s2pte_mkexec(pte_t pte)
-{
-       pte_val(pte) &= ~PTE_S2_XN;
-       return pte;
-}
-
-static inline pmd_t kvm_s2pmd_mkexec(pmd_t pmd)
-{
-       pmd_val(pmd) &= ~PMD_S2_XN;
-       return pmd;
-}
-
-static inline pud_t kvm_s2pud_mkexec(pud_t pud)
-{
-       pud_val(pud) &= ~PUD_S2_XN;
-       return pud;
-}
-
-static inline void kvm_set_s2pte_readonly(pte_t *ptep)
-{
-       pteval_t old_pteval, pteval;
-
-       pteval = READ_ONCE(pte_val(*ptep));
-       do {
-               old_pteval = pteval;
-               pteval &= ~PTE_S2_RDWR;
-               pteval |= PTE_S2_RDONLY;
-               pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval);
-       } while (pteval != old_pteval);
-}
-
-static inline bool kvm_s2pte_readonly(pte_t *ptep)
-{
-       return (READ_ONCE(pte_val(*ptep)) & PTE_S2_RDWR) == PTE_S2_RDONLY;
-}
-
-static inline bool kvm_s2pte_exec(pte_t *ptep)
-{
-       return !(READ_ONCE(pte_val(*ptep)) & PTE_S2_XN);
-}
-
-static inline void kvm_set_s2pmd_readonly(pmd_t *pmdp)
-{
-       kvm_set_s2pte_readonly((pte_t *)pmdp);
-}
-
-static inline bool kvm_s2pmd_readonly(pmd_t *pmdp)
-{
-       return kvm_s2pte_readonly((pte_t *)pmdp);
-}
-
-static inline bool kvm_s2pmd_exec(pmd_t *pmdp)
-{
-       return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN);
-}
-
-static inline void kvm_set_s2pud_readonly(pud_t *pudp)
-{
-       kvm_set_s2pte_readonly((pte_t *)pudp);
-}
-
-static inline bool kvm_s2pud_readonly(pud_t *pudp)
-{
-       return kvm_s2pte_readonly((pte_t *)pudp);
-}
-
-static inline bool kvm_s2pud_exec(pud_t *pudp)
-{
-       return !(READ_ONCE(pud_val(*pudp)) & PUD_S2_XN);
-}
-
-static inline pud_t kvm_s2pud_mkyoung(pud_t pud)
-{
-       return pud_mkyoung(pud);
-}
-
-static inline bool kvm_s2pud_young(pud_t pud)
-{
-       return pud_young(pud);
-}
-
 
 struct kvm;
 
@@ -293,30 +170,6 @@ static inline void __invalidate_icache_guest_page(kvm_pfn_t pfn,
        }
 }
 
-static inline void __kvm_flush_dcache_pte(pte_t pte)
-{
-       if (!cpus_have_const_cap(ARM64_HAS_STAGE2_FWB)) {
-               struct page *page = pte_page(pte);
-               kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
-       }
-}
-
-static inline void __kvm_flush_dcache_pmd(pmd_t pmd)
-{
-       if (!cpus_have_const_cap(ARM64_HAS_STAGE2_FWB)) {
-               struct page *page = pmd_page(pmd);
-               kvm_flush_dcache_to_poc(page_address(page), PMD_SIZE);
-       }
-}
-
-static inline void __kvm_flush_dcache_pud(pud_t pud)
-{
-       if (!cpus_have_const_cap(ARM64_HAS_STAGE2_FWB)) {
-               struct page *page = pud_page(pud);
-               kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE);
-       }
-}
-
 void kvm_set_way_flush(struct kvm_vcpu *vcpu);
 void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled);
 
@@ -477,30 +330,6 @@ static inline int hyp_map_aux_data(void)
 
 #define kvm_phys_to_vttbr(addr)                phys_to_ttbr(addr)
 
-/*
- * Get the magic number 'x' for VTTBR:BADDR of this KVM instance.
- * With v8.2 LVA extensions, 'x' should be a minimum of 6 with
- * 52bit IPS.
- */
-static inline int arm64_vttbr_x(u32 ipa_shift, u32 levels)
-{
-       int x = ARM64_VTTBR_X(ipa_shift, levels);
-
-       return (IS_ENABLED(CONFIG_ARM64_PA_BITS_52) && x < 6) ? 6 : x;
-}
-
-static inline u64 vttbr_baddr_mask(u32 ipa_shift, u32 levels)
-{
-       unsigned int x = arm64_vttbr_x(ipa_shift, levels);
-
-       return GENMASK_ULL(PHYS_MASK_SHIFT - 1, x);
-}
-
-static inline u64 kvm_vttbr_baddr_mask(struct kvm *kvm)
-{
-       return vttbr_baddr_mask(kvm_phys_shift(kvm), kvm_stage2_levels(kvm));
-}
-
 static __always_inline u64 kvm_get_vttbr(struct kvm_s2_mmu *mmu)
 {
        struct kvm_vmid *vmid = &mmu->vmid;
index 1a98935..2b5f822 100644 (file)
 #define PTE_CONT               (_AT(pteval_t, 1) << 52)        /* Contiguous range */
 #define PTE_PXN                        (_AT(pteval_t, 1) << 53)        /* Privileged XN */
 #define PTE_UXN                        (_AT(pteval_t, 1) << 54)        /* User XN */
-#define PTE_HYP_XN             (_AT(pteval_t, 1) << 54)        /* HYP XN */
 
 #define PTE_ADDR_LOW           (((_AT(pteval_t, 1) << (48 - PAGE_SHIFT)) - 1) << PAGE_SHIFT)
 #ifdef CONFIG_ARM64_PA_BITS_52
 #define PTE_ATTRINDX_MASK      (_AT(pteval_t, 7) << 2)
 
 /*
- * 2nd stage PTE definitions
- */
-#define PTE_S2_RDONLY          (_AT(pteval_t, 1) << 6)   /* HAP[2:1] */
-#define PTE_S2_RDWR            (_AT(pteval_t, 3) << 6)   /* HAP[2:1] */
-#define PTE_S2_XN              (_AT(pteval_t, 2) << 53)  /* XN[1:0] */
-#define PTE_S2_SW_RESVD                (_AT(pteval_t, 15) << 55) /* Reserved for SW */
-
-#define PMD_S2_RDONLY          (_AT(pmdval_t, 1) << 6)   /* HAP[2:1] */
-#define PMD_S2_RDWR            (_AT(pmdval_t, 3) << 6)   /* HAP[2:1] */
-#define PMD_S2_XN              (_AT(pmdval_t, 2) << 53)  /* XN[1:0] */
-#define PMD_S2_SW_RESVD                (_AT(pmdval_t, 15) << 55) /* Reserved for SW */
-
-#define PUD_S2_RDONLY          (_AT(pudval_t, 1) << 6)   /* HAP[2:1] */
-#define PUD_S2_RDWR            (_AT(pudval_t, 3) << 6)   /* HAP[2:1] */
-#define PUD_S2_XN              (_AT(pudval_t, 2) << 53)  /* XN[1:0] */
-
-/*
  * Memory Attribute override for Stage-2 (MemAttr[3:0])
  */
 #define PTE_S2_MEMATTR(t)      (_AT(pteval_t, (t)) << 2)
index 88acd7e..8f094c4 100644 (file)
@@ -73,19 +73,6 @@ extern bool arm64_use_ng_mappings;
                __val;                                                  \
         })
 
-#define PAGE_S2_XN                                                     \
-       ({                                                              \
-               u64 __val;                                              \
-               if (cpus_have_const_cap(ARM64_HAS_CACHE_DIC))           \
-                       __val = 0;                                      \
-               else                                                    \
-                       __val = PTE_S2_XN;                              \
-               __val;                                                  \
-       })
-
-#define PAGE_S2                        __pgprot(_PROT_DEFAULT | PAGE_S2_MEMATTR(NORMAL) | PTE_S2_RDONLY | PAGE_S2_XN)
-#define PAGE_S2_DEVICE         __pgprot(_PROT_DEFAULT | PAGE_S2_MEMATTR(DEVICE_nGnRE) | PTE_S2_RDONLY | PTE_S2_XN)
-
 #define PAGE_NONE              __pgprot(((_PAGE_DEFAULT) & ~PTE_VALID) | PTE_PROT_NONE | PTE_RDONLY | PTE_NG | PTE_PXN | PTE_UXN)
 /* shared+writable pages are clean by default, hence PTE_RDONLY|PTE_WRITE */
 #define PAGE_SHARED            __pgprot(_PAGE_DEFAULT | PTE_USER | PTE_RDONLY | PTE_NG | PTE_PXN | PTE_UXN | PTE_WRITE)
index 996bf98..fe341a6 100644 (file)
@@ -8,7 +8,6 @@
 #ifndef __ARM64_S2_PGTABLE_H_
 #define __ARM64_S2_PGTABLE_H_
 
-#include <linux/hugetlb.h>
 #include <linux/pgtable.h>
 
 /*
 #define stage2_pgdir_mask(kvm)         ~(stage2_pgdir_size(kvm) - 1)
 
 /*
- * The number of PTRS across all concatenated stage2 tables given by the
- * number of bits resolved at the initial level.
- * If we force more levels than necessary, we may have (stage2_pgdir_shift > IPA),
- * in which case, stage2_pgd_ptrs will have one entry.
- */
-#define pgd_ptrs_shift(ipa, pgdir_shift)       \
-       ((ipa) > (pgdir_shift) ? ((ipa) - (pgdir_shift)) : 0)
-#define __s2_pgd_ptrs(ipa, lvls)               \
-       (1 << (pgd_ptrs_shift((ipa), pt_levels_pgdir_shift(lvls))))
-#define __s2_pgd_size(ipa, lvls)       (__s2_pgd_ptrs((ipa), (lvls)) * sizeof(pgd_t))
-
-#define stage2_pgd_ptrs(kvm)           __s2_pgd_ptrs(kvm_phys_shift(kvm), kvm_stage2_levels(kvm))
-#define stage2_pgd_size(kvm)           __s2_pgd_size(kvm_phys_shift(kvm), kvm_stage2_levels(kvm))
-
-/*
  * kvm_mmmu_cache_min_pages() is the number of pages required to install
  * a stage-2 translation. We pre-allocate the entry level page table at
  * the VM creation.
  */
 #define kvm_mmu_cache_min_pages(kvm)   (kvm_stage2_levels(kvm) - 1)
 
-/* Stage2 PUD definitions when the level is present */
-static inline bool kvm_stage2_has_pud(struct kvm *kvm)
-{
-       return (CONFIG_PGTABLE_LEVELS > 3) && (kvm_stage2_levels(kvm) > 3);
-}
-
-#define S2_PUD_SHIFT                   ARM64_HW_PGTABLE_LEVEL_SHIFT(1)
-#define S2_PUD_SIZE                    (1UL << S2_PUD_SHIFT)
-#define S2_PUD_MASK                    (~(S2_PUD_SIZE - 1))
-
-#define stage2_pgd_none(kvm, pgd)              pgd_none(pgd)
-#define stage2_pgd_clear(kvm, pgd)             pgd_clear(pgd)
-#define stage2_pgd_present(kvm, pgd)           pgd_present(pgd)
-#define stage2_pgd_populate(kvm, pgd, p4d)     pgd_populate(NULL, pgd, p4d)
-
-static inline p4d_t *stage2_p4d_offset(struct kvm *kvm,
-                                      pgd_t *pgd, unsigned long address)
-{
-       return p4d_offset(pgd, address);
-}
-
-static inline void stage2_p4d_free(struct kvm *kvm, p4d_t *p4d)
-{
-}
-
-static inline bool stage2_p4d_table_empty(struct kvm *kvm, p4d_t *p4dp)
-{
-       return false;
-}
-
-static inline phys_addr_t stage2_p4d_addr_end(struct kvm *kvm,
-                                             phys_addr_t addr, phys_addr_t end)
-{
-       return end;
-}
-
-static inline bool stage2_p4d_none(struct kvm *kvm, p4d_t p4d)
-{
-       if (kvm_stage2_has_pud(kvm))
-               return p4d_none(p4d);
-       else
-               return 0;
-}
-
-static inline void stage2_p4d_clear(struct kvm *kvm, p4d_t *p4dp)
-{
-       if (kvm_stage2_has_pud(kvm))
-               p4d_clear(p4dp);
-}
-
-static inline bool stage2_p4d_present(struct kvm *kvm, p4d_t p4d)
-{
-       if (kvm_stage2_has_pud(kvm))
-               return p4d_present(p4d);
-       else
-               return 1;
-}
-
-static inline void stage2_p4d_populate(struct kvm *kvm, p4d_t *p4d, pud_t *pud)
-{
-       if (kvm_stage2_has_pud(kvm))
-               p4d_populate(NULL, p4d, pud);
-}
-
-static inline pud_t *stage2_pud_offset(struct kvm *kvm,
-                                      p4d_t *p4d, unsigned long address)
-{
-       if (kvm_stage2_has_pud(kvm))
-               return pud_offset(p4d, address);
-       else
-               return (pud_t *)p4d;
-}
-
-static inline void stage2_pud_free(struct kvm *kvm, pud_t *pud)
-{
-       if (kvm_stage2_has_pud(kvm))
-               free_page((unsigned long)pud);
-}
-
-static inline bool stage2_pud_table_empty(struct kvm *kvm, pud_t *pudp)
-{
-       if (kvm_stage2_has_pud(kvm))
-               return kvm_page_empty(pudp);
-       else
-               return false;
-}
-
-static inline phys_addr_t
-stage2_pud_addr_end(struct kvm *kvm, phys_addr_t addr, phys_addr_t end)
-{
-       if (kvm_stage2_has_pud(kvm)) {
-               phys_addr_t boundary = (addr + S2_PUD_SIZE) & S2_PUD_MASK;
-
-               return (boundary - 1 < end - 1) ? boundary : end;
-       } else {
-               return end;
-       }
-}
-
-/* Stage2 PMD definitions when the level is present */
-static inline bool kvm_stage2_has_pmd(struct kvm *kvm)
-{
-       return (CONFIG_PGTABLE_LEVELS > 2) && (kvm_stage2_levels(kvm) > 2);
-}
-
-#define S2_PMD_SHIFT                   ARM64_HW_PGTABLE_LEVEL_SHIFT(2)
-#define S2_PMD_SIZE                    (1UL << S2_PMD_SHIFT)
-#define S2_PMD_MASK                    (~(S2_PMD_SIZE - 1))
-
-static inline bool stage2_pud_none(struct kvm *kvm, pud_t pud)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               return pud_none(pud);
-       else
-               return 0;
-}
-
-static inline void stage2_pud_clear(struct kvm *kvm, pud_t *pud)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               pud_clear(pud);
-}
-
-static inline bool stage2_pud_present(struct kvm *kvm, pud_t pud)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               return pud_present(pud);
-       else
-               return 1;
-}
-
-static inline void stage2_pud_populate(struct kvm *kvm, pud_t *pud, pmd_t *pmd)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               pud_populate(NULL, pud, pmd);
-}
-
-static inline pmd_t *stage2_pmd_offset(struct kvm *kvm,
-                                      pud_t *pud, unsigned long address)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               return pmd_offset(pud, address);
-       else
-               return (pmd_t *)pud;
-}
-
-static inline void stage2_pmd_free(struct kvm *kvm, pmd_t *pmd)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               free_page((unsigned long)pmd);
-}
-
-static inline bool stage2_pud_huge(struct kvm *kvm, pud_t pud)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               return pud_huge(pud);
-       else
-               return 0;
-}
-
-static inline bool stage2_pmd_table_empty(struct kvm *kvm, pmd_t *pmdp)
-{
-       if (kvm_stage2_has_pmd(kvm))
-               return kvm_page_empty(pmdp);
-       else
-               return 0;
-}
-
-static inline phys_addr_t
-stage2_pmd_addr_end(struct kvm *kvm, phys_addr_t addr, phys_addr_t end)
-{
-       if (kvm_stage2_has_pmd(kvm)) {
-               phys_addr_t boundary = (addr + S2_PMD_SIZE) & S2_PMD_MASK;
-
-               return (boundary - 1 < end - 1) ? boundary : end;
-       } else {
-               return end;
-       }
-}
-
-static inline bool stage2_pte_table_empty(struct kvm *kvm, pte_t *ptep)
-{
-       return kvm_page_empty(ptep);
-}
-
-static inline unsigned long stage2_pgd_index(struct kvm *kvm, phys_addr_t addr)
-{
-       return (((addr) >> stage2_pgdir_shift(kvm)) & (stage2_pgd_ptrs(kvm) - 1));
-}
-
 static inline phys_addr_t
 stage2_pgd_addr_end(struct kvm *kvm, phys_addr_t addr, phys_addr_t end)
 {
@@ -256,13 +50,4 @@ stage2_pgd_addr_end(struct kvm *kvm, phys_addr_t addr, phys_addr_t end)
        return (boundary - 1 < end - 1) ? boundary : end;
 }
 
-/*
- * Level values for the ARMv8.4-TTL extension, mapping PUD/PMD/PTE and
- * the architectural page-table level.
- */
-#define S2_NO_LEVEL_HINT       0
-#define S2_PUD_LEVEL           1
-#define S2_PMD_LEVEL           2
-#define S2_PTE_LEVEL           3
-
 #endif /* __ARM64_S2_PGTABLE_H_ */
index ed5aefd..dd0ea7d 100644 (file)
@@ -31,13 +31,6 @@ static phys_addr_t hyp_idmap_vector;
 
 static unsigned long io_map_base;
 
-#define KVM_S2PTE_FLAG_IS_IOMAP                (1UL << 0)
-#define KVM_S2_FLAG_LOGGING_ACTIVE     (1UL << 1)
-
-static bool is_iomap(unsigned long flags)
-{
-       return flags & KVM_S2PTE_FLAG_IS_IOMAP;
-}
 
 /*
  * Release kvm_mmu_lock periodically if the memory region is large. Otherwise,
@@ -90,154 +83,11 @@ void kvm_flush_remote_tlbs(struct kvm *kvm)
        kvm_call_hyp(__kvm_tlb_flush_vmid, &kvm->arch.mmu);
 }
 
-static void kvm_tlb_flush_vmid_ipa(struct kvm_s2_mmu *mmu, phys_addr_t ipa,
-                                  int level)
-{
-       kvm_call_hyp(__kvm_tlb_flush_vmid_ipa, mmu, ipa, level);
-}
-
-/*
- * D-Cache management functions. They take the page table entries by
- * value, as they are flushing the cache using the kernel mapping (or
- * kmap on 32bit).
- */
-static void kvm_flush_dcache_pte(pte_t pte)
-{
-       __kvm_flush_dcache_pte(pte);
-}
-
-static void kvm_flush_dcache_pmd(pmd_t pmd)
-{
-       __kvm_flush_dcache_pmd(pmd);
-}
-
-static void kvm_flush_dcache_pud(pud_t pud)
-{
-       __kvm_flush_dcache_pud(pud);
-}
-
 static bool kvm_is_device_pfn(unsigned long pfn)
 {
        return !pfn_valid(pfn);
 }
 
-/**
- * stage2_dissolve_pmd() - clear and flush huge PMD entry
- * @mmu:       pointer to mmu structure to operate on
- * @addr:      IPA
- * @pmd:       pmd pointer for IPA
- *
- * Function clears a PMD entry, flushes addr 1st and 2nd stage TLBs.
- */
-static void stage2_dissolve_pmd(struct kvm_s2_mmu *mmu, phys_addr_t addr, pmd_t *pmd)
-{
-       if (!pmd_thp_or_huge(*pmd))
-               return;
-
-       pmd_clear(pmd);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PMD_LEVEL);
-       put_page(virt_to_page(pmd));
-}
-
-/**
- * stage2_dissolve_pud() - clear and flush huge PUD entry
- * @mmu:       pointer to mmu structure to operate on
- * @addr:      IPA
- * @pud:       pud pointer for IPA
- *
- * Function clears a PUD entry, flushes addr 1st and 2nd stage TLBs.
- */
-static void stage2_dissolve_pud(struct kvm_s2_mmu *mmu, phys_addr_t addr, pud_t *pudp)
-{
-       struct kvm *kvm = mmu->kvm;
-
-       if (!stage2_pud_huge(kvm, *pudp))
-               return;
-
-       stage2_pud_clear(kvm, pudp);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PUD_LEVEL);
-       put_page(virt_to_page(pudp));
-}
-
-static void clear_stage2_pgd_entry(struct kvm_s2_mmu *mmu, pgd_t *pgd, phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       p4d_t *p4d_table __maybe_unused = stage2_p4d_offset(kvm, pgd, 0UL);
-       stage2_pgd_clear(kvm, pgd);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_NO_LEVEL_HINT);
-       stage2_p4d_free(kvm, p4d_table);
-       put_page(virt_to_page(pgd));
-}
-
-static void clear_stage2_p4d_entry(struct kvm_s2_mmu *mmu, p4d_t *p4d, phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pud_table __maybe_unused = stage2_pud_offset(kvm, p4d, 0);
-       stage2_p4d_clear(kvm, p4d);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_NO_LEVEL_HINT);
-       stage2_pud_free(kvm, pud_table);
-       put_page(virt_to_page(p4d));
-}
-
-static void clear_stage2_pud_entry(struct kvm_s2_mmu *mmu, pud_t *pud, phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       pmd_t *pmd_table __maybe_unused = stage2_pmd_offset(kvm, pud, 0);
-
-       VM_BUG_ON(stage2_pud_huge(kvm, *pud));
-       stage2_pud_clear(kvm, pud);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_NO_LEVEL_HINT);
-       stage2_pmd_free(kvm, pmd_table);
-       put_page(virt_to_page(pud));
-}
-
-static void clear_stage2_pmd_entry(struct kvm_s2_mmu *mmu, pmd_t *pmd, phys_addr_t addr)
-{
-       pte_t *pte_table = pte_offset_kernel(pmd, 0);
-       VM_BUG_ON(pmd_thp_or_huge(*pmd));
-       pmd_clear(pmd);
-       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_NO_LEVEL_HINT);
-       free_page((unsigned long)pte_table);
-       put_page(virt_to_page(pmd));
-}
-
-static inline void kvm_set_pte(pte_t *ptep, pte_t new_pte)
-{
-       WRITE_ONCE(*ptep, new_pte);
-       dsb(ishst);
-}
-
-static inline void kvm_set_pmd(pmd_t *pmdp, pmd_t new_pmd)
-{
-       WRITE_ONCE(*pmdp, new_pmd);
-       dsb(ishst);
-}
-
-static inline void kvm_pmd_populate(pmd_t *pmdp, pte_t *ptep)
-{
-       kvm_set_pmd(pmdp, kvm_mk_pmd(ptep));
-}
-
-static inline void kvm_pud_populate(pud_t *pudp, pmd_t *pmdp)
-{
-       WRITE_ONCE(*pudp, kvm_mk_pud(pmdp));
-       dsb(ishst);
-}
-
-static inline void kvm_p4d_populate(p4d_t *p4dp, pud_t *pudp)
-{
-       WRITE_ONCE(*p4dp, kvm_mk_p4d(pudp));
-       dsb(ishst);
-}
-
-static inline void kvm_pgd_populate(pgd_t *pgdp, p4d_t *p4dp)
-{
-#ifndef __PAGETABLE_P4D_FOLDED
-       WRITE_ONCE(*pgdp, kvm_mk_pgd(p4dp));
-       dsb(ishst);
-#endif
-}
-
 /*
  * Unmapping vs dcache management:
  *
@@ -262,108 +112,6 @@ static inline void kvm_pgd_populate(pgd_t *pgdp, p4d_t *p4dp)
  * we then fully enforce cacheability of RAM, no matter what the guest
  * does.
  */
-static void unmap_stage2_ptes(struct kvm_s2_mmu *mmu, pmd_t *pmd,
-                      phys_addr_t addr, phys_addr_t end)
-{
-       phys_addr_t start_addr = addr;
-       pte_t *pte, *start_pte;
-
-       start_pte = pte = pte_offset_kernel(pmd, addr);
-       do {
-               if (!pte_none(*pte)) {
-                       pte_t old_pte = *pte;
-
-                       kvm_set_pte(pte, __pte(0));
-                       kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PTE_LEVEL);
-
-                       /* No need to invalidate the cache for device mappings */
-                       if (!kvm_is_device_pfn(pte_pfn(old_pte)))
-                               kvm_flush_dcache_pte(old_pte);
-
-                       put_page(virt_to_page(pte));
-               }
-       } while (pte++, addr += PAGE_SIZE, addr != end);
-
-       if (stage2_pte_table_empty(mmu->kvm, start_pte))
-               clear_stage2_pmd_entry(mmu, pmd, start_addr);
-}
-
-static void unmap_stage2_pmds(struct kvm_s2_mmu *mmu, pud_t *pud,
-                      phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       phys_addr_t next, start_addr = addr;
-       pmd_t *pmd, *start_pmd;
-
-       start_pmd = pmd = stage2_pmd_offset(kvm, pud, addr);
-       do {
-               next = stage2_pmd_addr_end(kvm, addr, end);
-               if (!pmd_none(*pmd)) {
-                       if (pmd_thp_or_huge(*pmd)) {
-                               pmd_t old_pmd = *pmd;
-
-                               pmd_clear(pmd);
-                               kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PMD_LEVEL);
-
-                               kvm_flush_dcache_pmd(old_pmd);
-
-                               put_page(virt_to_page(pmd));
-                       } else {
-                               unmap_stage2_ptes(mmu, pmd, addr, next);
-                       }
-               }
-       } while (pmd++, addr = next, addr != end);
-
-       if (stage2_pmd_table_empty(kvm, start_pmd))
-               clear_stage2_pud_entry(mmu, pud, start_addr);
-}
-
-static void unmap_stage2_puds(struct kvm_s2_mmu *mmu, p4d_t *p4d,
-                      phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       phys_addr_t next, start_addr = addr;
-       pud_t *pud, *start_pud;
-
-       start_pud = pud = stage2_pud_offset(kvm, p4d, addr);
-       do {
-               next = stage2_pud_addr_end(kvm, addr, end);
-               if (!stage2_pud_none(kvm, *pud)) {
-                       if (stage2_pud_huge(kvm, *pud)) {
-                               pud_t old_pud = *pud;
-
-                               stage2_pud_clear(kvm, pud);
-                               kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PUD_LEVEL);
-                               kvm_flush_dcache_pud(old_pud);
-                               put_page(virt_to_page(pud));
-                       } else {
-                               unmap_stage2_pmds(mmu, pud, addr, next);
-                       }
-               }
-       } while (pud++, addr = next, addr != end);
-
-       if (stage2_pud_table_empty(kvm, start_pud))
-               clear_stage2_p4d_entry(mmu, p4d, start_addr);
-}
-
-static void unmap_stage2_p4ds(struct kvm_s2_mmu *mmu, pgd_t *pgd,
-                      phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       phys_addr_t next, start_addr = addr;
-       p4d_t *p4d, *start_p4d;
-
-       start_p4d = p4d = stage2_p4d_offset(kvm, pgd, addr);
-       do {
-               next = stage2_p4d_addr_end(kvm, addr, end);
-               if (!stage2_p4d_none(kvm, *p4d))
-                       unmap_stage2_puds(mmu, p4d, addr, next);
-       } while (p4d++, addr = next, addr != end);
-
-       if (stage2_p4d_table_empty(kvm, start_p4d))
-               clear_stage2_pgd_entry(mmu, pgd, start_addr);
-}
-
 /**
  * unmap_stage2_range -- Clear stage2 page table entries to unmap a range
  * @kvm:   The VM pointer
@@ -392,71 +140,6 @@ static void unmap_stage2_range(struct kvm_s2_mmu *mmu, phys_addr_t start, u64 si
        __unmap_stage2_range(mmu, start, size, true);
 }
 
-static void stage2_flush_ptes(struct kvm_s2_mmu *mmu, pmd_t *pmd,
-                             phys_addr_t addr, phys_addr_t end)
-{
-       pte_t *pte;
-
-       pte = pte_offset_kernel(pmd, addr);
-       do {
-               if (!pte_none(*pte) && !kvm_is_device_pfn(pte_pfn(*pte)))
-                       kvm_flush_dcache_pte(*pte);
-       } while (pte++, addr += PAGE_SIZE, addr != end);
-}
-
-static void stage2_flush_pmds(struct kvm_s2_mmu *mmu, pud_t *pud,
-                             phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       pmd_t *pmd;
-       phys_addr_t next;
-
-       pmd = stage2_pmd_offset(kvm, pud, addr);
-       do {
-               next = stage2_pmd_addr_end(kvm, addr, end);
-               if (!pmd_none(*pmd)) {
-                       if (pmd_thp_or_huge(*pmd))
-                               kvm_flush_dcache_pmd(*pmd);
-                       else
-                               stage2_flush_ptes(mmu, pmd, addr, next);
-               }
-       } while (pmd++, addr = next, addr != end);
-}
-
-static void stage2_flush_puds(struct kvm_s2_mmu *mmu, p4d_t *p4d,
-                             phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pud;
-       phys_addr_t next;
-
-       pud = stage2_pud_offset(kvm, p4d, addr);
-       do {
-               next = stage2_pud_addr_end(kvm, addr, end);
-               if (!stage2_pud_none(kvm, *pud)) {
-                       if (stage2_pud_huge(kvm, *pud))
-                               kvm_flush_dcache_pud(*pud);
-                       else
-                               stage2_flush_pmds(mmu, pud, addr, next);
-               }
-       } while (pud++, addr = next, addr != end);
-}
-
-static void stage2_flush_p4ds(struct kvm_s2_mmu *mmu, pgd_t *pgd,
-                             phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       p4d_t *p4d;
-       phys_addr_t next;
-
-       p4d = stage2_p4d_offset(kvm, pgd, addr);
-       do {
-               next = stage2_p4d_addr_end(kvm, addr, end);
-               if (!stage2_p4d_none(kvm, *p4d))
-                       stage2_flush_puds(mmu, p4d, addr, next);
-       } while (p4d++, addr = next, addr != end);
-}
-
 static void stage2_flush_memslot(struct kvm *kvm,
                                 struct kvm_memory_slot *memslot)
 {
@@ -805,348 +488,6 @@ void kvm_free_stage2_pgd(struct kvm_s2_mmu *mmu)
        }
 }
 
-static p4d_t *stage2_get_p4d(struct kvm_s2_mmu *mmu, struct kvm_mmu_memory_cache *cache,
-                            phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       pgd_t *pgd;
-       p4d_t *p4d;
-
-       pgd = mmu->pgd + stage2_pgd_index(kvm, addr);
-       if (stage2_pgd_none(kvm, *pgd)) {
-               if (!cache)
-                       return NULL;
-               p4d = kvm_mmu_memory_cache_alloc(cache);
-               stage2_pgd_populate(kvm, pgd, p4d);
-               get_page(virt_to_page(pgd));
-       }
-
-       return stage2_p4d_offset(kvm, pgd, addr);
-}
-
-static pud_t *stage2_get_pud(struct kvm_s2_mmu *mmu, struct kvm_mmu_memory_cache *cache,
-                            phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       p4d_t *p4d;
-       pud_t *pud;
-
-       p4d = stage2_get_p4d(mmu, cache, addr);
-       if (stage2_p4d_none(kvm, *p4d)) {
-               if (!cache)
-                       return NULL;
-               pud = kvm_mmu_memory_cache_alloc(cache);
-               stage2_p4d_populate(kvm, p4d, pud);
-               get_page(virt_to_page(p4d));
-       }
-
-       return stage2_pud_offset(kvm, p4d, addr);
-}
-
-static pmd_t *stage2_get_pmd(struct kvm_s2_mmu *mmu, struct kvm_mmu_memory_cache *cache,
-                            phys_addr_t addr)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pud;
-       pmd_t *pmd;
-
-       pud = stage2_get_pud(mmu, cache, addr);
-       if (!pud || stage2_pud_huge(kvm, *pud))
-               return NULL;
-
-       if (stage2_pud_none(kvm, *pud)) {
-               if (!cache)
-                       return NULL;
-               pmd = kvm_mmu_memory_cache_alloc(cache);
-               stage2_pud_populate(kvm, pud, pmd);
-               get_page(virt_to_page(pud));
-       }
-
-       return stage2_pmd_offset(kvm, pud, addr);
-}
-
-static int stage2_set_pmd_huge(struct kvm_s2_mmu *mmu,
-                              struct kvm_mmu_memory_cache *cache,
-                              phys_addr_t addr, const pmd_t *new_pmd)
-{
-       pmd_t *pmd, old_pmd;
-
-retry:
-       pmd = stage2_get_pmd(mmu, cache, addr);
-       VM_BUG_ON(!pmd);
-
-       old_pmd = *pmd;
-       /*
-        * Multiple vcpus faulting on the same PMD entry, can
-        * lead to them sequentially updating the PMD with the
-        * same value. Following the break-before-make
-        * (pmd_clear() followed by tlb_flush()) process can
-        * hinder forward progress due to refaults generated
-        * on missing translations.
-        *
-        * Skip updating the page table if the entry is
-        * unchanged.
-        */
-       if (pmd_val(old_pmd) == pmd_val(*new_pmd))
-               return 0;
-
-       if (pmd_present(old_pmd)) {
-               /*
-                * If we already have PTE level mapping for this block,
-                * we must unmap it to avoid inconsistent TLB state and
-                * leaking the table page. We could end up in this situation
-                * if the memory slot was marked for dirty logging and was
-                * reverted, leaving PTE level mappings for the pages accessed
-                * during the period. So, unmap the PTE level mapping for this
-                * block and retry, as we could have released the upper level
-                * table in the process.
-                *
-                * Normal THP split/merge follows mmu_notifier callbacks and do
-                * get handled accordingly.
-                */
-               if (!pmd_thp_or_huge(old_pmd)) {
-                       unmap_stage2_range(mmu, addr & S2_PMD_MASK, S2_PMD_SIZE);
-                       goto retry;
-               }
-               /*
-                * Mapping in huge pages should only happen through a
-                * fault.  If a page is merged into a transparent huge
-                * page, the individual subpages of that huge page
-                * should be unmapped through MMU notifiers before we
-                * get here.
-                *
-                * Merging of CompoundPages is not supported; they
-                * should become splitting first, unmapped, merged,
-                * and mapped back in on-demand.
-                */
-               WARN_ON_ONCE(pmd_pfn(old_pmd) != pmd_pfn(*new_pmd));
-               pmd_clear(pmd);
-               kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PMD_LEVEL);
-       } else {
-               get_page(virt_to_page(pmd));
-       }
-
-       kvm_set_pmd(pmd, *new_pmd);
-       return 0;
-}
-
-static int stage2_set_pud_huge(struct kvm_s2_mmu *mmu,
-                              struct kvm_mmu_memory_cache *cache,
-                              phys_addr_t addr, const pud_t *new_pudp)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pudp, old_pud;
-
-retry:
-       pudp = stage2_get_pud(mmu, cache, addr);
-       VM_BUG_ON(!pudp);
-
-       old_pud = *pudp;
-
-       /*
-        * A large number of vcpus faulting on the same stage 2 entry,
-        * can lead to a refault due to the stage2_pud_clear()/tlb_flush().
-        * Skip updating the page tables if there is no change.
-        */
-       if (pud_val(old_pud) == pud_val(*new_pudp))
-               return 0;
-
-       if (stage2_pud_present(kvm, old_pud)) {
-               /*
-                * If we already have table level mapping for this block, unmap
-                * the range for this block and retry.
-                */
-               if (!stage2_pud_huge(kvm, old_pud)) {
-                       unmap_stage2_range(mmu, addr & S2_PUD_MASK, S2_PUD_SIZE);
-                       goto retry;
-               }
-
-               WARN_ON_ONCE(kvm_pud_pfn(old_pud) != kvm_pud_pfn(*new_pudp));
-               stage2_pud_clear(kvm, pudp);
-               kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PUD_LEVEL);
-       } else {
-               get_page(virt_to_page(pudp));
-       }
-
-       kvm_set_pud(pudp, *new_pudp);
-       return 0;
-}
-
-/*
- * stage2_get_leaf_entry - walk the stage2 VM page tables and return
- * true if a valid and present leaf-entry is found. A pointer to the
- * leaf-entry is returned in the appropriate level variable - pudpp,
- * pmdpp, ptepp.
- */
-static bool stage2_get_leaf_entry(struct kvm_s2_mmu *mmu, phys_addr_t addr,
-                                 pud_t **pudpp, pmd_t **pmdpp, pte_t **ptepp)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pudp;
-       pmd_t *pmdp;
-       pte_t *ptep;
-
-       *pudpp = NULL;
-       *pmdpp = NULL;
-       *ptepp = NULL;
-
-       pudp = stage2_get_pud(mmu, NULL, addr);
-       if (!pudp || stage2_pud_none(kvm, *pudp) || !stage2_pud_present(kvm, *pudp))
-               return false;
-
-       if (stage2_pud_huge(kvm, *pudp)) {
-               *pudpp = pudp;
-               return true;
-       }
-
-       pmdp = stage2_pmd_offset(kvm, pudp, addr);
-       if (!pmdp || pmd_none(*pmdp) || !pmd_present(*pmdp))
-               return false;
-
-       if (pmd_thp_or_huge(*pmdp)) {
-               *pmdpp = pmdp;
-               return true;
-       }
-
-       ptep = pte_offset_kernel(pmdp, addr);
-       if (!ptep || pte_none(*ptep) || !pte_present(*ptep))
-               return false;
-
-       *ptepp = ptep;
-       return true;
-}
-
-static bool stage2_is_exec(struct kvm_s2_mmu *mmu, phys_addr_t addr, unsigned long sz)
-{
-       pud_t *pudp;
-       pmd_t *pmdp;
-       pte_t *ptep;
-       bool found;
-
-       found = stage2_get_leaf_entry(mmu, addr, &pudp, &pmdp, &ptep);
-       if (!found)
-               return false;
-
-       if (pudp)
-               return sz <= PUD_SIZE && kvm_s2pud_exec(pudp);
-       else if (pmdp)
-               return sz <= PMD_SIZE && kvm_s2pmd_exec(pmdp);
-       else
-               return sz == PAGE_SIZE && kvm_s2pte_exec(ptep);
-}
-
-static int stage2_set_pte(struct kvm_s2_mmu *mmu,
-                         struct kvm_mmu_memory_cache *cache,
-                         phys_addr_t addr, const pte_t *new_pte,
-                         unsigned long flags)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pud;
-       pmd_t *pmd;
-       pte_t *pte, old_pte;
-       bool iomap = flags & KVM_S2PTE_FLAG_IS_IOMAP;
-       bool logging_active = flags & KVM_S2_FLAG_LOGGING_ACTIVE;
-
-       VM_BUG_ON(logging_active && !cache);
-
-       /* Create stage-2 page table mapping - Levels 0 and 1 */
-       pud = stage2_get_pud(mmu, cache, addr);
-       if (!pud) {
-               /*
-                * Ignore calls from kvm_set_spte_hva for unallocated
-                * address ranges.
-                */
-               return 0;
-       }
-
-       /*
-        * While dirty page logging - dissolve huge PUD, then continue
-        * on to allocate page.
-        */
-       if (logging_active)
-               stage2_dissolve_pud(mmu, addr, pud);
-
-       if (stage2_pud_none(kvm, *pud)) {
-               if (!cache)
-                       return 0; /* ignore calls from kvm_set_spte_hva */
-               pmd = kvm_mmu_memory_cache_alloc(cache);
-               stage2_pud_populate(kvm, pud, pmd);
-               get_page(virt_to_page(pud));
-       }
-
-       pmd = stage2_pmd_offset(kvm, pud, addr);
-       if (!pmd) {
-               /*
-                * Ignore calls from kvm_set_spte_hva for unallocated
-                * address ranges.
-                */
-               return 0;
-       }
-
-       /*
-        * While dirty page logging - dissolve huge PMD, then continue on to
-        * allocate page.
-        */
-       if (logging_active)
-               stage2_dissolve_pmd(mmu, addr, pmd);
-
-       /* Create stage-2 page mappings - Level 2 */
-       if (pmd_none(*pmd)) {
-               if (!cache)
-                       return 0; /* ignore calls from kvm_set_spte_hva */
-               pte = kvm_mmu_memory_cache_alloc(cache);
-               kvm_pmd_populate(pmd, pte);
-               get_page(virt_to_page(pmd));
-       }
-
-       pte = pte_offset_kernel(pmd, addr);
-
-       if (iomap && pte_present(*pte))
-               return -EFAULT;
-
-       /* Create 2nd stage page table mapping - Level 3 */
-       old_pte = *pte;
-       if (pte_present(old_pte)) {
-               /* Skip page table update if there is no change */
-               if (pte_val(old_pte) == pte_val(*new_pte))
-                       return 0;
-
-               kvm_set_pte(pte, __pte(0));
-               kvm_tlb_flush_vmid_ipa(mmu, addr, S2_PTE_LEVEL);
-       } else {
-               get_page(virt_to_page(pte));
-       }
-
-       kvm_set_pte(pte, *new_pte);
-       return 0;
-}
-
-#ifndef __HAVE_ARCH_PTEP_TEST_AND_CLEAR_YOUNG
-static int stage2_ptep_test_and_clear_young(pte_t *pte)
-{
-       if (pte_young(*pte)) {
-               *pte = pte_mkold(*pte);
-               return 1;
-       }
-       return 0;
-}
-#else
-static int stage2_ptep_test_and_clear_young(pte_t *pte)
-{
-       return __ptep_test_and_clear_young(pte);
-}
-#endif
-
-static int stage2_pmdp_test_and_clear_young(pmd_t *pmd)
-{
-       return stage2_ptep_test_and_clear_young((pte_t *)pmd);
-}
-
-static int stage2_pudp_test_and_clear_young(pud_t *pud)
-{
-       return stage2_ptep_test_and_clear_young((pte_t *)pud);
-}
-
 /**
  * kvm_phys_addr_ioremap - map a device range to guest IPA
  *
@@ -1190,102 +531,6 @@ int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
 }
 
 /**
- * stage2_wp_ptes - write protect PMD range
- * @pmd:       pointer to pmd entry
- * @addr:      range start address
- * @end:       range end address
- */
-static void stage2_wp_ptes(pmd_t *pmd, phys_addr_t addr, phys_addr_t end)
-{
-       pte_t *pte;
-
-       pte = pte_offset_kernel(pmd, addr);
-       do {
-               if (!pte_none(*pte)) {
-                       if (!kvm_s2pte_readonly(pte))
-                               kvm_set_s2pte_readonly(pte);
-               }
-       } while (pte++, addr += PAGE_SIZE, addr != end);
-}
-
-/**
- * stage2_wp_pmds - write protect PUD range
- * kvm:                kvm instance for the VM
- * @pud:       pointer to pud entry
- * @addr:      range start address
- * @end:       range end address
- */
-static void stage2_wp_pmds(struct kvm_s2_mmu *mmu, pud_t *pud,
-                          phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       pmd_t *pmd;
-       phys_addr_t next;
-
-       pmd = stage2_pmd_offset(kvm, pud, addr);
-
-       do {
-               next = stage2_pmd_addr_end(kvm, addr, end);
-               if (!pmd_none(*pmd)) {
-                       if (pmd_thp_or_huge(*pmd)) {
-                               if (!kvm_s2pmd_readonly(pmd))
-                                       kvm_set_s2pmd_readonly(pmd);
-                       } else {
-                               stage2_wp_ptes(pmd, addr, next);
-                       }
-               }
-       } while (pmd++, addr = next, addr != end);
-}
-
-/**
- * stage2_wp_puds - write protect P4D range
- * @p4d:       pointer to p4d entry
- * @addr:      range start address
- * @end:       range end address
- */
-static void  stage2_wp_puds(struct kvm_s2_mmu *mmu, p4d_t *p4d,
-                           phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       pud_t *pud;
-       phys_addr_t next;
-
-       pud = stage2_pud_offset(kvm, p4d, addr);
-       do {
-               next = stage2_pud_addr_end(kvm, addr, end);
-               if (!stage2_pud_none(kvm, *pud)) {
-                       if (stage2_pud_huge(kvm, *pud)) {
-                               if (!kvm_s2pud_readonly(pud))
-                                       kvm_set_s2pud_readonly(pud);
-                       } else {
-                               stage2_wp_pmds(mmu, pud, addr, next);
-                       }
-               }
-       } while (pud++, addr = next, addr != end);
-}
-
-/**
- * stage2_wp_p4ds - write protect PGD range
- * @pgd:       pointer to pgd entry
- * @addr:      range start address
- * @end:       range end address
- */
-static void  stage2_wp_p4ds(struct kvm_s2_mmu *mmu, pgd_t *pgd,
-                           phys_addr_t addr, phys_addr_t end)
-{
-       struct kvm *kvm = mmu->kvm;
-       p4d_t *p4d;
-       phys_addr_t next;
-
-       p4d = stage2_p4d_offset(kvm, pgd, addr);
-       do {
-               next = stage2_p4d_addr_end(kvm, addr, end);
-               if (!stage2_p4d_none(kvm, *p4d))
-                       stage2_wp_puds(mmu, p4d, addr, next);
-       } while (p4d++, addr = next, addr != end);
-}
-
-/**
  * stage2_wp_range() - write protect stage2 memory region range
  * @kvm:       The KVM pointer
  * @addr:      Start address of range