Lines Matching refs:phys_addr_t

44 static phys_addr_t hyp_idmap_vector;
70 static void kvm_tlb_flush_vmid_ipa(struct kvm *kvm, phys_addr_t ipa) in kvm_tlb_flush_vmid_ipa()
109 static void stage2_dissolve_pmd(struct kvm *kvm, phys_addr_t addr, pmd_t *pmd) in stage2_dissolve_pmd()
151 static void clear_stage2_pgd_entry(struct kvm *kvm, pgd_t *pgd, phys_addr_t addr) in clear_stage2_pgd_entry()
160 static void clear_stage2_pud_entry(struct kvm *kvm, pud_t *pud, phys_addr_t addr) in clear_stage2_pud_entry()
170 static void clear_stage2_pmd_entry(struct kvm *kvm, pmd_t *pmd, phys_addr_t addr) in clear_stage2_pmd_entry()
234 phys_addr_t addr, phys_addr_t end) in unmap_stage2_ptes()
236 phys_addr_t start_addr = addr; in unmap_stage2_ptes()
260 phys_addr_t addr, phys_addr_t end) in unmap_stage2_pmds()
262 phys_addr_t next, start_addr = addr; in unmap_stage2_pmds()
289 phys_addr_t addr, phys_addr_t end) in unmap_stage2_puds()
291 phys_addr_t next, start_addr = addr; in unmap_stage2_puds()
326 static void __unmap_stage2_range(struct kvm *kvm, phys_addr_t start, u64 size, in __unmap_stage2_range()
330 phys_addr_t addr = start, end = start + size; in __unmap_stage2_range()
331 phys_addr_t next; in __unmap_stage2_range()
357 static void unmap_stage2_range(struct kvm *kvm, phys_addr_t start, u64 size) in unmap_stage2_range()
363 phys_addr_t addr, phys_addr_t end) in stage2_flush_ptes()
375 phys_addr_t addr, phys_addr_t end) in stage2_flush_pmds()
378 phys_addr_t next; in stage2_flush_pmds()
393 phys_addr_t addr, phys_addr_t end) in stage2_flush_puds()
396 phys_addr_t next; in stage2_flush_puds()
413 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_flush_memslot()
414 phys_addr_t end = addr + PAGE_SIZE * memslot->npages; in stage2_flush_memslot()
415 phys_addr_t next; in stage2_flush_memslot()
476 static void unmap_hyp_ptes(pmd_t *pmd, phys_addr_t addr, phys_addr_t end) in unmap_hyp_ptes()
492 static void unmap_hyp_pmds(pud_t *pud, phys_addr_t addr, phys_addr_t end) in unmap_hyp_pmds()
494 phys_addr_t next; in unmap_hyp_pmds()
509 static void unmap_hyp_puds(pgd_t *pgd, phys_addr_t addr, phys_addr_t end) in unmap_hyp_puds()
511 phys_addr_t next; in unmap_hyp_puds()
532 phys_addr_t start, u64 size) in __unmap_hyp_range()
535 phys_addr_t addr = start, end = start + size; in __unmap_hyp_range()
536 phys_addr_t next; in __unmap_hyp_range()
550 static void unmap_hyp_range(pgd_t *pgdp, phys_addr_t start, u64 size) in unmap_hyp_range()
555 static void unmap_hyp_idmap_range(pgd_t *pgdp, phys_addr_t start, u64 size) in unmap_hyp_idmap_range()
726 static phys_addr_t kvm_kaddr_to_phys(void *kaddr) in kvm_kaddr_to_phys()
749 phys_addr_t phys_addr; in create_hyp_mappings()
775 static int __create_hyp_private_mapping(phys_addr_t phys_addr, size_t size, in __create_hyp_private_mapping()
832 int create_hyp_io_mappings(phys_addr_t phys_addr, size_t size, in create_hyp_io_mappings()
867 int create_hyp_exec_mappings(phys_addr_t phys_addr, size_t size, in create_hyp_exec_mappings()
919 phys_addr_t addr = memslot->base_gfn << PAGE_SHIFT; in stage2_unmap_memslot()
920 phys_addr_t size = PAGE_SIZE * memslot->npages; in stage2_unmap_memslot()
1008 phys_addr_t addr) in stage2_get_pud()
1026 phys_addr_t addr) in stage2_get_pmd()
1047 *cache, phys_addr_t addr, const pmd_t *new_pmd) in stage2_set_pmd_huge()
1093 static bool stage2_is_exec(struct kvm *kvm, phys_addr_t addr) in stage2_is_exec()
1113 phys_addr_t addr, const pte_t *new_pte, in stage2_set_pte()
1200 int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa, in kvm_phys_addr_ioremap()
1201 phys_addr_t pa, unsigned long size, bool writable) in kvm_phys_addr_ioremap()
1203 phys_addr_t addr, end; in kvm_phys_addr_ioremap()
1236 static bool transparent_hugepage_adjust(kvm_pfn_t *pfnp, phys_addr_t *ipap) in transparent_hugepage_adjust()
1300 static void stage2_wp_ptes(pmd_t *pmd, phys_addr_t addr, phys_addr_t end) in stage2_wp_ptes()
1319 static void stage2_wp_pmds(pud_t *pud, phys_addr_t addr, phys_addr_t end) in stage2_wp_pmds()
1322 phys_addr_t next; in stage2_wp_pmds()
1347 static void stage2_wp_puds(pgd_t *pgd, phys_addr_t addr, phys_addr_t end) in stage2_wp_puds()
1350 phys_addr_t next; in stage2_wp_puds()
1369 static void stage2_wp_range(struct kvm *kvm, phys_addr_t addr, phys_addr_t end) in stage2_wp_range()
1372 phys_addr_t next; in stage2_wp_range()
1411 phys_addr_t start = memslot->base_gfn << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1412 phys_addr_t end = (memslot->base_gfn + memslot->npages) << PAGE_SHIFT; in kvm_mmu_wp_memory_region()
1435 phys_addr_t base_gfn = slot->base_gfn + gfn_offset; in kvm_mmu_write_protect_pt_masked()
1436 phys_addr_t start = (base_gfn + __ffs(mask)) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1437 phys_addr_t end = (base_gfn + __fls(mask) + 1) << PAGE_SHIFT; in kvm_mmu_write_protect_pt_masked()
1485 static int user_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa, in user_mem_abort()
1650 static void handle_access_fault(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa) in handle_access_fault()
1700 phys_addr_t fault_ipa; in kvm_handle_guest_abort()
1948 phys_addr_t kvm_mmu_get_httbr(void) in kvm_mmu_get_httbr()
1956 phys_addr_t kvm_get_idmap_vector(void) in kvm_get_idmap_vector()
2125 phys_addr_t pa; in kvm_arch_prepare_memory_region()
2127 pa = (phys_addr_t)vma->vm_pgoff << PAGE_SHIFT; in kvm_arch_prepare_memory_region()
2183 phys_addr_t size = slot->npages << PAGE_SHIFT; in kvm_arch_flush_shadow_memslot()