/linux-4.19.296/include/trace/events/ |
D | thp.h | 13 TP_PROTO(unsigned long addr, unsigned long pte), 14 TP_ARGS(addr, pte), 17 __field(unsigned long, pte) 22 __entry->pte = pte; 26 __entry->addr, __entry->pte) 49 TP_PROTO(unsigned long addr, unsigned long pte, unsigned long clr, unsigned long set), 50 TP_ARGS(addr, pte, clr, set), 53 __field(unsigned long, pte) 60 __entry->pte = pte; 66 …r 0x%lx and pte = 0x%lx clr = 0x%lx, set = 0x%lx", __entry->addr, __entry->pte, __entry->clr, __en… [all …]
|
D | xen.h | 142 __entry->pteval = pteval.pte), 169 __entry->pteval = pteval.pte), 291 __entry->pteval = pteval.pte),
|
/linux-4.19.296/include/asm-generic/ |
D | hugetlb.h | 10 static inline unsigned long huge_pte_write(pte_t pte) in huge_pte_write() argument 12 return pte_write(pte); in huge_pte_write() 15 static inline unsigned long huge_pte_dirty(pte_t pte) in huge_pte_dirty() argument 17 return pte_dirty(pte); in huge_pte_dirty() 20 static inline pte_t huge_pte_mkwrite(pte_t pte) in huge_pte_mkwrite() argument 22 return pte_mkwrite(pte); in huge_pte_mkwrite() 25 static inline pte_t huge_pte_mkdirty(pte_t pte) in huge_pte_mkdirty() argument 27 return pte_mkdirty(pte); in huge_pte_mkdirty() 30 static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot) in huge_pte_modify() argument 32 return pte_modify(pte, newprot); in huge_pte_modify()
|
D | pgtable.h | 66 pte_t pte = *ptep; in ptep_test_and_clear_young() local 68 if (!pte_young(pte)) in ptep_test_and_clear_young() 71 set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); in ptep_test_and_clear_young() 129 pte_t pte = *ptep; in ptep_get_and_clear() local 131 return pte; in ptep_get_and_clear() 184 pte_t pte; in ptep_get_and_clear_full() local 185 pte = ptep_get_and_clear(mm, address, ptep); in ptep_get_and_clear_full() 186 return pte; in ptep_get_and_clear_full() 346 static inline int pte_unused(pte_t pte) in pte_unused() argument 353 #define pte_access_permitted(pte, write) \ argument [all …]
|
D | page.h | 38 unsigned long pte; member 51 #define pte_val(x) ((x).pte)
|
D | pgtable-nopmd.h | 36 #define pud_populate(mm, pmd, pte) do { } while (0) argument
|
/linux-4.19.296/virt/kvm/arm/ |
D | mmu.c | 80 static void kvm_flush_dcache_pte(pte_t pte) in kvm_flush_dcache_pte() argument 82 __kvm_flush_dcache_pte(pte); in kvm_flush_dcache_pte() 237 pte_t *pte, *start_pte; in unmap_stage2_ptes() local 239 start_pte = pte = pte_offset_kernel(pmd, addr); in unmap_stage2_ptes() 241 if (!pte_none(*pte)) { in unmap_stage2_ptes() 242 pte_t old_pte = *pte; in unmap_stage2_ptes() 244 kvm_set_pte(pte, __pte(0)); in unmap_stage2_ptes() 251 put_page(virt_to_page(pte)); in unmap_stage2_ptes() 253 } while (pte++, addr += PAGE_SIZE, addr != end); in unmap_stage2_ptes() 365 pte_t *pte; in stage2_flush_ptes() local [all …]
|
/linux-4.19.296/fs/proc/ |
D | task_mmu.c | 499 static void smaps_pte_entry(pte_t *pte, unsigned long addr, in smaps_pte_entry() argument 507 if (pte_present(*pte)) { in smaps_pte_entry() 508 page = vm_normal_page(vma, addr, *pte); in smaps_pte_entry() 509 } else if (is_swap_pte(*pte)) { in smaps_pte_entry() 510 swp_entry_t swpent = pte_to_swp_entry(*pte); in smaps_pte_entry() 530 && pte_none(*pte))) { in smaps_pte_entry() 547 smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked); in smaps_pte_entry() 584 pte_t *pte; in smaps_pte_range() local 602 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range() 603 for (; addr != end; pte++, addr += PAGE_SIZE) in smaps_pte_range() [all …]
|
/linux-4.19.296/include/linux/ |
D | swapops.h | 57 static inline int is_swap_pte(pte_t pte) in is_swap_pte() argument 59 return !pte_none(pte) && !pte_present(pte); in is_swap_pte() 67 static inline swp_entry_t pte_to_swp_entry(pte_t pte) in pte_to_swp_entry() argument 71 if (pte_swp_soft_dirty(pte)) in pte_to_swp_entry() 72 pte = pte_swp_clear_soft_dirty(pte); in pte_to_swp_entry() 73 arch_entry = __pte_to_swp_entry(pte); in pte_to_swp_entry() 229 struct mm_struct *mm, pte_t *pte); 254 struct mm_struct *mm, pte_t *pte) { } in migration_entry_wait_huge() argument
|
D | hugetlb.h | 161 bool is_hugetlb_entry_migration(pte_t pte); 501 struct mm_struct *mm, pte_t *pte) in huge_pte_lockptr() argument 504 return pmd_lockptr(mm, (pmd_t *) pte); in huge_pte_lockptr() 537 pte_t *ptep, pte_t pte, unsigned long sz) in set_huge_swap_pte_at() argument 539 set_huge_pte_at(mm, addr, ptep, pte); in set_huge_swap_pte_at() 600 struct mm_struct *mm, pte_t *pte) in huge_pte_lockptr() argument 618 pte_t *ptep, pte_t pte, unsigned long sz) in set_huge_swap_pte_at() argument 624 struct mm_struct *mm, pte_t *pte) in huge_pte_lock() argument 628 ptl = huge_pte_lockptr(h, mm, pte); in huge_pte_lock() 634 static inline bool hugetlb_pmd_shared(pte_t *pte) in hugetlb_pmd_shared() argument [all …]
|
D | mmu_notifier.h | 109 pte_t pte); 233 unsigned long address, pte_t pte); 277 unsigned long address, pte_t pte) in mmu_notifier_change_pte() argument 280 __mmu_notifier_change_pte(mm, address, pte); in mmu_notifier_change_pte() 466 unsigned long address, pte_t pte) in mmu_notifier_change_pte() argument
|
D | rmap.h | 214 pte_t *pte; member 222 if (pvmw->pte && !PageHuge(pvmw->page)) in page_vma_mapped_walk_done() 223 pte_unmap(pvmw->pte); in page_vma_mapped_walk_done()
|
D | mm.h | 364 pte_t *pte; /* Pointer to pte entry matching member 731 static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) in maybe_mkwrite() argument 734 pte = pte_mkwrite(pte); in maybe_mkwrite() 735 return pte; in maybe_mkwrite() 1345 pte_t pte, bool with_public_device); 1346 #define vm_normal_page(vma, addr, pte) _vm_normal_page(vma, addr, pte, false) argument 1387 int (*pte_entry)(pte_t *pte, unsigned long addr, 1391 int (*hugetlb_entry)(pte_t *pte, unsigned long hmask, 1708 static inline int pte_devmap(pte_t pte) in pte_devmap() argument 1957 #define pte_unmap_unlock(pte, ptl) do { \ argument [all …]
|
D | pfn_t.h | 117 pte_t pte_mkdevmap(pte_t pte);
|
/linux-4.19.296/lib/ |
D | ioremap.c | 64 pte_t *pte; in ioremap_pte_range() local 68 pte = pte_alloc_kernel(pmd, addr); in ioremap_pte_range() 69 if (!pte) in ioremap_pte_range() 72 BUG_ON(!pte_none(*pte)); in ioremap_pte_range() 73 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range() 75 } while (pte++, addr += PAGE_SIZE, addr != end); in ioremap_pte_range()
|
/linux-4.19.296/block/partitions/ |
D | efi.c | 291 gpt_entry *pte; in alloc_read_gpt_entries() local 300 pte = kmalloc(count, GFP_KERNEL); in alloc_read_gpt_entries() 301 if (!pte) in alloc_read_gpt_entries() 305 (u8 *) pte, count) < count) { in alloc_read_gpt_entries() 306 kfree(pte); in alloc_read_gpt_entries() 307 pte=NULL; in alloc_read_gpt_entries() 310 return pte; in alloc_read_gpt_entries() 477 is_pte_valid(const gpt_entry *pte, const u64 lastlba) in is_pte_valid() argument 479 if ((!efi_guidcmp(pte->partition_type_guid, NULL_GUID)) || in is_pte_valid() 480 le64_to_cpu(pte->starting_lba) > lastlba || in is_pte_valid() [all …]
|
/linux-4.19.296/drivers/misc/sgi-gru/ |
D | grufault.c | 225 pte_t pte; in atomic_pte_lookup() local 244 pte = *(pte_t *) pmdp; in atomic_pte_lookup() 247 pte = *pte_offset_kernel(pmdp, vaddr); in atomic_pte_lookup() 249 if (unlikely(!pte_present(pte) || in atomic_pte_lookup() 250 (write && (!pte_write(pte) || !pte_dirty(pte))))) in atomic_pte_lookup() 253 *paddr = pte_pfn(pte) << PAGE_SHIFT; in atomic_pte_lookup()
|
/linux-4.19.296/drivers/char/agp/ |
D | amd64-agp.c | 50 u32 pte; in amd64_insert_memory() local 86 pte = (tmp & 0x000000ff00000000ULL) >> 28; in amd64_insert_memory() 87 pte |=(tmp & 0x00000000fffff000ULL); in amd64_insert_memory() 88 pte |= GPTE_VALID | GPTE_COHERENT; in amd64_insert_memory() 90 writel(pte, agp_bridge->gatt_table+j); in amd64_insert_memory()
|
/linux-4.19.296/drivers/xen/ |
D | xlate_mmu.c | 101 pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot)); in remap_pte_fn() local 139 set_pte_at(info->vma->vm_mm, addr, ptep, pte); in remap_pte_fn()
|
D | privcmd.c | 738 pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot)); in remap_pfn_fn() local 740 set_pte_at(r->mm, addr, ptep, pte); in remap_pfn_fn() 982 static int is_mapped_fn(pte_t *pte, struct page *pmd_page, in is_mapped_fn() argument 985 return pte_none(*pte) ? 0 : -EBUSY; in is_mapped_fn()
|
D | gntdev.c | 301 static int find_grant_ptes(pte_t *pte, pgtable_t token, in find_grant_ptes() argument 310 pte_maddr = arbitrary_virt_to_machine(pte).maddr; in find_grant_ptes() 329 static int set_grant_ptes_as_special(pte_t *pte, pgtable_t token, in set_grant_ptes_as_special() argument 332 set_pte_at(current->mm, addr, pte, pte_mkspecial(*pte)); in set_grant_ptes_as_special()
|
/linux-4.19.296/fs/ |
D | userfaultfd.c | 236 pte_t *ptep, pte; in userfaultfd_huge_must_wait() local 247 pte = huge_ptep_get(ptep); in userfaultfd_huge_must_wait() 253 if (huge_pte_none(pte)) in userfaultfd_huge_must_wait() 255 if (!huge_pte_write(pte) && (reason & VM_UFFD_WP)) in userfaultfd_huge_must_wait() 288 pte_t *pte; in userfaultfd_must_wait() local 326 pte = pte_offset_map(pmd, address); in userfaultfd_must_wait() 331 if (pte_none(*pte)) in userfaultfd_must_wait() 333 pte_unmap(pte); in userfaultfd_must_wait()
|
D | dax.c | 871 pte_t pte, *ptep = NULL; in dax_mapping_entry_mkclean() local 926 pte = ptep_clear_flush(vma, address, ptep); in dax_mapping_entry_mkclean() 927 pte = pte_wrprotect(pte); in dax_mapping_entry_mkclean() 928 pte = pte_mkclean(pte); in dax_mapping_entry_mkclean() 929 set_pte_at(vma->vm_mm, address, ptep, pte); in dax_mapping_entry_mkclean()
|
/linux-4.19.296/virt/kvm/ |
D | kvm_main.c | 386 pte_t pte) in kvm_mmu_notifier_change_pte() argument 394 kvm_set_spte_hva(kvm, address, pte); in kvm_mmu_notifier_change_pte()
|