Home
last modified time | relevance | path

Searched refs:pte (Results 1 – 24 of 24) sorted by relevance

/linux-4.19.296/include/trace/events/
Dthp.h13 TP_PROTO(unsigned long addr, unsigned long pte),
14 TP_ARGS(addr, pte),
17 __field(unsigned long, pte)
22 __entry->pte = pte;
26 __entry->addr, __entry->pte)
49 TP_PROTO(unsigned long addr, unsigned long pte, unsigned long clr, unsigned long set),
50 TP_ARGS(addr, pte, clr, set),
53 __field(unsigned long, pte)
60 __entry->pte = pte;
66 …r 0x%lx and pte = 0x%lx clr = 0x%lx, set = 0x%lx", __entry->addr, __entry->pte, __entry->clr, __en…
[all …]
Dxen.h142 __entry->pteval = pteval.pte),
169 __entry->pteval = pteval.pte),
291 __entry->pteval = pteval.pte),
/linux-4.19.296/include/asm-generic/
Dhugetlb.h10 static inline unsigned long huge_pte_write(pte_t pte) in huge_pte_write() argument
12 return pte_write(pte); in huge_pte_write()
15 static inline unsigned long huge_pte_dirty(pte_t pte) in huge_pte_dirty() argument
17 return pte_dirty(pte); in huge_pte_dirty()
20 static inline pte_t huge_pte_mkwrite(pte_t pte) in huge_pte_mkwrite() argument
22 return pte_mkwrite(pte); in huge_pte_mkwrite()
25 static inline pte_t huge_pte_mkdirty(pte_t pte) in huge_pte_mkdirty() argument
27 return pte_mkdirty(pte); in huge_pte_mkdirty()
30 static inline pte_t huge_pte_modify(pte_t pte, pgprot_t newprot) in huge_pte_modify() argument
32 return pte_modify(pte, newprot); in huge_pte_modify()
Dpgtable.h66 pte_t pte = *ptep; in ptep_test_and_clear_young() local
68 if (!pte_young(pte)) in ptep_test_and_clear_young()
71 set_pte_at(vma->vm_mm, address, ptep, pte_mkold(pte)); in ptep_test_and_clear_young()
129 pte_t pte = *ptep; in ptep_get_and_clear() local
131 return pte; in ptep_get_and_clear()
184 pte_t pte; in ptep_get_and_clear_full() local
185 pte = ptep_get_and_clear(mm, address, ptep); in ptep_get_and_clear_full()
186 return pte; in ptep_get_and_clear_full()
346 static inline int pte_unused(pte_t pte) in pte_unused() argument
353 #define pte_access_permitted(pte, write) \ argument
[all …]
Dpage.h38 unsigned long pte; member
51 #define pte_val(x) ((x).pte)
Dpgtable-nopmd.h36 #define pud_populate(mm, pmd, pte) do { } while (0) argument
/linux-4.19.296/virt/kvm/arm/
Dmmu.c80 static void kvm_flush_dcache_pte(pte_t pte) in kvm_flush_dcache_pte() argument
82 __kvm_flush_dcache_pte(pte); in kvm_flush_dcache_pte()
237 pte_t *pte, *start_pte; in unmap_stage2_ptes() local
239 start_pte = pte = pte_offset_kernel(pmd, addr); in unmap_stage2_ptes()
241 if (!pte_none(*pte)) { in unmap_stage2_ptes()
242 pte_t old_pte = *pte; in unmap_stage2_ptes()
244 kvm_set_pte(pte, __pte(0)); in unmap_stage2_ptes()
251 put_page(virt_to_page(pte)); in unmap_stage2_ptes()
253 } while (pte++, addr += PAGE_SIZE, addr != end); in unmap_stage2_ptes()
365 pte_t *pte; in stage2_flush_ptes() local
[all …]
/linux-4.19.296/fs/proc/
Dtask_mmu.c499 static void smaps_pte_entry(pte_t *pte, unsigned long addr, in smaps_pte_entry() argument
507 if (pte_present(*pte)) { in smaps_pte_entry()
508 page = vm_normal_page(vma, addr, *pte); in smaps_pte_entry()
509 } else if (is_swap_pte(*pte)) { in smaps_pte_entry()
510 swp_entry_t swpent = pte_to_swp_entry(*pte); in smaps_pte_entry()
530 && pte_none(*pte))) { in smaps_pte_entry()
547 smaps_account(mss, page, false, pte_young(*pte), pte_dirty(*pte), locked); in smaps_pte_entry()
584 pte_t *pte; in smaps_pte_range() local
602 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range()
603 for (; addr != end; pte++, addr += PAGE_SIZE) in smaps_pte_range()
[all …]
/linux-4.19.296/include/linux/
Dswapops.h57 static inline int is_swap_pte(pte_t pte) in is_swap_pte() argument
59 return !pte_none(pte) && !pte_present(pte); in is_swap_pte()
67 static inline swp_entry_t pte_to_swp_entry(pte_t pte) in pte_to_swp_entry() argument
71 if (pte_swp_soft_dirty(pte)) in pte_to_swp_entry()
72 pte = pte_swp_clear_soft_dirty(pte); in pte_to_swp_entry()
73 arch_entry = __pte_to_swp_entry(pte); in pte_to_swp_entry()
229 struct mm_struct *mm, pte_t *pte);
254 struct mm_struct *mm, pte_t *pte) { } in migration_entry_wait_huge() argument
Dhugetlb.h161 bool is_hugetlb_entry_migration(pte_t pte);
501 struct mm_struct *mm, pte_t *pte) in huge_pte_lockptr() argument
504 return pmd_lockptr(mm, (pmd_t *) pte); in huge_pte_lockptr()
537 pte_t *ptep, pte_t pte, unsigned long sz) in set_huge_swap_pte_at() argument
539 set_huge_pte_at(mm, addr, ptep, pte); in set_huge_swap_pte_at()
600 struct mm_struct *mm, pte_t *pte) in huge_pte_lockptr() argument
618 pte_t *ptep, pte_t pte, unsigned long sz) in set_huge_swap_pte_at() argument
624 struct mm_struct *mm, pte_t *pte) in huge_pte_lock() argument
628 ptl = huge_pte_lockptr(h, mm, pte); in huge_pte_lock()
634 static inline bool hugetlb_pmd_shared(pte_t *pte) in hugetlb_pmd_shared() argument
[all …]
Dmmu_notifier.h109 pte_t pte);
233 unsigned long address, pte_t pte);
277 unsigned long address, pte_t pte) in mmu_notifier_change_pte() argument
280 __mmu_notifier_change_pte(mm, address, pte); in mmu_notifier_change_pte()
466 unsigned long address, pte_t pte) in mmu_notifier_change_pte() argument
Drmap.h214 pte_t *pte; member
222 if (pvmw->pte && !PageHuge(pvmw->page)) in page_vma_mapped_walk_done()
223 pte_unmap(pvmw->pte); in page_vma_mapped_walk_done()
Dmm.h364 pte_t *pte; /* Pointer to pte entry matching member
731 static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) in maybe_mkwrite() argument
734 pte = pte_mkwrite(pte); in maybe_mkwrite()
735 return pte; in maybe_mkwrite()
1345 pte_t pte, bool with_public_device);
1346 #define vm_normal_page(vma, addr, pte) _vm_normal_page(vma, addr, pte, false) argument
1387 int (*pte_entry)(pte_t *pte, unsigned long addr,
1391 int (*hugetlb_entry)(pte_t *pte, unsigned long hmask,
1708 static inline int pte_devmap(pte_t pte) in pte_devmap() argument
1957 #define pte_unmap_unlock(pte, ptl) do { \ argument
[all …]
Dpfn_t.h117 pte_t pte_mkdevmap(pte_t pte);
/linux-4.19.296/lib/
Dioremap.c64 pte_t *pte; in ioremap_pte_range() local
68 pte = pte_alloc_kernel(pmd, addr); in ioremap_pte_range()
69 if (!pte) in ioremap_pte_range()
72 BUG_ON(!pte_none(*pte)); in ioremap_pte_range()
73 set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); in ioremap_pte_range()
75 } while (pte++, addr += PAGE_SIZE, addr != end); in ioremap_pte_range()
/linux-4.19.296/block/partitions/
Defi.c291 gpt_entry *pte; in alloc_read_gpt_entries() local
300 pte = kmalloc(count, GFP_KERNEL); in alloc_read_gpt_entries()
301 if (!pte) in alloc_read_gpt_entries()
305 (u8 *) pte, count) < count) { in alloc_read_gpt_entries()
306 kfree(pte); in alloc_read_gpt_entries()
307 pte=NULL; in alloc_read_gpt_entries()
310 return pte; in alloc_read_gpt_entries()
477 is_pte_valid(const gpt_entry *pte, const u64 lastlba) in is_pte_valid() argument
479 if ((!efi_guidcmp(pte->partition_type_guid, NULL_GUID)) || in is_pte_valid()
480 le64_to_cpu(pte->starting_lba) > lastlba || in is_pte_valid()
[all …]
/linux-4.19.296/drivers/misc/sgi-gru/
Dgrufault.c225 pte_t pte; in atomic_pte_lookup() local
244 pte = *(pte_t *) pmdp; in atomic_pte_lookup()
247 pte = *pte_offset_kernel(pmdp, vaddr); in atomic_pte_lookup()
249 if (unlikely(!pte_present(pte) || in atomic_pte_lookup()
250 (write && (!pte_write(pte) || !pte_dirty(pte))))) in atomic_pte_lookup()
253 *paddr = pte_pfn(pte) << PAGE_SHIFT; in atomic_pte_lookup()
/linux-4.19.296/drivers/char/agp/
Damd64-agp.c50 u32 pte; in amd64_insert_memory() local
86 pte = (tmp & 0x000000ff00000000ULL) >> 28; in amd64_insert_memory()
87 pte |=(tmp & 0x00000000fffff000ULL); in amd64_insert_memory()
88 pte |= GPTE_VALID | GPTE_COHERENT; in amd64_insert_memory()
90 writel(pte, agp_bridge->gatt_table+j); in amd64_insert_memory()
/linux-4.19.296/drivers/xen/
Dxlate_mmu.c101 pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), info->prot)); in remap_pte_fn() local
139 set_pte_at(info->vma->vm_mm, addr, ptep, pte); in remap_pte_fn()
Dprivcmd.c738 pte_t pte = pte_mkspecial(pfn_pte(page_to_pfn(page), r->prot)); in remap_pfn_fn() local
740 set_pte_at(r->mm, addr, ptep, pte); in remap_pfn_fn()
982 static int is_mapped_fn(pte_t *pte, struct page *pmd_page, in is_mapped_fn() argument
985 return pte_none(*pte) ? 0 : -EBUSY; in is_mapped_fn()
Dgntdev.c301 static int find_grant_ptes(pte_t *pte, pgtable_t token, in find_grant_ptes() argument
310 pte_maddr = arbitrary_virt_to_machine(pte).maddr; in find_grant_ptes()
329 static int set_grant_ptes_as_special(pte_t *pte, pgtable_t token, in set_grant_ptes_as_special() argument
332 set_pte_at(current->mm, addr, pte, pte_mkspecial(*pte)); in set_grant_ptes_as_special()
/linux-4.19.296/fs/
Duserfaultfd.c236 pte_t *ptep, pte; in userfaultfd_huge_must_wait() local
247 pte = huge_ptep_get(ptep); in userfaultfd_huge_must_wait()
253 if (huge_pte_none(pte)) in userfaultfd_huge_must_wait()
255 if (!huge_pte_write(pte) && (reason & VM_UFFD_WP)) in userfaultfd_huge_must_wait()
288 pte_t *pte; in userfaultfd_must_wait() local
326 pte = pte_offset_map(pmd, address); in userfaultfd_must_wait()
331 if (pte_none(*pte)) in userfaultfd_must_wait()
333 pte_unmap(pte); in userfaultfd_must_wait()
Ddax.c871 pte_t pte, *ptep = NULL; in dax_mapping_entry_mkclean() local
926 pte = ptep_clear_flush(vma, address, ptep); in dax_mapping_entry_mkclean()
927 pte = pte_wrprotect(pte); in dax_mapping_entry_mkclean()
928 pte = pte_mkclean(pte); in dax_mapping_entry_mkclean()
929 set_pte_at(vma->vm_mm, address, ptep, pte); in dax_mapping_entry_mkclean()
/linux-4.19.296/virt/kvm/
Dkvm_main.c386 pte_t pte) in kvm_mmu_notifier_change_pte() argument
394 kvm_set_spte_hva(kvm, address, pte); in kvm_mmu_notifier_change_pte()