/linux-4.19.296/include/linux/ |
D | page-flags.h | 138 struct page; /* forward declaration */ 140 static inline struct page *compound_head(struct page *page) in compound_head() argument 142 unsigned long head = READ_ONCE(page->compound_head); in compound_head() 145 return (struct page *) (head - 1); in compound_head() 146 return page; in compound_head() 149 static __always_inline int PageTail(struct page *page) in PageTail() argument 151 return READ_ONCE(page->compound_head) & 1; in PageTail() 154 static __always_inline int PageCompound(struct page *page) in PageCompound() argument 156 return test_bit(PG_head, &page->flags) || PageTail(page); in PageCompound() 160 static inline int PagePoisoned(const struct page *page) in PagePoisoned() argument [all …]
|
D | page_ref.h | 29 extern void __page_ref_set(struct page *page, int v); 30 extern void __page_ref_mod(struct page *page, int v); 31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret); 32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret); 33 extern void __page_ref_mod_unless(struct page *page, int v, int u); 34 extern void __page_ref_freeze(struct page *page, int v, int ret); 35 extern void __page_ref_unfreeze(struct page *page, int v); 41 static inline void __page_ref_set(struct page *page, int v) in __page_ref_set() argument 44 static inline void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument 47 static inline void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument [all …]
|
D | balloon_compaction.h | 66 int (*migratepage)(struct balloon_dev_info *, struct page *newpage, 67 struct page *page, enum migrate_mode mode); 71 extern struct page *balloon_page_alloc(void); 73 struct page *page); 74 extern struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info); 87 extern bool balloon_page_isolate(struct page *page, 89 extern void balloon_page_putback(struct page *page); 91 struct page *newpage, 92 struct page *page, enum migrate_mode mode); 104 struct page *page) in balloon_page_insert() argument [all …]
|
D | page_idle.h | 12 static inline bool page_is_young(struct page *page) in page_is_young() argument 14 return PageYoung(page); in page_is_young() 17 static inline void set_page_young(struct page *page) in set_page_young() argument 19 SetPageYoung(page); in set_page_young() 22 static inline bool test_and_clear_page_young(struct page *page) in test_and_clear_page_young() argument 24 return TestClearPageYoung(page); in test_and_clear_page_young() 27 static inline bool page_is_idle(struct page *page) in page_is_idle() argument 29 return PageIdle(page); in page_is_idle() 32 static inline void set_page_idle(struct page *page) in set_page_idle() argument 34 SetPageIdle(page); in set_page_idle() [all …]
|
D | pagemap.h | 121 void release_pages(struct page **pages, int nr); 167 static inline int page_cache_get_speculative(struct page *page) in page_cache_get_speculative() argument 182 VM_BUG_ON_PAGE(page_count(page) == 0, page); in page_cache_get_speculative() 183 page_ref_inc(page); in page_cache_get_speculative() 186 if (unlikely(!get_page_unless_zero(page))) { in page_cache_get_speculative() 195 VM_BUG_ON_PAGE(PageTail(page), page); in page_cache_get_speculative() 203 static inline int page_cache_add_speculative(struct page *page, int count) in page_cache_add_speculative() argument 211 VM_BUG_ON_PAGE(page_count(page) == 0, page); in page_cache_add_speculative() 212 page_ref_add(page, count); in page_cache_add_speculative() 215 if (unlikely(!page_ref_add_unless(page, count, 0))) in page_cache_add_speculative() [all …]
|
D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned int order); 12 extern void __set_page_owner(struct page *page, 14 extern void __split_page_owner(struct page *page, unsigned int order); 15 extern void __copy_page_owner(struct page *oldpage, struct page *newpage); 16 extern void __set_page_owner_migrate_reason(struct page *page, int reason); 17 extern void __dump_page_owner(struct page *page); 21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 24 __reset_page_owner(page, order); in reset_page_owner() 27 static inline void set_page_owner(struct page *page, in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() [all …]
|
D | migrate.h | 10 typedef struct page *new_page_t(struct page *page, unsigned long private); 11 typedef void free_page_t(struct page *page, unsigned long private); 34 static inline struct page *new_page_nodemask(struct page *page, in new_page_nodemask() argument 39 struct page *new_page = NULL; in new_page_nodemask() 41 if (PageHuge(page)) in new_page_nodemask() 42 return alloc_huge_page_nodemask(page_hstate(compound_head(page)), in new_page_nodemask() 45 if (PageTransHuge(page)) { in new_page_nodemask() 50 if (PageHighMem(page) || (zone_idx(page_zone(page)) == ZONE_MOVABLE)) in new_page_nodemask() 66 struct page *newpage, struct page *page, 70 extern int isolate_movable_page(struct page *page, isolate_mode_t mode); [all …]
|
D | mm.h | 105 #define mm_zero_struct_page(pp) ((void)memset((pp), 0, sizeof(struct page))) 141 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) argument 356 struct page *cow_page; /* Page handler may use for COW fault */ 358 struct page *page; /* ->fault handlers should return a member 451 struct page *(*find_special_page)(struct vm_area_struct *vma, 476 #define page_private(page) ((page)->private) argument 477 #define set_page_private(page, v) ((page)->private = (v)) argument 517 static inline int put_page_testzero(struct page *page) in put_page_testzero() argument 519 VM_BUG_ON_PAGE(page_ref_count(page) == 0, page); in put_page_testzero() 520 return page_ref_dec_and_test(page); in put_page_testzero() [all …]
|
D | swap.h | 307 void *workingset_eviction(struct address_space *mapping, struct page *page); 309 void workingset_activation(struct page *page); 334 extern void lru_cache_add(struct page *); 335 extern void lru_cache_add_anon(struct page *page); 336 extern void lru_cache_add_file(struct page *page); 337 extern void lru_add_page_tail(struct page *page, struct page *page_tail, 339 extern void activate_page(struct page *); 340 extern void mark_page_accessed(struct page *); 344 extern void rotate_reclaimable_page(struct page *page); 345 extern void deactivate_file_page(struct page *page); [all …]
|
D | highmem.h | 15 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma… in flush_anon_page() argument 21 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument 43 struct page *kmap_to_page(void *addr); 49 static inline struct page *kmap_to_page(void *addr) in kmap_to_page() 57 static inline void *kmap(struct page *page) in kmap() argument 60 return page_address(page); in kmap() 63 static inline void kunmap(struct page *page) in kunmap() argument 67 static inline void *kmap_atomic(struct page *page) in kmap_atomic() argument 71 return page_address(page); in kmap_atomic() 73 #define kmap_atomic_prot(page, prot) kmap_atomic(page) argument [all …]
|
D | hugetlb_cgroup.h | 29 static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) in hugetlb_cgroup_from_page() argument 31 VM_BUG_ON_PAGE(!PageHuge(page), page); in hugetlb_cgroup_from_page() 33 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in hugetlb_cgroup_from_page() 35 return (struct hugetlb_cgroup *)page[2].private; in hugetlb_cgroup_from_page() 39 int set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) in set_hugetlb_cgroup() argument 41 VM_BUG_ON_PAGE(!PageHuge(page), page); in set_hugetlb_cgroup() 43 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in set_hugetlb_cgroup() 45 page[2].private = (unsigned long)h_cg; in set_hugetlb_cgroup() 58 struct page *page); 60 struct page *page); [all …]
|
D | mm_inline.h | 21 static inline int page_is_file_cache(struct page *page) in page_is_file_cache() argument 23 return !PageSwapBacked(page); in page_is_file_cache() 47 static __always_inline void add_page_to_lru_list(struct page *page, in add_page_to_lru_list() argument 50 update_lru_size(lruvec, lru, page_zonenum(page), hpage_nr_pages(page)); in add_page_to_lru_list() 51 list_add(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list() 54 static __always_inline void add_page_to_lru_list_tail(struct page *page, in add_page_to_lru_list_tail() argument 57 update_lru_size(lruvec, lru, page_zonenum(page), hpage_nr_pages(page)); in add_page_to_lru_list_tail() 58 list_add_tail(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list_tail() 61 static __always_inline void del_page_from_lru_list(struct page *page, in del_page_from_lru_list() argument 64 list_del(&page->lru); in del_page_from_lru_list() [all …]
|
D | rmap.h | 167 struct anon_vma *page_get_anon_vma(struct page *page); 176 void page_move_anon_rmap(struct page *, struct vm_area_struct *); 177 void page_add_anon_rmap(struct page *, struct vm_area_struct *, 179 void do_page_add_anon_rmap(struct page *, struct vm_area_struct *, 181 void page_add_new_anon_rmap(struct page *, struct vm_area_struct *, 183 void page_add_file_rmap(struct page *, bool); 184 void page_remove_rmap(struct page *, bool); 186 void hugepage_add_anon_rmap(struct page *, struct vm_area_struct *, 188 void hugepage_add_new_anon_rmap(struct page *, struct vm_area_struct *, 191 static inline void page_dup_rmap(struct page *page, bool compound) in page_dup_rmap() argument [all …]
|
D | pageblock-flags.h | 66 struct page; 68 unsigned long get_pfnblock_flags_mask(struct page *page, 73 void set_pfnblock_flags_mask(struct page *page, 80 #define get_pageblock_flags_group(page, start_bitidx, end_bitidx) \ argument 81 get_pfnblock_flags_mask(page, page_to_pfn(page), \ 84 #define set_pageblock_flags_group(page, flags, start_bitidx, end_bitidx) \ argument 85 set_pfnblock_flags_mask(page, flags, page_to_pfn(page), \ 90 #define get_pageblock_skip(page) \ argument 91 get_pageblock_flags_group(page, PB_migrate_skip, \ 93 #define clear_pageblock_skip(page) \ argument [all …]
|
/linux-4.19.296/fs/9p/ |
D | cache.h | 45 extern int __v9fs_fscache_release_page(struct page *page, gfp_t gfp); 46 extern void __v9fs_fscache_invalidate_page(struct page *page); 48 struct page *page); 53 extern void __v9fs_readpage_to_fscache(struct inode *inode, struct page *page); 55 struct page *page); 57 static inline int v9fs_fscache_release_page(struct page *page, in v9fs_fscache_release_page() argument 60 return __v9fs_fscache_release_page(page, gfp); in v9fs_fscache_release_page() 63 static inline void v9fs_fscache_invalidate_page(struct page *page) in v9fs_fscache_invalidate_page() argument 65 __v9fs_fscache_invalidate_page(page); in v9fs_fscache_invalidate_page() 69 struct page *page) in v9fs_readpage_from_fscache() argument [all …]
|
D | vfs_addr.c | 53 static int v9fs_fid_readpage(void *data, struct page *page) in v9fs_fid_readpage() argument 56 struct inode *inode = page->mapping->host; in v9fs_fid_readpage() 57 struct bio_vec bvec = {.bv_page = page, .bv_len = PAGE_SIZE}; in v9fs_fid_readpage() 63 BUG_ON(!PageLocked(page)); in v9fs_fid_readpage() 65 retval = v9fs_readpage_from_fscache(inode, page); in v9fs_fid_readpage() 71 retval = p9_client_read(fid, page_offset(page), &to, &err); in v9fs_fid_readpage() 73 v9fs_uncache_page(inode, page); in v9fs_fid_readpage() 78 zero_user(page, retval, PAGE_SIZE - retval); in v9fs_fid_readpage() 79 flush_dcache_page(page); in v9fs_fid_readpage() 80 SetPageUptodate(page); in v9fs_fid_readpage() [all …]
|
/linux-4.19.296/fs/jfs/ |
D | jfs_metapage.c | 61 unlock_page(mp->page); in __lock_metapage() 63 lock_page(mp->page); in __lock_metapage() 92 #define mp_anchor(page) ((struct meta_anchor *)page_private(page)) argument 94 static inline struct metapage *page_to_mp(struct page *page, int offset) in page_to_mp() argument 96 if (!PagePrivate(page)) in page_to_mp() 98 return mp_anchor(page)->mp[offset >> L2PSIZE]; in page_to_mp() 101 static inline int insert_metapage(struct page *page, struct metapage *mp) in insert_metapage() argument 107 if (PagePrivate(page)) in insert_metapage() 108 a = mp_anchor(page); in insert_metapage() 113 set_page_private(page, (unsigned long)a); in insert_metapage() [all …]
|
/linux-4.19.296/fs/sysv/ |
D | dir.c | 31 static inline void dir_put_page(struct page *page) in dir_put_page() argument 33 kunmap(page); in dir_put_page() 34 put_page(page); in dir_put_page() 37 static int dir_commit_chunk(struct page *page, loff_t pos, unsigned len) in dir_commit_chunk() argument 39 struct address_space *mapping = page->mapping; in dir_commit_chunk() 43 block_write_end(NULL, mapping, pos, len, len, page, NULL); in dir_commit_chunk() 49 err = write_one_page(page); in dir_commit_chunk() 51 unlock_page(page); in dir_commit_chunk() 55 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page() 58 struct page *page = read_mapping_page(mapping, n, NULL); in dir_get_page() local [all …]
|
/linux-4.19.296/include/trace/events/ |
D | page_ref.h | 15 TP_PROTO(struct page *page, int v), 17 TP_ARGS(page, v), 30 __entry->pfn = page_to_pfn(page); 31 __entry->flags = page->flags; 32 __entry->count = page_ref_count(page); 33 __entry->mapcount = page_mapcount(page); 34 __entry->mapping = page->mapping; 35 __entry->mt = get_pageblock_migratetype(page); 49 TP_PROTO(struct page *page, int v), 51 TP_ARGS(page, v) [all …]
|
D | pagemap.h | 19 #define trace_pagemap_flags(page) ( \ argument 20 (PageAnon(page) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \ 21 (page_mapped(page) ? PAGEMAP_MAPPED : 0) | \ 22 (PageSwapCache(page) ? PAGEMAP_SWAPCACHE : 0) | \ 23 (PageSwapBacked(page) ? PAGEMAP_SWAPBACKED : 0) | \ 24 (PageMappedToDisk(page) ? PAGEMAP_MAPPEDDISK : 0) | \ 25 (page_has_private(page) ? PAGEMAP_BUFFERS : 0) \ 31 struct page *page, 35 TP_ARGS(page, lru), 38 __field(struct page *, page ) [all …]
|
/linux-4.19.296/fs/ecryptfs/ |
D | mmap.c | 47 struct page *ecryptfs_get_locked_page(struct inode *inode, loff_t index) in ecryptfs_get_locked_page() 49 struct page *page = read_mapping_page(inode->i_mapping, index, NULL); in ecryptfs_get_locked_page() local 50 if (!IS_ERR(page)) in ecryptfs_get_locked_page() 51 lock_page(page); in ecryptfs_get_locked_page() 52 return page; in ecryptfs_get_locked_page() 65 static int ecryptfs_writepage(struct page *page, struct writeback_control *wbc) in ecryptfs_writepage() argument 69 rc = ecryptfs_encrypt_page(page); in ecryptfs_writepage() 72 "page (upper index [0x%.16lx])\n", page->index); in ecryptfs_writepage() 73 ClearPageUptodate(page); in ecryptfs_writepage() 76 SetPageUptodate(page); in ecryptfs_writepage() [all …]
|
/linux-4.19.296/include/drm/ttm/ |
D | ttm_set_memory.h | 40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() 45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc() 50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc() 55 static inline int ttm_set_pages_wb(struct page *page, int numpages) in ttm_set_pages_wb() argument 57 return set_pages_wb(page, numpages); in ttm_set_pages_wb() 60 static inline int ttm_set_pages_wc(struct page *page, int numpages) in ttm_set_pages_wc() argument 62 unsigned long addr = (unsigned long)page_address(page); in ttm_set_pages_wc() 67 static inline int ttm_set_pages_uc(struct page *page, int numpages) in ttm_set_pages_uc() argument 69 return set_pages_uc(page, numpages); in ttm_set_pages_uc() 78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb() [all …]
|
/linux-4.19.296/fs/nilfs2/ |
D | page.c | 29 __nilfs_get_page_block(struct page *page, unsigned long block, pgoff_t index, in __nilfs_get_page_block() argument 36 if (!page_has_buffers(page)) in __nilfs_get_page_block() 37 create_empty_buffers(page, 1 << blkbits, b_state); in __nilfs_get_page_block() 40 bh = nilfs_page_get_nth_block(page, block - first_block); in __nilfs_get_page_block() 54 struct page *page; in nilfs_grab_buffer() local 57 page = grab_cache_page(mapping, index); in nilfs_grab_buffer() 58 if (unlikely(!page)) in nilfs_grab_buffer() 61 bh = __nilfs_get_page_block(page, blkoff, index, blkbits, b_state); in nilfs_grab_buffer() 63 unlock_page(page); in nilfs_grab_buffer() 64 put_page(page); in nilfs_grab_buffer() [all …]
|
D | dir.c | 67 static inline void nilfs_put_page(struct page *page) in nilfs_put_page() argument 69 kunmap(page); in nilfs_put_page() 70 put_page(page); in nilfs_put_page() 87 static int nilfs_prepare_chunk(struct page *page, unsigned int from, in nilfs_prepare_chunk() argument 90 loff_t pos = page_offset(page) + from; in nilfs_prepare_chunk() 92 return __block_write_begin(page, pos, to - from, nilfs_get_block); in nilfs_prepare_chunk() 95 static void nilfs_commit_chunk(struct page *page, in nilfs_commit_chunk() argument 100 loff_t pos = page_offset(page) + from; in nilfs_commit_chunk() 105 nr_dirty = nilfs_page_count_clean_buffers(page, from, to); in nilfs_commit_chunk() 106 copied = block_write_end(NULL, mapping, pos, len, len, page, NULL); in nilfs_commit_chunk() [all …]
|
/linux-4.19.296/fs/squashfs/ |
D | file_direct.c | 23 static int squashfs_read_cache(struct page *target_page, u64 block, int bsize, 24 int pages, struct page **page, int bytes); 27 int squashfs_readpage_block(struct page *target_page, u64 block, int bsize, in squashfs_readpage_block() 39 struct page **page; in squashfs_readpage_block() local 48 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block() 49 if (page == NULL) in squashfs_readpage_block() 56 actor = squashfs_page_actor_init_special(page, pages, 0); in squashfs_readpage_block() 62 page[i] = (n == target_page->index) ? target_page : in squashfs_readpage_block() 65 if (page[i] == NULL) { in squashfs_readpage_block() 70 if (PageUptodate(page[i])) { in squashfs_readpage_block() [all …]
|