Home
last modified time | relevance | path

Searched refs:page (Results 1 – 25 of 709) sorted by relevance

12345678910>>...29

/linux-4.19.296/include/linux/
Dpage-flags.h138 struct page; /* forward declaration */
140 static inline struct page *compound_head(struct page *page) in compound_head() argument
142 unsigned long head = READ_ONCE(page->compound_head); in compound_head()
145 return (struct page *) (head - 1); in compound_head()
146 return page; in compound_head()
149 static __always_inline int PageTail(struct page *page) in PageTail() argument
151 return READ_ONCE(page->compound_head) & 1; in PageTail()
154 static __always_inline int PageCompound(struct page *page) in PageCompound() argument
156 return test_bit(PG_head, &page->flags) || PageTail(page); in PageCompound()
160 static inline int PagePoisoned(const struct page *page) in PagePoisoned() argument
[all …]
Dpage_ref.h29 extern void __page_ref_set(struct page *page, int v);
30 extern void __page_ref_mod(struct page *page, int v);
31 extern void __page_ref_mod_and_test(struct page *page, int v, int ret);
32 extern void __page_ref_mod_and_return(struct page *page, int v, int ret);
33 extern void __page_ref_mod_unless(struct page *page, int v, int u);
34 extern void __page_ref_freeze(struct page *page, int v, int ret);
35 extern void __page_ref_unfreeze(struct page *page, int v);
41 static inline void __page_ref_set(struct page *page, int v) in __page_ref_set() argument
44 static inline void __page_ref_mod(struct page *page, int v) in __page_ref_mod() argument
47 static inline void __page_ref_mod_and_test(struct page *page, int v, int ret) in __page_ref_mod_and_test() argument
[all …]
Dballoon_compaction.h66 int (*migratepage)(struct balloon_dev_info *, struct page *newpage,
67 struct page *page, enum migrate_mode mode);
71 extern struct page *balloon_page_alloc(void);
73 struct page *page);
74 extern struct page *balloon_page_dequeue(struct balloon_dev_info *b_dev_info);
87 extern bool balloon_page_isolate(struct page *page,
89 extern void balloon_page_putback(struct page *page);
91 struct page *newpage,
92 struct page *page, enum migrate_mode mode);
104 struct page *page) in balloon_page_insert() argument
[all …]
Dpage_idle.h12 static inline bool page_is_young(struct page *page) in page_is_young() argument
14 return PageYoung(page); in page_is_young()
17 static inline void set_page_young(struct page *page) in set_page_young() argument
19 SetPageYoung(page); in set_page_young()
22 static inline bool test_and_clear_page_young(struct page *page) in test_and_clear_page_young() argument
24 return TestClearPageYoung(page); in test_and_clear_page_young()
27 static inline bool page_is_idle(struct page *page) in page_is_idle() argument
29 return PageIdle(page); in page_is_idle()
32 static inline void set_page_idle(struct page *page) in set_page_idle() argument
34 SetPageIdle(page); in set_page_idle()
[all …]
Dpagemap.h121 void release_pages(struct page **pages, int nr);
167 static inline int page_cache_get_speculative(struct page *page) in page_cache_get_speculative() argument
182 VM_BUG_ON_PAGE(page_count(page) == 0, page); in page_cache_get_speculative()
183 page_ref_inc(page); in page_cache_get_speculative()
186 if (unlikely(!get_page_unless_zero(page))) { in page_cache_get_speculative()
195 VM_BUG_ON_PAGE(PageTail(page), page); in page_cache_get_speculative()
203 static inline int page_cache_add_speculative(struct page *page, int count) in page_cache_add_speculative() argument
211 VM_BUG_ON_PAGE(page_count(page) == 0, page); in page_cache_add_speculative()
212 page_ref_add(page, count); in page_cache_add_speculative()
215 if (unlikely(!page_ref_add_unless(page, count, 0))) in page_cache_add_speculative()
[all …]
Dpage_owner.h11 extern void __reset_page_owner(struct page *page, unsigned int order);
12 extern void __set_page_owner(struct page *page,
14 extern void __split_page_owner(struct page *page, unsigned int order);
15 extern void __copy_page_owner(struct page *oldpage, struct page *newpage);
16 extern void __set_page_owner_migrate_reason(struct page *page, int reason);
17 extern void __dump_page_owner(struct page *page);
21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument
24 __reset_page_owner(page, order); in reset_page_owner()
27 static inline void set_page_owner(struct page *page, in set_page_owner() argument
31 __set_page_owner(page, order, gfp_mask); in set_page_owner()
[all …]
Dmigrate.h10 typedef struct page *new_page_t(struct page *page, unsigned long private);
11 typedef void free_page_t(struct page *page, unsigned long private);
34 static inline struct page *new_page_nodemask(struct page *page, in new_page_nodemask() argument
39 struct page *new_page = NULL; in new_page_nodemask()
41 if (PageHuge(page)) in new_page_nodemask()
42 return alloc_huge_page_nodemask(page_hstate(compound_head(page)), in new_page_nodemask()
45 if (PageTransHuge(page)) { in new_page_nodemask()
50 if (PageHighMem(page) || (zone_idx(page_zone(page)) == ZONE_MOVABLE)) in new_page_nodemask()
66 struct page *newpage, struct page *page,
70 extern int isolate_movable_page(struct page *page, isolate_mode_t mode);
[all …]
Dmm.h105 #define mm_zero_struct_page(pp) ((void)memset((pp), 0, sizeof(struct page)))
141 #define nth_page(page,n) pfn_to_page(page_to_pfn((page)) + (n)) argument
356 struct page *cow_page; /* Page handler may use for COW fault */
358 struct page *page; /* ->fault handlers should return a member
451 struct page *(*find_special_page)(struct vm_area_struct *vma,
476 #define page_private(page) ((page)->private) argument
477 #define set_page_private(page, v) ((page)->private = (v)) argument
517 static inline int put_page_testzero(struct page *page) in put_page_testzero() argument
519 VM_BUG_ON_PAGE(page_ref_count(page) == 0, page); in put_page_testzero()
520 return page_ref_dec_and_test(page); in put_page_testzero()
[all …]
Dswap.h307 void *workingset_eviction(struct address_space *mapping, struct page *page);
309 void workingset_activation(struct page *page);
334 extern void lru_cache_add(struct page *);
335 extern void lru_cache_add_anon(struct page *page);
336 extern void lru_cache_add_file(struct page *page);
337 extern void lru_add_page_tail(struct page *page, struct page *page_tail,
339 extern void activate_page(struct page *);
340 extern void mark_page_accessed(struct page *);
344 extern void rotate_reclaimable_page(struct page *page);
345 extern void deactivate_file_page(struct page *page);
[all …]
Dhighmem.h15 static inline void flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vma… in flush_anon_page() argument
21 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument
43 struct page *kmap_to_page(void *addr);
49 static inline struct page *kmap_to_page(void *addr) in kmap_to_page()
57 static inline void *kmap(struct page *page) in kmap() argument
60 return page_address(page); in kmap()
63 static inline void kunmap(struct page *page) in kunmap() argument
67 static inline void *kmap_atomic(struct page *page) in kmap_atomic() argument
71 return page_address(page); in kmap_atomic()
73 #define kmap_atomic_prot(page, prot) kmap_atomic(page) argument
[all …]
Dhugetlb_cgroup.h29 static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) in hugetlb_cgroup_from_page() argument
31 VM_BUG_ON_PAGE(!PageHuge(page), page); in hugetlb_cgroup_from_page()
33 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in hugetlb_cgroup_from_page()
35 return (struct hugetlb_cgroup *)page[2].private; in hugetlb_cgroup_from_page()
39 int set_hugetlb_cgroup(struct page *page, struct hugetlb_cgroup *h_cg) in set_hugetlb_cgroup() argument
41 VM_BUG_ON_PAGE(!PageHuge(page), page); in set_hugetlb_cgroup()
43 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in set_hugetlb_cgroup()
45 page[2].private = (unsigned long)h_cg; in set_hugetlb_cgroup()
58 struct page *page);
60 struct page *page);
[all …]
Dmm_inline.h21 static inline int page_is_file_cache(struct page *page) in page_is_file_cache() argument
23 return !PageSwapBacked(page); in page_is_file_cache()
47 static __always_inline void add_page_to_lru_list(struct page *page, in add_page_to_lru_list() argument
50 update_lru_size(lruvec, lru, page_zonenum(page), hpage_nr_pages(page)); in add_page_to_lru_list()
51 list_add(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list()
54 static __always_inline void add_page_to_lru_list_tail(struct page *page, in add_page_to_lru_list_tail() argument
57 update_lru_size(lruvec, lru, page_zonenum(page), hpage_nr_pages(page)); in add_page_to_lru_list_tail()
58 list_add_tail(&page->lru, &lruvec->lists[lru]); in add_page_to_lru_list_tail()
61 static __always_inline void del_page_from_lru_list(struct page *page, in del_page_from_lru_list() argument
64 list_del(&page->lru); in del_page_from_lru_list()
[all …]
Drmap.h167 struct anon_vma *page_get_anon_vma(struct page *page);
176 void page_move_anon_rmap(struct page *, struct vm_area_struct *);
177 void page_add_anon_rmap(struct page *, struct vm_area_struct *,
179 void do_page_add_anon_rmap(struct page *, struct vm_area_struct *,
181 void page_add_new_anon_rmap(struct page *, struct vm_area_struct *,
183 void page_add_file_rmap(struct page *, bool);
184 void page_remove_rmap(struct page *, bool);
186 void hugepage_add_anon_rmap(struct page *, struct vm_area_struct *,
188 void hugepage_add_new_anon_rmap(struct page *, struct vm_area_struct *,
191 static inline void page_dup_rmap(struct page *page, bool compound) in page_dup_rmap() argument
[all …]
Dpageblock-flags.h66 struct page;
68 unsigned long get_pfnblock_flags_mask(struct page *page,
73 void set_pfnblock_flags_mask(struct page *page,
80 #define get_pageblock_flags_group(page, start_bitidx, end_bitidx) \ argument
81 get_pfnblock_flags_mask(page, page_to_pfn(page), \
84 #define set_pageblock_flags_group(page, flags, start_bitidx, end_bitidx) \ argument
85 set_pfnblock_flags_mask(page, flags, page_to_pfn(page), \
90 #define get_pageblock_skip(page) \ argument
91 get_pageblock_flags_group(page, PB_migrate_skip, \
93 #define clear_pageblock_skip(page) \ argument
[all …]
/linux-4.19.296/fs/9p/
Dcache.h45 extern int __v9fs_fscache_release_page(struct page *page, gfp_t gfp);
46 extern void __v9fs_fscache_invalidate_page(struct page *page);
48 struct page *page);
53 extern void __v9fs_readpage_to_fscache(struct inode *inode, struct page *page);
55 struct page *page);
57 static inline int v9fs_fscache_release_page(struct page *page, in v9fs_fscache_release_page() argument
60 return __v9fs_fscache_release_page(page, gfp); in v9fs_fscache_release_page()
63 static inline void v9fs_fscache_invalidate_page(struct page *page) in v9fs_fscache_invalidate_page() argument
65 __v9fs_fscache_invalidate_page(page); in v9fs_fscache_invalidate_page()
69 struct page *page) in v9fs_readpage_from_fscache() argument
[all …]
Dvfs_addr.c53 static int v9fs_fid_readpage(void *data, struct page *page) in v9fs_fid_readpage() argument
56 struct inode *inode = page->mapping->host; in v9fs_fid_readpage()
57 struct bio_vec bvec = {.bv_page = page, .bv_len = PAGE_SIZE}; in v9fs_fid_readpage()
63 BUG_ON(!PageLocked(page)); in v9fs_fid_readpage()
65 retval = v9fs_readpage_from_fscache(inode, page); in v9fs_fid_readpage()
71 retval = p9_client_read(fid, page_offset(page), &to, &err); in v9fs_fid_readpage()
73 v9fs_uncache_page(inode, page); in v9fs_fid_readpage()
78 zero_user(page, retval, PAGE_SIZE - retval); in v9fs_fid_readpage()
79 flush_dcache_page(page); in v9fs_fid_readpage()
80 SetPageUptodate(page); in v9fs_fid_readpage()
[all …]
/linux-4.19.296/fs/jfs/
Djfs_metapage.c61 unlock_page(mp->page); in __lock_metapage()
63 lock_page(mp->page); in __lock_metapage()
92 #define mp_anchor(page) ((struct meta_anchor *)page_private(page)) argument
94 static inline struct metapage *page_to_mp(struct page *page, int offset) in page_to_mp() argument
96 if (!PagePrivate(page)) in page_to_mp()
98 return mp_anchor(page)->mp[offset >> L2PSIZE]; in page_to_mp()
101 static inline int insert_metapage(struct page *page, struct metapage *mp) in insert_metapage() argument
107 if (PagePrivate(page)) in insert_metapage()
108 a = mp_anchor(page); in insert_metapage()
113 set_page_private(page, (unsigned long)a); in insert_metapage()
[all …]
/linux-4.19.296/fs/sysv/
Ddir.c31 static inline void dir_put_page(struct page *page) in dir_put_page() argument
33 kunmap(page); in dir_put_page()
34 put_page(page); in dir_put_page()
37 static int dir_commit_chunk(struct page *page, loff_t pos, unsigned len) in dir_commit_chunk() argument
39 struct address_space *mapping = page->mapping; in dir_commit_chunk()
43 block_write_end(NULL, mapping, pos, len, len, page, NULL); in dir_commit_chunk()
49 err = write_one_page(page); in dir_commit_chunk()
51 unlock_page(page); in dir_commit_chunk()
55 static struct page * dir_get_page(struct inode *dir, unsigned long n) in dir_get_page()
58 struct page *page = read_mapping_page(mapping, n, NULL); in dir_get_page() local
[all …]
/linux-4.19.296/include/trace/events/
Dpage_ref.h15 TP_PROTO(struct page *page, int v),
17 TP_ARGS(page, v),
30 __entry->pfn = page_to_pfn(page);
31 __entry->flags = page->flags;
32 __entry->count = page_ref_count(page);
33 __entry->mapcount = page_mapcount(page);
34 __entry->mapping = page->mapping;
35 __entry->mt = get_pageblock_migratetype(page);
49 TP_PROTO(struct page *page, int v),
51 TP_ARGS(page, v)
[all …]
Dpagemap.h19 #define trace_pagemap_flags(page) ( \ argument
20 (PageAnon(page) ? PAGEMAP_ANONYMOUS : PAGEMAP_FILE) | \
21 (page_mapped(page) ? PAGEMAP_MAPPED : 0) | \
22 (PageSwapCache(page) ? PAGEMAP_SWAPCACHE : 0) | \
23 (PageSwapBacked(page) ? PAGEMAP_SWAPBACKED : 0) | \
24 (PageMappedToDisk(page) ? PAGEMAP_MAPPEDDISK : 0) | \
25 (page_has_private(page) ? PAGEMAP_BUFFERS : 0) \
31 struct page *page,
35 TP_ARGS(page, lru),
38 __field(struct page *, page )
[all …]
/linux-4.19.296/fs/ecryptfs/
Dmmap.c47 struct page *ecryptfs_get_locked_page(struct inode *inode, loff_t index) in ecryptfs_get_locked_page()
49 struct page *page = read_mapping_page(inode->i_mapping, index, NULL); in ecryptfs_get_locked_page() local
50 if (!IS_ERR(page)) in ecryptfs_get_locked_page()
51 lock_page(page); in ecryptfs_get_locked_page()
52 return page; in ecryptfs_get_locked_page()
65 static int ecryptfs_writepage(struct page *page, struct writeback_control *wbc) in ecryptfs_writepage() argument
69 rc = ecryptfs_encrypt_page(page); in ecryptfs_writepage()
72 "page (upper index [0x%.16lx])\n", page->index); in ecryptfs_writepage()
73 ClearPageUptodate(page); in ecryptfs_writepage()
76 SetPageUptodate(page); in ecryptfs_writepage()
[all …]
/linux-4.19.296/include/drm/ttm/
Dttm_set_memory.h40 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb()
45 static inline int ttm_set_pages_array_wc(struct page **pages, int addrinarray) in ttm_set_pages_array_wc()
50 static inline int ttm_set_pages_array_uc(struct page **pages, int addrinarray) in ttm_set_pages_array_uc()
55 static inline int ttm_set_pages_wb(struct page *page, int numpages) in ttm_set_pages_wb() argument
57 return set_pages_wb(page, numpages); in ttm_set_pages_wb()
60 static inline int ttm_set_pages_wc(struct page *page, int numpages) in ttm_set_pages_wc() argument
62 unsigned long addr = (unsigned long)page_address(page); in ttm_set_pages_wc()
67 static inline int ttm_set_pages_uc(struct page *page, int numpages) in ttm_set_pages_uc() argument
69 return set_pages_uc(page, numpages); in ttm_set_pages_uc()
78 static inline int ttm_set_pages_array_wb(struct page **pages, int addrinarray) in ttm_set_pages_array_wb()
[all …]
/linux-4.19.296/fs/nilfs2/
Dpage.c29 __nilfs_get_page_block(struct page *page, unsigned long block, pgoff_t index, in __nilfs_get_page_block() argument
36 if (!page_has_buffers(page)) in __nilfs_get_page_block()
37 create_empty_buffers(page, 1 << blkbits, b_state); in __nilfs_get_page_block()
40 bh = nilfs_page_get_nth_block(page, block - first_block); in __nilfs_get_page_block()
54 struct page *page; in nilfs_grab_buffer() local
57 page = grab_cache_page(mapping, index); in nilfs_grab_buffer()
58 if (unlikely(!page)) in nilfs_grab_buffer()
61 bh = __nilfs_get_page_block(page, blkoff, index, blkbits, b_state); in nilfs_grab_buffer()
63 unlock_page(page); in nilfs_grab_buffer()
64 put_page(page); in nilfs_grab_buffer()
[all …]
Ddir.c67 static inline void nilfs_put_page(struct page *page) in nilfs_put_page() argument
69 kunmap(page); in nilfs_put_page()
70 put_page(page); in nilfs_put_page()
87 static int nilfs_prepare_chunk(struct page *page, unsigned int from, in nilfs_prepare_chunk() argument
90 loff_t pos = page_offset(page) + from; in nilfs_prepare_chunk()
92 return __block_write_begin(page, pos, to - from, nilfs_get_block); in nilfs_prepare_chunk()
95 static void nilfs_commit_chunk(struct page *page, in nilfs_commit_chunk() argument
100 loff_t pos = page_offset(page) + from; in nilfs_commit_chunk()
105 nr_dirty = nilfs_page_count_clean_buffers(page, from, to); in nilfs_commit_chunk()
106 copied = block_write_end(NULL, mapping, pos, len, len, page, NULL); in nilfs_commit_chunk()
[all …]
/linux-4.19.296/fs/squashfs/
Dfile_direct.c23 static int squashfs_read_cache(struct page *target_page, u64 block, int bsize,
24 int pages, struct page **page, int bytes);
27 int squashfs_readpage_block(struct page *target_page, u64 block, int bsize, in squashfs_readpage_block()
39 struct page **page; in squashfs_readpage_block() local
48 page = kmalloc_array(pages, sizeof(void *), GFP_KERNEL); in squashfs_readpage_block()
49 if (page == NULL) in squashfs_readpage_block()
56 actor = squashfs_page_actor_init_special(page, pages, 0); in squashfs_readpage_block()
62 page[i] = (n == target_page->index) ? target_page : in squashfs_readpage_block()
65 if (page[i] == NULL) { in squashfs_readpage_block()
70 if (PageUptodate(page[i])) { in squashfs_readpage_block()
[all …]

12345678910>>...29