/linux-4.19.296/fs/jffs2/ |
D | file.c | 26 struct page *pg, void *fsdata); 30 static int jffs2_readpage (struct file *filp, struct page *pg); 79 static int jffs2_do_readpage_nolock (struct inode *inode, struct page *pg) in jffs2_do_readpage_nolock() argument 87 __func__, inode->i_ino, pg->index << PAGE_SHIFT); in jffs2_do_readpage_nolock() 89 BUG_ON(!PageLocked(pg)); in jffs2_do_readpage_nolock() 91 pg_buf = kmap(pg); in jffs2_do_readpage_nolock() 94 ret = jffs2_read_inode_range(c, f, pg_buf, pg->index << PAGE_SHIFT, in jffs2_do_readpage_nolock() 98 ClearPageUptodate(pg); in jffs2_do_readpage_nolock() 99 SetPageError(pg); in jffs2_do_readpage_nolock() 101 SetPageUptodate(pg); in jffs2_do_readpage_nolock() [all …]
|
D | fs.c | 687 struct page *pg; in jffs2_gc_fetch_page() local 689 pg = read_cache_page(inode->i_mapping, offset >> PAGE_SHIFT, in jffs2_gc_fetch_page() 691 if (IS_ERR(pg)) in jffs2_gc_fetch_page() 692 return (void *)pg; in jffs2_gc_fetch_page() 694 *priv = (unsigned long)pg; in jffs2_gc_fetch_page() 695 return kmap(pg); in jffs2_gc_fetch_page() 702 struct page *pg = (void *)*priv; in jffs2_gc_release_page() local 704 kunmap(pg); in jffs2_gc_release_page() 705 put_page(pg); in jffs2_gc_release_page()
|
D | os-linux.h | 158 int jffs2_do_readpage_unlock (struct inode *inode, struct page *pg); 187 unsigned char *pg,
|
/linux-4.19.296/drivers/gpio/ |
D | gpio-palmas.c | 40 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_get() local 41 struct palmas *palmas = pg->palmas; in palmas_gpio_get() 72 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_set() local 73 struct palmas *palmas = pg->palmas; in palmas_gpio_set() 94 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_output() local 95 struct palmas *palmas = pg->palmas; in palmas_gpio_output() 116 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_input() local 117 struct palmas *palmas = pg->palmas; in palmas_gpio_input() 134 struct palmas_gpio *pg = gpiochip_get_data(gc); in palmas_gpio_to_irq() local 135 struct palmas *palmas = pg->palmas; in palmas_gpio_to_irq()
|
/linux-4.19.296/include/asm-generic/ |
D | memory_model.h | 44 #define __page_to_pfn(pg) \ argument 45 ({ const struct page *__pg = (pg); \ 62 #define __page_to_pfn(pg) \ argument 63 ({ const struct page *__pg = (pg); \
|
D | cacheflush.h | 22 #define flush_icache_page(vma,pg) do { } while (0) argument 23 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) argument
|
D | page.h | 31 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 32 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
/linux-4.19.296/include/scsi/ |
D | osd_attributes.h | 12 #define ATTR_SET(pg, id, l, ptr) \ argument 13 { .attr_page = pg, .attr_id = id, .len = l, .val_ptr = ptr } 15 #define ATTR_DEF(pg, id, l) ATTR_SET(pg, id, l, NULL) argument
|
/linux-4.19.296/drivers/char/agp/ |
D | i460-agp.c | 403 int i, start_offset, end_offset, idx, pg, num_entries; in i460_insert_memory_large_io_page() local 443 pg = lp - i460.lp_desc; in i460_insert_memory_large_io_page() 444 WR_GATT(pg, i460_mask_memory(agp_bridge, in i460_insert_memory_large_io_page() 446 WR_FLUSH_GATT(pg); in i460_insert_memory_large_io_page() 464 int i, pg, start_offset, end_offset, idx, num_entries; in i460_remove_memory_large_io_page() local 489 pg = lp - i460.lp_desc; in i460_remove_memory_large_io_page() 490 WR_GATT(pg, 0); in i460_remove_memory_large_io_page() 491 WR_FLUSH_GATT(pg); in i460_remove_memory_large_io_page()
|
D | intel-gtt.c | 847 unsigned int pg, in intel_gtt_insert_page() argument 850 intel_private.driver->write_entry(addr, pg, flags); in intel_gtt_insert_page() 851 readl(intel_private.gtt + pg); in intel_gtt_insert_page()
|
/linux-4.19.296/drivers/xen/ |
D | xen-scsiback.c | 417 struct page **pg, grant_handle_t *grant, int cnt) in scsiback_gnttab_data_map_batch() argument 424 err = gnttab_map_refs(map, NULL, pg, cnt); in scsiback_gnttab_data_map_batch() 432 get_page(pg[i]); in scsiback_gnttab_data_map_batch() 440 struct scsiif_request_segment *seg, struct page **pg, in scsiback_gnttab_data_map_list() argument 448 if (get_free_page(pg + mapcount)) { in scsiback_gnttab_data_map_list() 449 put_free_pages(pg, mapcount); in scsiback_gnttab_data_map_list() 453 gnttab_set_map_op(&map[mapcount], vaddr_page(pg[mapcount]), in scsiback_gnttab_data_map_list() 458 err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); in scsiback_gnttab_data_map_list() 459 pg += mapcount; in scsiback_gnttab_data_map_list() 466 err = scsiback_gnttab_data_map_batch(map, pg, grant, mapcount); in scsiback_gnttab_data_map_list() [all …]
|
/linux-4.19.296/drivers/media/common/saa7146/ |
D | saa7146_core.c | 152 struct page *pg; in vmalloc_to_sg() local 160 pg = vmalloc_to_page(virt); in vmalloc_to_sg() 161 if (NULL == pg) in vmalloc_to_sg() 163 BUG_ON(PageHighMem(pg)); in vmalloc_to_sg() 164 sg_set_page(&sglist[i], pg, PAGE_SIZE, 0); in vmalloc_to_sg()
|
/linux-4.19.296/include/drm/ |
D | intel-gtt.h | 19 unsigned int pg,
|
/linux-4.19.296/crypto/ |
D | ahash.c | 53 walk->data = kmap(walk->pg); in hash_walk_next() 55 walk->data = kmap_atomic(walk->pg); in hash_walk_next() 75 walk->pg = sg_page(walk->sg) + (walk->offset >> PAGE_SHIFT); in hash_walk_new_entry() 106 kunmap(walk->pg); in crypto_hash_walk_done() 121 walk->pg++; in crypto_hash_walk_done()
|
/linux-4.19.296/drivers/media/v4l2-core/ |
D | videobuf-dma-sg.c | 69 struct page *pg; in videobuf_vmalloc_to_sg() local 77 pg = vmalloc_to_page(virt); in videobuf_vmalloc_to_sg() 78 if (NULL == pg) in videobuf_vmalloc_to_sg() 80 BUG_ON(PageHighMem(pg)); in videobuf_vmalloc_to_sg() 81 sg_set_page(&sglist[i], pg, PAGE_SIZE, 0); in videobuf_vmalloc_to_sg()
|
/linux-4.19.296/include/linux/ |
D | lightnvm.h | 56 u64 pg : NVM_12_PG_BITS; member 444 l.ppa |= ((u64)r.g.pg) << ppaf->pg_offset; in generic_to_dev_addr() 473 l.g.pg = (r.ppa & ppaf->pg_mask) >> ppaf->pg_offset; in dev_to_generic_addr()
|
/linux-4.19.296/drivers/s390/crypto/ |
D | pkey_api.c | 778 u8 *rarray, *varray, *pg; in fetch_mkvp() local 780 pg = (u8 *) __get_free_page(GFP_KERNEL); in fetch_mkvp() 781 if (!pg) in fetch_mkvp() 783 rarray = pg; in fetch_mkvp() 784 varray = pg + PAGE_SIZE/2; in fetch_mkvp() 798 free_page((unsigned long) pg); in fetch_mkvp()
|
/linux-4.19.296/drivers/lightnvm/ |
D | pblk.h | 1009 ppa.g.pg = (paddr & ppaf->pg_mask) >> ppaf->pg_offset; in addr_to_gen_ppa() 1049 paddr |= (u64)p.g.pg << ppaf->pg_offset; in pblk_dev_ppa_to_line_addr() 1093 ppa64.g.pg = (ppa32 & ppaf->pg_mask) >> in pblk_ppa32_to_ppa64() 1136 ppa32 |= ppa64.g.pg << ppaf->pg_offset; in pblk_ppa64_to_ppa32() 1314 p->g.pg, p->g.pl, p->g.sec); in print_ppa() 1356 ppa->g.pg < geo->num_pg && in pblk_boundary_ppa_checks()
|
/linux-4.19.296/fs/fuse/ |
D | dev.c | 712 struct page *pg; member 735 } else if (cs->pg) { in fuse_copy_finish() 737 flush_dcache_page(cs->pg); in fuse_copy_finish() 738 set_page_dirty_lock(cs->pg); in fuse_copy_finish() 740 put_page(cs->pg); in fuse_copy_finish() 742 cs->pg = NULL; in fuse_copy_finish() 769 cs->pg = buf->page; in fuse_copy_fill() 787 cs->pg = page; in fuse_copy_fill() 801 cs->pg = page; in fuse_copy_fill() 813 void *pgaddr = kmap_atomic(cs->pg); in fuse_copy_do() [all …]
|
/linux-4.19.296/drivers/edac/ |
D | edac_mc.c | 837 struct page *pg; in edac_mc_scrub_block() local 848 pg = pfn_to_page(page); in edac_mc_scrub_block() 850 if (PageHighMem(pg)) in edac_mc_scrub_block() 853 virt_addr = kmap_atomic(pg); in edac_mc_scrub_block() 861 if (PageHighMem(pg)) in edac_mc_scrub_block()
|
/linux-4.19.296/fs/exofs/ |
D | ore.c | 611 unsigned pg = *cur_pg; in _ore_add_stripe_unit() local 644 added_len = bio_add_pc_page(q, per_dev->bio, pages[pg], in _ore_add_stripe_unit() 656 _add_stripe_page(ios->sp2d, &ios->si, pages[pg]); in _ore_add_stripe_unit() 659 ++pg; in _ore_add_stripe_unit() 664 *cur_pg = pg; in _ore_add_stripe_unit()
|
/linux-4.19.296/drivers/memstick/host/ |
D | tifm_ms.c | 189 struct page *pg; in tifm_ms_transfer_data() local 206 pg = nth_page(sg_page(&host->req->sg), in tifm_ms_transfer_data() 213 buf = kmap_atomic(pg) + p_off; in tifm_ms_transfer_data()
|
D | jmb38x_ms.c | 307 struct page *pg; in jmb38x_ms_transfer_data() local 322 pg = nth_page(sg_page(&host->req->sg), in jmb38x_ms_transfer_data() 329 buf = kmap_atomic(pg) + p_off; in jmb38x_ms_transfer_data()
|
/linux-4.19.296/include/crypto/internal/ |
D | hash.h | 28 struct page *pg; member
|
/linux-4.19.296/include/linux/qed/ |
D | qed_eth_if.h | 280 int (*cee_peer_getpg)(struct qed_dev *cdev, struct cee_pg *pg);
|