/linux-4.19.296/include/asm-generic/ |
D | sections.h | 94 static inline bool memory_contains(void *begin, void *end, void *virt, in memory_contains() argument 97 return virt >= begin && virt + size <= end; in memory_contains() 111 static inline bool memory_intersects(void *begin, void *end, void *virt, in memory_intersects() argument 114 void *vend = virt + size; in memory_intersects() 116 if (virt < end && vend > begin) in memory_intersects() 131 static inline bool init_section_contains(void *virt, size_t size) in init_section_contains() argument 133 return memory_contains(__init_begin, __init_end, virt, size); in init_section_contains() 145 static inline bool init_section_intersects(void *virt, size_t size) in init_section_intersects() argument 147 return memory_intersects(__init_begin, __init_end, virt, size); in init_section_intersects()
|
/linux-4.19.296/drivers/gpio/ |
D | gpio-stp-xway.c | 82 void __iomem *virt; member 103 return (xway_stp_r32(chip->virt, XWAY_STP_CPU0) & BIT(gpio)); in xway_stp_get() 122 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set() 123 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_set() 167 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); in xway_stp_hw_init() 168 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); in xway_stp_hw_init() 169 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); in xway_stp_hw_init() 170 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_hw_init() 171 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); in xway_stp_hw_init() 174 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MASK, in xway_stp_hw_init() [all …]
|
/linux-4.19.296/include/crypto/ |
D | cbc.h | 26 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_segment() 27 u8 *dst = walk->dst.virt.addr; in crypto_cbc_encrypt_segment() 48 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_inplace() 75 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_cbc_encrypt_walk() 91 u8 *src = walk->src.virt.addr; in crypto_cbc_decrypt_segment() 92 u8 *dst = walk->dst.virt.addr; in crypto_cbc_decrypt_segment() 115 u8 *src = walk->src.virt.addr; in crypto_cbc_decrypt_inplace() 140 if (walk->src.virt.addr == walk->dst.virt.addr) in crypto_cbc_decrypt_blocks()
|
/linux-4.19.296/drivers/char/tpm/eventlog/ |
D | acpi.c | 54 void __iomem *virt; in tpm_read_log_acpi() local 99 virt = acpi_os_map_iomem(start, len); in tpm_read_log_acpi() 100 if (!virt) in tpm_read_log_acpi() 103 memcpy_fromio(log->bios_event_log, virt, len); in tpm_read_log_acpi() 105 acpi_os_unmap_iomem(virt, len); in tpm_read_log_acpi()
|
/linux-4.19.296/crypto/ |
D | blkcipher.c | 46 walk->src.virt.addr = scatterwalk_map(&walk->in); in blkcipher_map_src() 51 walk->dst.virt.addr = scatterwalk_map(&walk->out); in blkcipher_map_dst() 56 scatterwalk_unmap(walk->src.virt.addr); in blkcipher_unmap_src() 61 scatterwalk_unmap(walk->dst.virt.addr); in blkcipher_unmap_dst() 88 memcpy(walk->dst.virt.addr, walk->page, n); in blkcipher_done_fast() 166 walk->dst.virt.addr = (u8 *)ALIGN((unsigned long)walk->buffer, in blkcipher_next_slow() 168 walk->dst.virt.addr = blkcipher_get_spot(walk->dst.virt.addr, bsize); in blkcipher_next_slow() 169 walk->src.virt.addr = blkcipher_get_spot(walk->dst.virt.addr + in blkcipher_next_slow() 172 scatterwalk_copychunks(walk->src.virt.addr, &walk->in, bsize, 0); in blkcipher_next_slow() 185 memcpy(tmp, walk->src.virt.addr, walk->nbytes); in blkcipher_next_copy() [all …]
|
D | cfb.c | 59 u8 *src = walk->src.virt.addr; in crypto_cfb_final() 60 u8 *dst = walk->dst.virt.addr; in crypto_cfb_final() 73 u8 *src = walk->src.virt.addr; in crypto_cfb_encrypt_segment() 74 u8 *dst = walk->dst.virt.addr; in crypto_cfb_encrypt_segment() 96 u8 *src = walk->src.virt.addr; in crypto_cfb_encrypt_inplace() 123 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_cfb_encrypt() 143 u8 *src = walk->src.virt.addr; in crypto_cfb_decrypt_segment() 144 u8 *dst = walk->dst.virt.addr; in crypto_cfb_decrypt_segment() 166 u8 *src = walk->src.virt.addr; in crypto_cfb_decrypt_inplace() 183 if (walk->src.virt.addr == walk->dst.virt.addr) in crypto_cfb_decrypt_blocks()
|
D | pcbc.c | 52 u8 *src = walk->src.virt.addr; in crypto_pcbc_encrypt_segment() 53 u8 *dst = walk->dst.virt.addr; in crypto_pcbc_encrypt_segment() 74 u8 *src = walk->src.virt.addr; in crypto_pcbc_encrypt_inplace() 102 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_pcbc_encrypt() 120 u8 *src = walk->src.virt.addr; in crypto_pcbc_decrypt_segment() 121 u8 *dst = walk->dst.virt.addr; in crypto_pcbc_decrypt_segment() 142 u8 *src = walk->src.virt.addr; in crypto_pcbc_decrypt_inplace() 170 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_pcbc_decrypt()
|
D | skcipher.c | 65 walk->src.virt.addr = skcipher_map(&walk->in); in skcipher_map_src() 70 walk->dst.virt.addr = skcipher_map(&walk->out); in skcipher_map_dst() 75 skcipher_unmap(&walk->in, walk->src.virt.addr); in skcipher_unmap_src() 80 skcipher_unmap(&walk->out, walk->dst.virt.addr); in skcipher_unmap_dst() 133 memcpy(walk->dst.virt.addr, walk->page, n); in skcipher_walk_done() 278 walk->dst.virt.addr = PTR_ALIGN(buffer, alignmask + 1); in skcipher_next_slow() 279 walk->dst.virt.addr = skcipher_get_spot(walk->dst.virt.addr, bsize); in skcipher_next_slow() 280 walk->src.virt.addr = walk->dst.virt.addr; in skcipher_next_slow() 282 scatterwalk_copychunks(walk->src.virt.addr, &walk->in, bsize, 0); in skcipher_next_slow() 296 memcpy(tmp, walk->src.virt.addr, walk->nbytes); in skcipher_next_copy() [all …]
|
D | crypto_null.c | 88 if (walk.src.virt.addr != walk.dst.virt.addr) in skcipher_null_crypt() 89 memcpy(walk.dst.virt.addr, walk.src.virt.addr, in skcipher_null_crypt()
|
D | ctr.c | 63 u8 *src = walk->src.virt.addr; in crypto_ctr_crypt_final() 64 u8 *dst = walk->dst.virt.addr; in crypto_ctr_crypt_final() 80 u8 *src = walk->src.virt.addr; in crypto_ctr_crypt_segment() 81 u8 *dst = walk->dst.virt.addr; in crypto_ctr_crypt_segment() 108 u8 *src = walk->src.virt.addr; in crypto_ctr_crypt_inplace() 141 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_ctr_crypt()
|
D | chacha20_generic.c | 94 chacha20_docrypt(state, walk.dst.virt.addr, walk.src.virt.addr, in crypto_chacha20_crypt()
|
D | arc4.c | 107 u8 *wsrc = walk.src.virt.addr; in ecb_arc4_crypt() 108 u8 *wdst = walk.dst.virt.addr; in ecb_arc4_crypt()
|
/linux-4.19.296/include/linux/mtd/ |
D | map.h | 211 void __iomem *virt; member 408 r.x[0] = __raw_readb(map->virt + ofs); in inline_map_read() 410 r.x[0] = __raw_readw(map->virt + ofs); in inline_map_read() 412 r.x[0] = __raw_readl(map->virt + ofs); in inline_map_read() 415 r.x[0] = __raw_readq(map->virt + ofs); in inline_map_read() 418 memcpy_fromio(r.x, map->virt + ofs, map->bankwidth); in inline_map_read() 428 __raw_writeb(datum.x[0], map->virt + ofs); in inline_map_write() 430 __raw_writew(datum.x[0], map->virt + ofs); in inline_map_write() 432 __raw_writel(datum.x[0], map->virt + ofs); in inline_map_write() 435 __raw_writeq(datum.x[0], map->virt + ofs); in inline_map_write() [all …]
|
/linux-4.19.296/fs/ecryptfs/ |
D | read_write.c | 75 char *virt; in ecryptfs_write_lower_page_segment() local 81 virt = kmap(page_for_lower); in ecryptfs_write_lower_page_segment() 82 rc = ecryptfs_write_lower(ecryptfs_inode, virt, offset, size); in ecryptfs_write_lower_page_segment() 263 char *virt; in ecryptfs_read_lower_page_segment() local 268 virt = kmap(page_for_ecryptfs); in ecryptfs_read_lower_page_segment() 269 rc = ecryptfs_read_lower(virt, offset, size, ecryptfs_inode); in ecryptfs_read_lower_page_segment()
|
D | crypto.c | 1032 ecryptfs_write_header_metadata(char *virt, in ecryptfs_write_header_metadata() argument 1042 put_unaligned_be32(header_extent_size, virt); in ecryptfs_write_header_metadata() 1043 virt += 4; in ecryptfs_write_header_metadata() 1044 put_unaligned_be16(num_header_extents_at_front, virt); in ecryptfs_write_header_metadata() 1114 char *virt, size_t virt_len) in ecryptfs_write_metadata_to_contents() argument 1118 rc = ecryptfs_write_lower(ecryptfs_inode, virt, in ecryptfs_write_metadata_to_contents() 1170 char *virt; in ecryptfs_write_metadata() local 1190 virt = (char *)ecryptfs_get_zeroed_pages(GFP_KERNEL, order); in ecryptfs_write_metadata() 1191 if (!virt) { in ecryptfs_write_metadata() 1197 rc = ecryptfs_write_headers_virt(virt, virt_len, &size, crypt_stat, in ecryptfs_write_metadata() [all …]
|
/linux-4.19.296/block/ |
D | t10-pi.c | 212 u32 virt = bip_get_seed(bip) & 0xffffffff; in t10_pi_prepare() local 229 if (be32_to_cpu(pi->ref_tag) == virt) in t10_pi_prepare() 231 virt++; in t10_pi_prepare() 271 u32 virt = bip_get_seed(bip) & 0xffffffff; in t10_pi_complete() local 285 pi->ref_tag = cpu_to_be32(virt); in t10_pi_complete() 286 virt++; in t10_pi_complete()
|
/linux-4.19.296/include/linux/ |
D | dma-debug.h | 52 dma_addr_t dma_addr, void *virt); 55 void *virt, dma_addr_t addr); 136 dma_addr_t dma_addr, void *virt) in debug_dma_alloc_coherent() argument 141 void *virt, dma_addr_t addr) in debug_dma_free_coherent() argument
|
/linux-4.19.296/drivers/cpufreq/ |
D | tegra186-cpufreq.c | 121 void *virt; in init_vhint_table() local 123 virt = dma_alloc_coherent(bpmp->dev, sizeof(*data), &phys, in init_vhint_table() 125 if (!virt) in init_vhint_table() 128 data = (struct cpu_vhint_data *)virt; in init_vhint_table() 189 dma_free_coherent(bpmp->dev, sizeof(*data), virt, phys); in init_vhint_table()
|
/linux-4.19.296/drivers/isdn/hardware/eicon/ |
D | diva_dma.c | 59 int nr, void *virt, dword phys, in diva_init_dma_map_entry() argument 62 pmap[nr].local_ram_addr = virt; in diva_init_dma_map_entry()
|
D | diva_dma.h | 31 int nr, void *virt, dword phys,
|
/linux-4.19.296/drivers/sfi/ |
D | sfi_core.c | 110 static void __ref sfi_unmap_memory(void __iomem *virt, u32 size) in sfi_unmap_memory() argument 112 if (!virt || !size) in sfi_unmap_memory() 116 memunmap(virt); in sfi_unmap_memory() 118 early_memunmap(virt, size); in sfi_unmap_memory()
|
/linux-4.19.296/drivers/misc/mic/scif/ |
D | scif_map.h | 105 scif_iounmap(void *virt, size_t len, struct scif_dev *scifdev) in scif_iounmap() argument 110 sdev->hw_ops->iounmap(sdev, (void __force __iomem *)virt); in scif_iounmap()
|
/linux-4.19.296/drivers/xen/ |
D | gntdev.c | 924 static int gntdev_get_page(struct gntdev_copy_batch *batch, void __user *virt, in gntdev_get_page() argument 927 unsigned long addr = (unsigned long)virt; in gntdev_get_page() 1018 void __user *virt; in gntdev_grant_copy_seg() local 1040 virt = seg->source.virt + copied; in gntdev_grant_copy_seg() 1041 off = (unsigned long)virt & ~XEN_PAGE_MASK; in gntdev_grant_copy_seg() 1045 ret = gntdev_get_page(batch, virt, &gfn); in gntdev_grant_copy_seg() 1060 virt = seg->dest.virt + copied; in gntdev_grant_copy_seg() 1061 off = (unsigned long)virt & ~XEN_PAGE_MASK; in gntdev_grant_copy_seg() 1065 ret = gntdev_get_page(batch, virt, &gfn); in gntdev_grant_copy_seg()
|
/linux-4.19.296/include/acpi/ |
D | acpi_io.h | 21 void __ref acpi_os_unmap_iomem(void __iomem *virt, acpi_size size);
|
/linux-4.19.296/virt/kvm/arm/vgic/ |
D | trace.h | 33 #define TRACE_INCLUDE_PATH ../../../virt/kvm/arm/vgic
|