/linux-4.19.296/include/linux/ |
D | dma-debug.h | 49 int nelems, int dir); 85 int nelems, int direction); 89 int nelems, int direction); 131 int nelems, int dir) in debug_dma_unmap_sg() argument 187 int nelems, int direction) in debug_dma_sync_sg_for_cpu() argument 193 int nelems, int direction) in debug_dma_sync_sg_for_device() argument
|
D | swiotlb.h | 84 swiotlb_map_sg_attrs(struct device *hwdev, struct scatterlist *sgl, int nelems, 90 int nelems, enum dma_data_direction dir, 99 int nelems, enum dma_data_direction dir); 107 int nelems, enum dma_data_direction dir);
|
D | pci-dma-compat.h | 94 int nelems, int direction) in pci_dma_sync_sg_for_cpu() argument 96 dma_sync_sg_for_cpu(&hwdev->dev, sg, nelems, (enum dma_data_direction)direction); in pci_dma_sync_sg_for_cpu() 101 int nelems, int direction) in pci_dma_sync_sg_for_device() argument 103 dma_sync_sg_for_device(&hwdev->dev, sg, nelems, (enum dma_data_direction)direction); in pci_dma_sync_sg_for_device()
|
D | rhashtable.h | 159 return atomic_read(&ht->nelems) > (tbl->size / 4 * 3) && in rht_grow_above_75() 172 return atomic_read(&ht->nelems) < (tbl->size * 3 / 10) && in rht_shrink_below_30() 184 return atomic_read(&ht->nelems) > tbl->size && in rht_grow_above_100() 196 return atomic_read(&ht->nelems) >= ht->max_elems; in rht_grow_above_max() 663 atomic_inc(&ht->nelems); in __rhashtable_insert_fast() 935 atomic_dec(&ht->nelems); in __rhashtable_remove_fast_one()
|
D | dma-mapping.h | 406 int nelems, enum dma_data_direction dir) in dma_sync_sg_for_cpu() argument 412 ops->sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu() 413 debug_dma_sync_sg_for_cpu(dev, sg, nelems, dir); in dma_sync_sg_for_cpu() 418 int nelems, enum dma_data_direction dir) in dma_sync_sg_for_device() argument 424 ops->sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device() 425 debug_dma_sync_sg_for_device(dev, sg, nelems, dir); in dma_sync_sg_for_device()
|
D | rhashtable-types.h | 92 atomic_t nelems; member
|
/linux-4.19.296/drivers/xen/ |
D | swiotlb-xen.c | 523 int nelems, enum dma_data_direction dir, in xen_swiotlb_unmap_sg_attrs() argument 531 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_unmap_sg_attrs() 554 int nelems, enum dma_data_direction dir, in xen_swiotlb_map_sg_attrs() argument 562 for_each_sg(sgl, sg, nelems, i) { in xen_swiotlb_map_sg_attrs() 607 return nelems; in xen_swiotlb_map_sg_attrs() 619 int nelems, enum dma_data_direction dir, in xen_swiotlb_sync_sg() argument 625 for_each_sg(sgl, sg, nelems, i) in xen_swiotlb_sync_sg() 632 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_cpu() argument 634 xen_swiotlb_sync_sg(hwdev, sg, nelems, dir, SYNC_FOR_CPU); in xen_swiotlb_sync_sg_for_cpu() 639 int nelems, enum dma_data_direction dir) in xen_swiotlb_sync_sg_for_device() argument [all …]
|
/linux-4.19.296/include/uapi/linux/ |
D | btf.h | 99 __u32 nelems; member
|
/linux-4.19.296/lib/ |
D | rhashtable.c | 383 unsigned int nelems = atomic_read(&ht->nelems); in rhashtable_shrink() local 386 if (nelems) in rhashtable_shrink() 387 size = roundup_pow_of_two(nelems * 3 / 2); in rhashtable_shrink() 572 atomic_inc(&ht->nelems); in rhashtable_insert_one() 1075 atomic_set(&ht->nelems, 0); in rhashtable_init()
|
D | test_rhashtable.c | 210 total, atomic_read(&ht->nelems), entries, chain_len); in test_bucket_stats() 212 if (total != atomic_read(&ht->nelems) || total != entries) in test_bucket_stats()
|
/linux-4.19.296/include/linux/mtd/ |
D | rawnand.h | 1118 unsigned int nelems; member 1126 .nelems = sizeof((struct nand_op_parser_pattern_elem[]) { __VA_ARGS__ }) / \
|
/linux-4.19.296/include/net/netfilter/ |
D | nf_tables.h | 417 atomic_t nelems; member
|