/linux-4.19.296/drivers/iio/potentiometer/ |
D | mcp4531.c | 41 int avail[3]; member 81 [MCP453x_502] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 5, }, 82 [MCP453x_103] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 10, }, 83 [MCP453x_503] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 50, }, 84 [MCP453x_104] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 100, }, 85 [MCP454x_502] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 5, }, 86 [MCP454x_103] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 10, }, 87 [MCP454x_503] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 50, }, 88 [MCP454x_104] = { .wipers = 1, .avail = { 0, 1, 128 }, .kohms = 100, }, 89 [MCP455x_502] = { .wipers = 1, .avail = { 0, 1, 256 }, .kohms = 5, }, [all …]
|
/linux-4.19.296/drivers/xen/xenbus/ |
D | xenbus_comms.c | 114 unsigned int avail; in xb_write() local 129 dst = get_output_chunk(cons, prod, intf->req, &avail); in xb_write() 130 if (avail == 0) in xb_write() 132 if (avail > len) in xb_write() 133 avail = len; in xb_write() 135 memcpy(dst, data, avail); in xb_write() 136 data += avail; in xb_write() 137 len -= avail; in xb_write() 138 bytes += avail; in xb_write() 142 intf->req_prod += avail; in xb_write() [all …]
|
/linux-4.19.296/drivers/s390/cio/ |
D | airq.c | 137 iv->avail = kmalloc(size, GFP_KERNEL); in airq_iv_create() 138 if (!iv->avail) in airq_iv_create() 140 memset(iv->avail, 0xff, size); in airq_iv_create() 167 kfree(iv->avail); in airq_iv_create() 185 kfree(iv->avail); in airq_iv_release() 203 if (!iv->avail || num == 0) in airq_iv_alloc() 206 bit = find_first_bit_inv(iv->avail, iv->bits); in airq_iv_alloc() 209 if (!test_bit_inv(bit + i, iv->avail)) in airq_iv_alloc() 214 clear_bit_inv(bit + i, iv->avail); in airq_iv_alloc() 219 bit = find_next_bit_inv(iv->avail, iv->bits, bit + i + 1); in airq_iv_alloc() [all …]
|
/linux-4.19.296/drivers/uwb/ |
D | drp-avail.c | 61 void uwb_drp_available(struct uwb_rc *rc, struct uwb_mas_bm *avail) in uwb_drp_available() argument 63 bitmap_and(avail->bm, rc->drp_avail.global, rc->drp_avail.local, UWB_NUM_MAS); in uwb_drp_available() 64 bitmap_and(avail->bm, avail->bm, rc->drp_avail.pending, UWB_NUM_MAS); in uwb_drp_available() 76 struct uwb_mas_bm avail; in uwb_drp_avail_reserve_pending() local 78 uwb_drp_available(rc, &avail); in uwb_drp_avail_reserve_pending() 79 if (!bitmap_subset(mas->bm, avail.bm, UWB_NUM_MAS)) in uwb_drp_avail_reserve_pending() 119 struct uwb_mas_bm avail; in uwb_drp_avail_ie_update() local 121 bitmap_and(avail.bm, rc->drp_avail.global, rc->drp_avail.local, UWB_NUM_MAS); in uwb_drp_avail_ie_update() 125 uwb_mas_bm_copy_le(rc->drp_avail.ie.bmp, &avail); in uwb_drp_avail_ie_update()
|
/linux-4.19.296/fs/squashfs/ |
D | lzo_wrapper.c | 84 int avail, i, bytes = length, res; in lzo_uncompress() local 88 avail = min(bytes, msblk->devblksize - offset); in lzo_uncompress() 89 memcpy(buff, bh[i]->b_data + offset, avail); in lzo_uncompress() 90 buff += avail; in lzo_uncompress() 91 bytes -= avail; in lzo_uncompress()
|
D | lz4_wrapper.c | 99 int avail, i, bytes = length, res; in lz4_uncompress() local 102 avail = min(bytes, msblk->devblksize - offset); in lz4_uncompress() 103 memcpy(buff, bh[i]->b_data + offset, avail); in lz4_uncompress() 104 buff += avail; in lz4_uncompress() 105 bytes -= avail; in lz4_uncompress()
|
D | block.c | 98 int bytes, compressed, b = 0, k = 0, avail, i; in squashfs_read_data() local 191 avail = min_t(int, in, PAGE_SIZE - in squashfs_read_data() 194 avail); in squashfs_read_data() 195 in -= avail; in squashfs_read_data() 196 pg_offset += avail; in squashfs_read_data() 197 offset += avail; in squashfs_read_data()
|
D | xz_wrapper.c | 137 int avail, total = 0, k = 0; in squashfs_xz_uncompress() local 149 avail = min(length, msblk->devblksize - offset); in squashfs_xz_uncompress() 150 length -= avail; in squashfs_xz_uncompress() 152 stream->buf.in_size = avail; in squashfs_xz_uncompress()
|
D | zlib_wrapper.c | 78 int avail = min(length, msblk->devblksize - offset); in zlib_uncompress() local 79 length -= avail; in zlib_uncompress() 81 stream->avail_in = avail; in zlib_uncompress()
|
D | zstd_wrapper.c | 94 int avail = min(length, msblk->devblksize - offset); in zstd_uncompress() local 96 length -= avail; in zstd_uncompress() 98 in_buf.size = avail; in zstd_uncompress()
|
D | file.c | 377 …d squashfs_fill_page(struct page *page, struct squashfs_cache_entry *buffer, int offset, int avail) in squashfs_fill_page() argument 383 copied = squashfs_copy_data(pageaddr, buffer, offset, avail); in squashfs_fill_page() 388 if (copied == avail) in squashfs_fill_page() 412 int avail = buffer ? min_t(int, bytes, PAGE_SIZE) : 0; in squashfs_copy_cache() local 414 TRACE("bytes %d, i %d, available_bytes %d\n", bytes, i, avail); in squashfs_copy_cache() 425 squashfs_fill_page(push_page, buffer, offset, avail); in squashfs_copy_cache()
|
/linux-4.19.296/include/uapi/linux/ |
D | virtio_ring.h | 102 struct vring_avail *avail; member 140 #define vring_used_event(vr) ((vr)->avail->ring[(vr)->num]) 148 vr->avail = p + num*sizeof(struct vring_desc); in vring_init() 149 vr->used = (void *)(((uintptr_t)&vr->avail->ring[num] + sizeof(__virtio16) in vring_init()
|
/linux-4.19.296/drivers/misc/mic/vop/ |
D | vop_debugfs.c | 141 struct vring_avail *avail; in vop_vdev_info_show() local 166 avail = vrh->vring.avail; in vop_vdev_info_show() 168 vringh16_to_cpu(vrh, avail->flags), in vop_vdev_info_show() 170 avail->idx) & (num - 1)); in vop_vdev_info_show() 172 vringh16_to_cpu(vrh, avail->flags), in vop_vdev_info_show() 173 vringh16_to_cpu(vrh, avail->idx)); in vop_vdev_info_show() 176 j, avail->ring[j]); in vop_vdev_info_show()
|
/linux-4.19.296/drivers/misc/ |
D | cs5535-mfgpt.c | 39 DECLARE_BITMAP(avail, MFGPT_MAX_TIMERS); 175 t = find_first_bit(mfgpt->avail, max); in cs5535_mfgpt_alloc_timer() 180 if (!test_bit(timer_nr, mfgpt->avail)) in cs5535_mfgpt_alloc_timer() 186 __clear_bit(timer_nr, mfgpt->avail); in cs5535_mfgpt_alloc_timer() 196 __set_bit(timer_nr, mfgpt->avail); in cs5535_mfgpt_alloc_timer() 223 __set_bit(timer->nr, timer->chip->avail); in cs5535_mfgpt_free_timer() 310 __set_bit(i, mfgpt->avail); in scan_timers()
|
/linux-4.19.296/fs/xfs/libxfs/ |
D | xfs_ag_resv.c | 79 xfs_extlen_t avail; in xfs_ag_resv_critical() local 84 avail = pag->pagf_freeblks - pag->pag_rmapbt_resv.ar_reserved; in xfs_ag_resv_critical() 88 avail = pag->pagf_freeblks + pag->pagf_flcount - in xfs_ag_resv_critical() 97 trace_xfs_ag_resv_critical(pag, type, avail); in xfs_ag_resv_critical() 100 return XFS_TEST_ERROR(avail < orig / 10 || avail < XFS_BTREE_MAXLEVELS, in xfs_ag_resv_critical()
|
/linux-4.19.296/crypto/ |
D | md4.c | 155 const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f); in md4_update() local 159 if (avail > len) { in md4_update() 160 memcpy((char *)mctx->block + (sizeof(mctx->block) - avail), in md4_update() 165 memcpy((char *)mctx->block + (sizeof(mctx->block) - avail), in md4_update() 166 data, avail); in md4_update() 169 data += avail; in md4_update() 170 len -= avail; in md4_update()
|
D | md5.c | 148 const u32 avail = sizeof(mctx->block) - (mctx->byte_count & 0x3f); in md5_update() local 152 if (avail > len) { in md5_update() 153 memcpy((char *)mctx->block + (sizeof(mctx->block) - avail), in md5_update() 158 memcpy((char *)mctx->block + (sizeof(mctx->block) - avail), in md5_update() 159 data, avail); in md5_update() 162 data += avail; in md5_update() 163 len -= avail; in md5_update()
|
/linux-4.19.296/drivers/pci/ |
D | bus.c | 165 struct resource *r, avail; in pci_bus_alloc_from_region() local 186 avail = *r; in pci_bus_alloc_from_region() 187 pci_clip_resource_to_region(bus, &avail, region); in pci_bus_alloc_from_region() 195 if (avail.start) in pci_bus_alloc_from_region() 196 min_used = avail.start; in pci_bus_alloc_from_region() 198 max = avail.end; in pci_bus_alloc_from_region()
|
/linux-4.19.296/lib/ |
D | genalloc.c | 199 atomic_long_set(&chunk->avail, size); in gen_pool_add_virt() 309 if (size > atomic_long_read(&chunk->avail)) in gen_pool_alloc_algo() 329 atomic_long_sub(size, &chunk->avail); in gen_pool_alloc_algo() 395 atomic_long_add(size, &chunk->avail); in gen_pool_free() 465 size_t avail = 0; in gen_pool_avail() local 469 avail += atomic_long_read(&chunk->avail); in gen_pool_avail() 471 return avail; in gen_pool_avail()
|
/linux-4.19.296/fs/ubifs/ |
D | tnc_commit.c | 439 int lnum, offs, len, next_len, buf_len, buf_offs, used, avail; in layout_in_empty_space() local 452 avail = buf_len; in layout_in_empty_space() 473 avail = buf_len; in layout_in_empty_space() 519 avail -= ALIGN(len, 8); in layout_in_empty_space() 523 avail > 0) in layout_in_empty_space() 526 if (avail <= 0 && next_len && in layout_in_empty_space() 546 avail = buf_len - used; in layout_in_empty_space() 834 int avail, wlen, err, lnum_pos = 0, blen, nxt_offs; in write_index() local 850 avail = buf_len; in write_index() 897 avail = buf_len; in write_index() [all …]
|
D | io.c | 524 ubifs_assert(c, !(wbuf->avail & 7)); in ubifs_wbuf_sync_nolock() 566 wbuf->avail = wbuf->size; in ubifs_wbuf_sync_nolock() 607 wbuf->avail = wbuf->size; in ubifs_wbuf_seek_nolock() 705 ubifs_assert(c, wbuf->avail > 0 && wbuf->avail <= wbuf->size); in ubifs_wbuf_write_nolock() 725 if (aligned_len <= wbuf->avail) { in ubifs_wbuf_write_nolock() 736 if (aligned_len == wbuf->avail) { in ubifs_wbuf_write_nolock() 750 wbuf->avail = wbuf->size; in ubifs_wbuf_write_nolock() 756 wbuf->avail -= aligned_len; in ubifs_wbuf_write_nolock() 774 memcpy(wbuf->buf + wbuf->used, buf, wbuf->avail); in ubifs_wbuf_write_nolock() 781 len -= wbuf->avail; in ubifs_wbuf_write_nolock() [all …]
|
D | orphan.c | 204 int avail_lebs, avail, gap; in avail_orphs() local 207 avail = avail_lebs * in avail_orphs() 211 avail += (gap - UBIFS_ORPH_NODE_SZ) / sizeof(__le64); in avail_orphs() 212 return avail; in avail_orphs() 224 int avail_lebs, avail; in tot_avail_orphs() local 227 avail = avail_lebs * in tot_avail_orphs() 229 return avail / 2; in tot_avail_orphs() 416 int avail, atomic = 0, err; in commit_orphans() local 419 avail = avail_orphs(c); in commit_orphans() 420 if (avail < c->cmt_orphans) { in commit_orphans()
|
/linux-4.19.296/virt/kvm/ |
D | coalesced_mmio.c | 46 unsigned avail; in coalesced_mmio_has_room() local 55 avail = (ring->first - last - 1) % KVM_COALESCED_MMIO_MAX; in coalesced_mmio_has_room() 56 if (avail == 0) { in coalesced_mmio_has_room()
|
/linux-4.19.296/drivers/media/dvb-core/ |
D | dvb_ringbuffer.c | 84 ssize_t avail; in dvb_ringbuffer_avail() local 90 avail = smp_load_acquire(&rbuf->pwrite) - rbuf->pread; in dvb_ringbuffer_avail() 91 if (avail < 0) in dvb_ringbuffer_avail() 92 avail += rbuf->size; in dvb_ringbuffer_avail() 93 return avail; in dvb_ringbuffer_avail()
|
/linux-4.19.296/fs/jffs2/ |
D | nodemgmt.c | 26 uint32_t avail; in jffs2_rp_can_write() local 29 avail = c->dirty_size + c->free_size + c->unchecked_size + in jffs2_rp_can_write() 33 if (avail < 2 * opts->rp_size) in jffs2_rp_can_write() 39 c->nr_erasing_blocks, avail, c->nospc_dirty_size); in jffs2_rp_can_write() 41 if (avail > opts->rp_size) in jffs2_rp_can_write() 101 uint32_t dirty, avail; in jffs2_reserve_space() local 140 avail = c->free_size + c->dirty_size + c->erasing_size + c->unchecked_size; in jffs2_reserve_space() 141 if ( (avail / c->sector_size) <= blocksneeded) { in jffs2_reserve_space() 149 avail, blocksneeded * c->sector_size); in jffs2_reserve_space()
|