/linux-4.19.296/drivers/i2c/busses/ |
D | i2c-stm32.c | 18 struct stm32_i2c_dma *dma; in stm32_i2c_dma_request() local 22 dma = devm_kzalloc(dev, sizeof(*dma), GFP_KERNEL); in stm32_i2c_dma_request() 23 if (!dma) in stm32_i2c_dma_request() 27 dma->chan_tx = dma_request_chan(dev, "tx"); in stm32_i2c_dma_request() 28 if (IS_ERR(dma->chan_tx)) { in stm32_i2c_dma_request() 30 ret = PTR_ERR(dma->chan_tx); in stm32_i2c_dma_request() 39 ret = dmaengine_slave_config(dma->chan_tx, &dma_sconfig); in stm32_i2c_dma_request() 46 dma->chan_rx = dma_request_chan(dev, "rx"); in stm32_i2c_dma_request() 47 if (IS_ERR(dma->chan_rx)) { in stm32_i2c_dma_request() 49 ret = PTR_ERR(dma->chan_rx); in stm32_i2c_dma_request() [all …]
|
D | i2c-imx.c | 204 struct imx_i2c_dma *dma; member 279 struct imx_i2c_dma *dma; in i2c_imx_dma_request() local 284 dma = devm_kzalloc(dev, sizeof(*dma), GFP_KERNEL); in i2c_imx_dma_request() 285 if (!dma) in i2c_imx_dma_request() 288 dma->chan_tx = dma_request_slave_channel(dev, "tx"); in i2c_imx_dma_request() 289 if (!dma->chan_tx) { in i2c_imx_dma_request() 299 ret = dmaengine_slave_config(dma->chan_tx, &dma_sconfig); in i2c_imx_dma_request() 305 dma->chan_rx = dma_request_slave_channel(dev, "rx"); in i2c_imx_dma_request() 306 if (!dma->chan_rx) { in i2c_imx_dma_request() 316 ret = dmaengine_slave_config(dma->chan_rx, &dma_sconfig); in i2c_imx_dma_request() [all …]
|
D | i2c-at91.c | 148 struct at91_twi_dma dma; member 244 struct at91_twi_dma *dma = &dev->dma; in at91_twi_dma_cleanup() local 248 if (dma->xfer_in_progress) { in at91_twi_dma_cleanup() 249 if (dma->direction == DMA_FROM_DEVICE) in at91_twi_dma_cleanup() 250 dmaengine_terminate_all(dma->chan_rx); in at91_twi_dma_cleanup() 252 dmaengine_terminate_all(dma->chan_tx); in at91_twi_dma_cleanup() 253 dma->xfer_in_progress = false; in at91_twi_dma_cleanup() 255 if (dma->buf_mapped) { in at91_twi_dma_cleanup() 256 dma_unmap_single(dev->dev, sg_dma_address(&dma->sg[0]), in at91_twi_dma_cleanup() 257 dev->buf_len, dma->direction); in at91_twi_dma_cleanup() [all …]
|
D | i2c-stm32f7.c | 298 struct stm32_i2c_dma *dma; member 622 struct stm32_i2c_dma *dma = i2c_dev->dma; in stm32f7_i2c_dma_callback() local 623 struct device *dev = dma->chan_using->device->dev; in stm32f7_i2c_dma_callback() 626 dma_unmap_single(dev, dma->dma_buf, dma->dma_len, dma->dma_data_dir); in stm32f7_i2c_dma_callback() 627 complete(&dma->dma_complete); in stm32f7_i2c_dma_callback() 821 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN) { in stm32f7_i2c_xfer_msg() 822 ret = stm32_i2c_prep_dma_xfer(i2c_dev->dev, i2c_dev->dma, in stm32f7_i2c_xfer_msg() 984 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN) { in stm32f7_i2c_smbus_xfer_msg() 985 ret = stm32_i2c_prep_dma_xfer(i2c_dev->dev, i2c_dev->dma, in stm32f7_i2c_smbus_xfer_msg() 1074 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN && in stm32f7_i2c_smbus_rep_start() [all …]
|
D | i2c-stm32.h | 51 void stm32_i2c_dma_free(struct stm32_i2c_dma *dma); 53 int stm32_i2c_prep_dma_xfer(struct device *dev, struct stm32_i2c_dma *dma,
|
D | i2c-qup.c | 224 struct dma_chan *dma; member 618 if (qup->btx.dma) in qup_i2c_rel_dma() 619 dma_release_channel(qup->btx.dma); in qup_i2c_rel_dma() 620 if (qup->brx.dma) in qup_i2c_rel_dma() 621 dma_release_channel(qup->brx.dma); in qup_i2c_rel_dma() 622 qup->btx.dma = NULL; in qup_i2c_rel_dma() 623 qup->brx.dma = NULL; in qup_i2c_rel_dma() 630 if (!qup->btx.dma) { in qup_i2c_req_dma() 631 qup->btx.dma = dma_request_slave_channel_reason(qup->dev, "tx"); in qup_i2c_req_dma() 632 if (IS_ERR(qup->btx.dma)) { in qup_i2c_req_dma() [all …]
|
/linux-4.19.296/drivers/media/v4l2-core/ |
D | videobuf-dma-sg.c | 144 return &mem->dma; in videobuf_to_dma() 148 static void videobuf_dma_init(struct videobuf_dmabuf *dma) in videobuf_dma_init() argument 150 memset(dma, 0, sizeof(*dma)); in videobuf_dma_init() 151 dma->magic = MAGIC_DMABUF; in videobuf_dma_init() 154 static int videobuf_dma_init_user_locked(struct videobuf_dmabuf *dma, in videobuf_dma_init_user_locked() argument 161 dma->direction = direction; in videobuf_dma_init_user_locked() 162 switch (dma->direction) { in videobuf_dma_init_user_locked() 175 dma->offset = data & ~PAGE_MASK; in videobuf_dma_init_user_locked() 176 dma->size = size; in videobuf_dma_init_user_locked() 177 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked() [all …]
|
/linux-4.19.296/drivers/thunderbolt/ |
D | dma_port.c | 206 struct tb_dma_port *dma; in dma_port_alloc() local 213 dma = kzalloc(sizeof(*dma), GFP_KERNEL); in dma_port_alloc() 214 if (!dma) in dma_port_alloc() 217 dma->buf = kmalloc_array(MAIL_DATA_DWORDS, sizeof(u32), GFP_KERNEL); in dma_port_alloc() 218 if (!dma->buf) { in dma_port_alloc() 219 kfree(dma); in dma_port_alloc() 223 dma->sw = sw; in dma_port_alloc() 224 dma->port = port; in dma_port_alloc() 225 dma->base = DMA_PORT_CAP; in dma_port_alloc() 227 return dma; in dma_port_alloc() [all …]
|
D | dma_port.h | 25 void dma_port_free(struct tb_dma_port *dma); 26 int dma_port_flash_read(struct tb_dma_port *dma, unsigned int address, 28 int dma_port_flash_update_auth(struct tb_dma_port *dma); 29 int dma_port_flash_update_auth_status(struct tb_dma_port *dma, u32 *status); 30 int dma_port_flash_write(struct tb_dma_port *dma, unsigned int address, 32 int dma_port_power_cycle(struct tb_dma_port *dma);
|
/linux-4.19.296/drivers/iio/adc/ |
D | ti_am335x_adc.c | 51 struct tiadc_dma dma; member 234 struct tiadc_dma *dma = &adc_dev->dma; in tiadc_dma_rx_complete() local 238 data = dma->buf + dma->current_period * dma->period_size; in tiadc_dma_rx_complete() 239 dma->current_period = 1 - dma->current_period; /* swap the buffer ID */ in tiadc_dma_rx_complete() 241 for (i = 0; i < dma->period_size; i += indio_dev->scan_bytes) { in tiadc_dma_rx_complete() 250 struct tiadc_dma *dma = &adc_dev->dma; in tiadc_start_dma() local 253 dma->current_period = 0; /* We start to fill period 0 */ in tiadc_start_dma() 261 dma->fifo_thresh = rounddown(FIFO1_THRESHOLD + 1, in tiadc_start_dma() 264 dma->period_size = rounddown(DMA_BUFFER_SIZE / 2, in tiadc_start_dma() 265 (dma->fifo_thresh + 1) * sizeof(u16)); in tiadc_start_dma() [all …]
|
/linux-4.19.296/drivers/soundwire/ |
D | intel.c | 466 static void intel_port_cleanup(struct sdw_cdns_dma_data *dma) in intel_port_cleanup() argument 470 for (i = 0; i < dma->nr_ports; i++) { in intel_port_cleanup() 471 if (dma->port[i]) { in intel_port_cleanup() 472 dma->port[i]->pdi->assigned = false; in intel_port_cleanup() 473 dma->port[i]->pdi = NULL; in intel_port_cleanup() 474 dma->port[i]->assigned = false; in intel_port_cleanup() 475 dma->port[i] = NULL; in intel_port_cleanup() 486 struct sdw_cdns_dma_data *dma; in intel_hw_params() local 492 dma = snd_soc_dai_get_dma_data(dai, substream); in intel_hw_params() 493 if (!dma) in intel_hw_params() [all …]
|
D | cadence_master.c | 994 struct sdw_cdns_dma_data *dma; in cdns_set_sdw_stream() local 996 dma = kzalloc(sizeof(*dma), GFP_KERNEL); in cdns_set_sdw_stream() 997 if (!dma) in cdns_set_sdw_stream() 1001 dma->stream_type = SDW_STREAM_PCM; in cdns_set_sdw_stream() 1003 dma->stream_type = SDW_STREAM_PDM; in cdns_set_sdw_stream() 1005 dma->bus = &cdns->bus; in cdns_set_sdw_stream() 1006 dma->link_id = cdns->instance; in cdns_set_sdw_stream() 1008 dma->stream = stream; in cdns_set_sdw_stream() 1011 dai->playback_dma_data = dma; in cdns_set_sdw_stream() 1013 dai->capture_dma_data = dma; in cdns_set_sdw_stream() [all …]
|
/linux-4.19.296/include/linux/soc/ti/ |
D | knav_qmss.h | 75 int knav_queue_push(void *qhandle, dma_addr_t dma, 85 dma_addr_t *dma, unsigned *dma_sz); 86 void *knav_pool_desc_unmap(void *ph, dma_addr_t dma, unsigned dma_sz); 88 void *knav_pool_desc_dma_to_virt(void *ph, dma_addr_t dma);
|
/linux-4.19.296/drivers/parport/ |
D | parport_pc.c | 100 int dma; member 619 disable_dma(port->dma); in parport_pc_fifo_write_block_dma() 620 clear_dma_ff(port->dma); in parport_pc_fifo_write_block_dma() 621 set_dma_mode(port->dma, DMA_MODE_WRITE); in parport_pc_fifo_write_block_dma() 622 set_dma_addr(port->dma, dma_addr); in parport_pc_fifo_write_block_dma() 623 set_dma_count(port->dma, count); in parport_pc_fifo_write_block_dma() 631 enable_dma(port->dma); in parport_pc_fifo_write_block_dma() 659 disable_dma(port->dma); in parport_pc_fifo_write_block_dma() 660 clear_dma_ff(port->dma); in parport_pc_fifo_write_block_dma() 661 count = get_dma_residue(port->dma); in parport_pc_fifo_write_block_dma() [all …]
|
D | parport_gsc.c | 237 int dma, struct parisc_device *padev) in parport_gsc_probe_port() argument 264 p->dma = dma; in parport_gsc_probe_port() 298 if (p->dma == PARPORT_DMA_AUTO) { in parport_gsc_probe_port() 299 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port() 302 if (p->dma == PARPORT_DMA_AUTO) /* To use DMA, giving the irq in parport_gsc_probe_port() 304 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port() 327 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port() 391 if (p->dma != PARPORT_DMA_NONE) in parport_remove_chip() 392 free_dma(p->dma); in parport_remove_chip()
|
/linux-4.19.296/drivers/isdn/hardware/avm/ |
D | b1dma.c | 356 skb_queue_tail(&card->dma->send_queue, skb); in b1dma_queue_tx() 370 avmcard_dmainfo *dma = card->dma; in b1dma_dispatch_tx() local 377 skb = skb_dequeue(&dma->send_queue); in b1dma_dispatch_tx() 385 p = dma->sendbuf.dmabuf; in b1dma_dispatch_tx() 396 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf; in b1dma_dispatch_tx() 410 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf, in b1dma_dispatch_tx() 415 b1dma_writel(card, dma->sendbuf.dmaaddr, AMCC_TXPTR); in b1dma_dispatch_tx() 450 avmcard_dmainfo *dma = card->dma; in b1dma_handle_rx() local 453 void *p = dma->recvbuf.dmabuf + 4; in b1dma_handle_rx() 458 printk(KERN_DEBUG "rx: 0x%x %lu\n", b1cmd, (unsigned long)dma->recvlen); in b1dma_handle_rx() [all …]
|
D | c4.c | 412 avmcard_dmainfo *dma = card->dma; in c4_dispatch_tx() local 424 skb = skb_dequeue(&dma->send_queue); in c4_dispatch_tx() 438 p = dma->sendbuf.dmabuf; in c4_dispatch_tx() 449 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf; in c4_dispatch_tx() 463 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf, in c4_dispatch_tx() 468 c4outmeml(card->mbase + MBOX_DOWN_ADDR, dma->sendbuf.dmaaddr); in c4_dispatch_tx() 497 skb_queue_tail(&card->dma->send_queue, skb); in queue_pollack() 505 avmcard_dmainfo *dma = card->dma; in c4_handle_rx() local 509 void *p = dma->recvbuf.dmabuf; in c4_handle_rx() 517 b1cmd, (unsigned long)dma->recvlen); in c4_handle_rx() [all …]
|
D | t1pci.c | 62 card->dma = avmcard_dma_alloc("t1pci", pdev, 2048 + 128, 2048 + 128); in t1pci_add_card() 63 if (!card->dma) { in t1pci_add_card() 146 avmcard_dma_free(card->dma); in t1pci_add_card() 166 avmcard_dma_free(card->dma); in t1pci_remove()
|
/linux-4.19.296/drivers/dma-buf/ |
D | Makefile | 1 obj-y := dma-buf.o dma-fence.o dma-fence-array.o reservation.o seqno-fence.o
|
/linux-4.19.296/include/media/ |
D | videobuf-dma-sg.h | 74 struct videobuf_dmabuf dma; member 87 int videobuf_dma_free(struct videobuf_dmabuf *dma); 89 int videobuf_dma_unmap(struct device *dev, struct videobuf_dmabuf *dma);
|
/linux-4.19.296/drivers/message/fusion/ |
D | mptlan.c | 84 dma_addr_t dma; member 522 pci_unmap_single(mpt_dev->pcidev, priv->RcvCtl[i].dma, in mpt_lan_close() 534 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[i].dma, in mpt_lan_close() 588 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[ctx].dma, in mpt_lan_send_turbo() 654 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[ctx].dma, in mpt_lan_send_reply() 684 dma_addr_t dma; in mpt_lan_sdu_send() local 726 dma = pci_map_single(mpt_dev->pcidev, skb->data, skb->len, in mpt_lan_sdu_send() 730 priv->SendCtl[ctx].dma = dma; in mpt_lan_sdu_send() 778 pSimple->Address.Low = cpu_to_le32((u32) dma); in mpt_lan_sdu_send() 780 pSimple->Address.High = cpu_to_le32((u32) ((u64) dma >> 32)); in mpt_lan_sdu_send() [all …]
|
/linux-4.19.296/drivers/media/common/videobuf2/ |
D | Makefile | 12 obj-$(CONFIG_VIDEOBUF2_DMA_CONTIG) += videobuf2-dma-contig.o 13 obj-$(CONFIG_VIDEOBUF2_DMA_SG) += videobuf2-dma-sg.o
|
/linux-4.19.296/crypto/async_tx/ |
D | async_raid6_recov.c | 37 struct dma_device *dma = chan ? chan->device : NULL; in async_sum_product() local 43 if (dma) in async_sum_product() 44 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product() 47 struct device *dev = dma->dev; in async_sum_product() 64 tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef, in async_sum_product() 102 struct dma_device *dma = chan ? chan->device : NULL; in async_mult() local 107 if (dma) in async_mult() 108 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_mult() 112 struct device *dev = dma->dev; in async_mult() 129 tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr, in async_mult()
|
/linux-4.19.296/include/linux/ |
D | async_tx.h | 105 struct dma_device *dma = chan->device; in async_tx_issue_pending() local 107 dma->device_issue_pending(chan); in async_tx_issue_pending() 155 dma_addr_t dma; member
|
D | dmaengine.h | 1081 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) in dma_set_maxpq() argument 1083 dma->max_pq = maxpq; in dma_set_maxpq() 1085 dma->max_pq |= DMA_HAS_PQ_CONTINUE; in dma_set_maxpq() 1100 static inline bool dma_dev_has_pq_continue(struct dma_device *dma) in dma_dev_has_pq_continue() argument 1102 return (dma->max_pq & DMA_HAS_PQ_CONTINUE) == DMA_HAS_PQ_CONTINUE; in dma_dev_has_pq_continue() 1105 static inline unsigned short dma_dev_to_maxpq(struct dma_device *dma) in dma_dev_to_maxpq() argument 1107 return dma->max_pq & ~DMA_HAS_PQ_CONTINUE; in dma_dev_to_maxpq() 1123 static inline int dma_maxpq(struct dma_device *dma, enum dma_ctrl_flags flags) in dma_maxpq() argument 1125 if (dma_dev_has_pq_continue(dma) || !dmaf_continue(flags)) in dma_maxpq() 1126 return dma_dev_to_maxpq(dma); in dma_maxpq() [all …]
|