Home
last modified time | relevance | path

Searched refs:dma (Results 1 – 25 of 99) sorted by relevance

1234

/linux-4.19.296/drivers/i2c/busses/
Di2c-stm32.c18 struct stm32_i2c_dma *dma; in stm32_i2c_dma_request() local
22 dma = devm_kzalloc(dev, sizeof(*dma), GFP_KERNEL); in stm32_i2c_dma_request()
23 if (!dma) in stm32_i2c_dma_request()
27 dma->chan_tx = dma_request_chan(dev, "tx"); in stm32_i2c_dma_request()
28 if (IS_ERR(dma->chan_tx)) { in stm32_i2c_dma_request()
30 ret = PTR_ERR(dma->chan_tx); in stm32_i2c_dma_request()
39 ret = dmaengine_slave_config(dma->chan_tx, &dma_sconfig); in stm32_i2c_dma_request()
46 dma->chan_rx = dma_request_chan(dev, "rx"); in stm32_i2c_dma_request()
47 if (IS_ERR(dma->chan_rx)) { in stm32_i2c_dma_request()
49 ret = PTR_ERR(dma->chan_rx); in stm32_i2c_dma_request()
[all …]
Di2c-imx.c204 struct imx_i2c_dma *dma; member
279 struct imx_i2c_dma *dma; in i2c_imx_dma_request() local
284 dma = devm_kzalloc(dev, sizeof(*dma), GFP_KERNEL); in i2c_imx_dma_request()
285 if (!dma) in i2c_imx_dma_request()
288 dma->chan_tx = dma_request_slave_channel(dev, "tx"); in i2c_imx_dma_request()
289 if (!dma->chan_tx) { in i2c_imx_dma_request()
299 ret = dmaengine_slave_config(dma->chan_tx, &dma_sconfig); in i2c_imx_dma_request()
305 dma->chan_rx = dma_request_slave_channel(dev, "rx"); in i2c_imx_dma_request()
306 if (!dma->chan_rx) { in i2c_imx_dma_request()
316 ret = dmaengine_slave_config(dma->chan_rx, &dma_sconfig); in i2c_imx_dma_request()
[all …]
Di2c-at91.c148 struct at91_twi_dma dma; member
244 struct at91_twi_dma *dma = &dev->dma; in at91_twi_dma_cleanup() local
248 if (dma->xfer_in_progress) { in at91_twi_dma_cleanup()
249 if (dma->direction == DMA_FROM_DEVICE) in at91_twi_dma_cleanup()
250 dmaengine_terminate_all(dma->chan_rx); in at91_twi_dma_cleanup()
252 dmaengine_terminate_all(dma->chan_tx); in at91_twi_dma_cleanup()
253 dma->xfer_in_progress = false; in at91_twi_dma_cleanup()
255 if (dma->buf_mapped) { in at91_twi_dma_cleanup()
256 dma_unmap_single(dev->dev, sg_dma_address(&dma->sg[0]), in at91_twi_dma_cleanup()
257 dev->buf_len, dma->direction); in at91_twi_dma_cleanup()
[all …]
Di2c-stm32f7.c298 struct stm32_i2c_dma *dma; member
622 struct stm32_i2c_dma *dma = i2c_dev->dma; in stm32f7_i2c_dma_callback() local
623 struct device *dev = dma->chan_using->device->dev; in stm32f7_i2c_dma_callback()
626 dma_unmap_single(dev, dma->dma_buf, dma->dma_len, dma->dma_data_dir); in stm32f7_i2c_dma_callback()
627 complete(&dma->dma_complete); in stm32f7_i2c_dma_callback()
821 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN) { in stm32f7_i2c_xfer_msg()
822 ret = stm32_i2c_prep_dma_xfer(i2c_dev->dev, i2c_dev->dma, in stm32f7_i2c_xfer_msg()
984 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN) { in stm32f7_i2c_smbus_xfer_msg()
985 ret = stm32_i2c_prep_dma_xfer(i2c_dev->dev, i2c_dev->dma, in stm32f7_i2c_smbus_xfer_msg()
1074 if (i2c_dev->dma && f7_msg->count >= STM32F7_I2C_DMA_LEN_MIN && in stm32f7_i2c_smbus_rep_start()
[all …]
Di2c-stm32.h51 void stm32_i2c_dma_free(struct stm32_i2c_dma *dma);
53 int stm32_i2c_prep_dma_xfer(struct device *dev, struct stm32_i2c_dma *dma,
Di2c-qup.c224 struct dma_chan *dma; member
618 if (qup->btx.dma) in qup_i2c_rel_dma()
619 dma_release_channel(qup->btx.dma); in qup_i2c_rel_dma()
620 if (qup->brx.dma) in qup_i2c_rel_dma()
621 dma_release_channel(qup->brx.dma); in qup_i2c_rel_dma()
622 qup->btx.dma = NULL; in qup_i2c_rel_dma()
623 qup->brx.dma = NULL; in qup_i2c_rel_dma()
630 if (!qup->btx.dma) { in qup_i2c_req_dma()
631 qup->btx.dma = dma_request_slave_channel_reason(qup->dev, "tx"); in qup_i2c_req_dma()
632 if (IS_ERR(qup->btx.dma)) { in qup_i2c_req_dma()
[all …]
/linux-4.19.296/drivers/media/v4l2-core/
Dvideobuf-dma-sg.c144 return &mem->dma; in videobuf_to_dma()
148 static void videobuf_dma_init(struct videobuf_dmabuf *dma) in videobuf_dma_init() argument
150 memset(dma, 0, sizeof(*dma)); in videobuf_dma_init()
151 dma->magic = MAGIC_DMABUF; in videobuf_dma_init()
154 static int videobuf_dma_init_user_locked(struct videobuf_dmabuf *dma, in videobuf_dma_init_user_locked() argument
161 dma->direction = direction; in videobuf_dma_init_user_locked()
162 switch (dma->direction) { in videobuf_dma_init_user_locked()
175 dma->offset = data & ~PAGE_MASK; in videobuf_dma_init_user_locked()
176 dma->size = size; in videobuf_dma_init_user_locked()
177 dma->nr_pages = last-first+1; in videobuf_dma_init_user_locked()
[all …]
/linux-4.19.296/drivers/thunderbolt/
Ddma_port.c206 struct tb_dma_port *dma; in dma_port_alloc() local
213 dma = kzalloc(sizeof(*dma), GFP_KERNEL); in dma_port_alloc()
214 if (!dma) in dma_port_alloc()
217 dma->buf = kmalloc_array(MAIL_DATA_DWORDS, sizeof(u32), GFP_KERNEL); in dma_port_alloc()
218 if (!dma->buf) { in dma_port_alloc()
219 kfree(dma); in dma_port_alloc()
223 dma->sw = sw; in dma_port_alloc()
224 dma->port = port; in dma_port_alloc()
225 dma->base = DMA_PORT_CAP; in dma_port_alloc()
227 return dma; in dma_port_alloc()
[all …]
Ddma_port.h25 void dma_port_free(struct tb_dma_port *dma);
26 int dma_port_flash_read(struct tb_dma_port *dma, unsigned int address,
28 int dma_port_flash_update_auth(struct tb_dma_port *dma);
29 int dma_port_flash_update_auth_status(struct tb_dma_port *dma, u32 *status);
30 int dma_port_flash_write(struct tb_dma_port *dma, unsigned int address,
32 int dma_port_power_cycle(struct tb_dma_port *dma);
/linux-4.19.296/drivers/iio/adc/
Dti_am335x_adc.c51 struct tiadc_dma dma; member
234 struct tiadc_dma *dma = &adc_dev->dma; in tiadc_dma_rx_complete() local
238 data = dma->buf + dma->current_period * dma->period_size; in tiadc_dma_rx_complete()
239 dma->current_period = 1 - dma->current_period; /* swap the buffer ID */ in tiadc_dma_rx_complete()
241 for (i = 0; i < dma->period_size; i += indio_dev->scan_bytes) { in tiadc_dma_rx_complete()
250 struct tiadc_dma *dma = &adc_dev->dma; in tiadc_start_dma() local
253 dma->current_period = 0; /* We start to fill period 0 */ in tiadc_start_dma()
261 dma->fifo_thresh = rounddown(FIFO1_THRESHOLD + 1, in tiadc_start_dma()
264 dma->period_size = rounddown(DMA_BUFFER_SIZE / 2, in tiadc_start_dma()
265 (dma->fifo_thresh + 1) * sizeof(u16)); in tiadc_start_dma()
[all …]
/linux-4.19.296/drivers/soundwire/
Dintel.c466 static void intel_port_cleanup(struct sdw_cdns_dma_data *dma) in intel_port_cleanup() argument
470 for (i = 0; i < dma->nr_ports; i++) { in intel_port_cleanup()
471 if (dma->port[i]) { in intel_port_cleanup()
472 dma->port[i]->pdi->assigned = false; in intel_port_cleanup()
473 dma->port[i]->pdi = NULL; in intel_port_cleanup()
474 dma->port[i]->assigned = false; in intel_port_cleanup()
475 dma->port[i] = NULL; in intel_port_cleanup()
486 struct sdw_cdns_dma_data *dma; in intel_hw_params() local
492 dma = snd_soc_dai_get_dma_data(dai, substream); in intel_hw_params()
493 if (!dma) in intel_hw_params()
[all …]
Dcadence_master.c994 struct sdw_cdns_dma_data *dma; in cdns_set_sdw_stream() local
996 dma = kzalloc(sizeof(*dma), GFP_KERNEL); in cdns_set_sdw_stream()
997 if (!dma) in cdns_set_sdw_stream()
1001 dma->stream_type = SDW_STREAM_PCM; in cdns_set_sdw_stream()
1003 dma->stream_type = SDW_STREAM_PDM; in cdns_set_sdw_stream()
1005 dma->bus = &cdns->bus; in cdns_set_sdw_stream()
1006 dma->link_id = cdns->instance; in cdns_set_sdw_stream()
1008 dma->stream = stream; in cdns_set_sdw_stream()
1011 dai->playback_dma_data = dma; in cdns_set_sdw_stream()
1013 dai->capture_dma_data = dma; in cdns_set_sdw_stream()
[all …]
/linux-4.19.296/include/linux/soc/ti/
Dknav_qmss.h75 int knav_queue_push(void *qhandle, dma_addr_t dma,
85 dma_addr_t *dma, unsigned *dma_sz);
86 void *knav_pool_desc_unmap(void *ph, dma_addr_t dma, unsigned dma_sz);
88 void *knav_pool_desc_dma_to_virt(void *ph, dma_addr_t dma);
/linux-4.19.296/drivers/parport/
Dparport_pc.c100 int dma; member
619 disable_dma(port->dma); in parport_pc_fifo_write_block_dma()
620 clear_dma_ff(port->dma); in parport_pc_fifo_write_block_dma()
621 set_dma_mode(port->dma, DMA_MODE_WRITE); in parport_pc_fifo_write_block_dma()
622 set_dma_addr(port->dma, dma_addr); in parport_pc_fifo_write_block_dma()
623 set_dma_count(port->dma, count); in parport_pc_fifo_write_block_dma()
631 enable_dma(port->dma); in parport_pc_fifo_write_block_dma()
659 disable_dma(port->dma); in parport_pc_fifo_write_block_dma()
660 clear_dma_ff(port->dma); in parport_pc_fifo_write_block_dma()
661 count = get_dma_residue(port->dma); in parport_pc_fifo_write_block_dma()
[all …]
Dparport_gsc.c237 int dma, struct parisc_device *padev) in parport_gsc_probe_port() argument
264 p->dma = dma; in parport_gsc_probe_port()
298 if (p->dma == PARPORT_DMA_AUTO) { in parport_gsc_probe_port()
299 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port()
302 if (p->dma == PARPORT_DMA_AUTO) /* To use DMA, giving the irq in parport_gsc_probe_port()
304 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port()
327 p->dma = PARPORT_DMA_NONE; in parport_gsc_probe_port()
391 if (p->dma != PARPORT_DMA_NONE) in parport_remove_chip()
392 free_dma(p->dma); in parport_remove_chip()
/linux-4.19.296/drivers/isdn/hardware/avm/
Db1dma.c356 skb_queue_tail(&card->dma->send_queue, skb); in b1dma_queue_tx()
370 avmcard_dmainfo *dma = card->dma; in b1dma_dispatch_tx() local
377 skb = skb_dequeue(&dma->send_queue); in b1dma_dispatch_tx()
385 p = dma->sendbuf.dmabuf; in b1dma_dispatch_tx()
396 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf; in b1dma_dispatch_tx()
410 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf, in b1dma_dispatch_tx()
415 b1dma_writel(card, dma->sendbuf.dmaaddr, AMCC_TXPTR); in b1dma_dispatch_tx()
450 avmcard_dmainfo *dma = card->dma; in b1dma_handle_rx() local
453 void *p = dma->recvbuf.dmabuf + 4; in b1dma_handle_rx()
458 printk(KERN_DEBUG "rx: 0x%x %lu\n", b1cmd, (unsigned long)dma->recvlen); in b1dma_handle_rx()
[all …]
Dc4.c412 avmcard_dmainfo *dma = card->dma; in c4_dispatch_tx() local
424 skb = skb_dequeue(&dma->send_queue); in c4_dispatch_tx()
438 p = dma->sendbuf.dmabuf; in c4_dispatch_tx()
449 txlen = (u8 *)p - (u8 *)dma->sendbuf.dmabuf; in c4_dispatch_tx()
463 skb_copy_from_linear_data_offset(skb, 2, dma->sendbuf.dmabuf, in c4_dispatch_tx()
468 c4outmeml(card->mbase + MBOX_DOWN_ADDR, dma->sendbuf.dmaaddr); in c4_dispatch_tx()
497 skb_queue_tail(&card->dma->send_queue, skb); in queue_pollack()
505 avmcard_dmainfo *dma = card->dma; in c4_handle_rx() local
509 void *p = dma->recvbuf.dmabuf; in c4_handle_rx()
517 b1cmd, (unsigned long)dma->recvlen); in c4_handle_rx()
[all …]
Dt1pci.c62 card->dma = avmcard_dma_alloc("t1pci", pdev, 2048 + 128, 2048 + 128); in t1pci_add_card()
63 if (!card->dma) { in t1pci_add_card()
146 avmcard_dma_free(card->dma); in t1pci_add_card()
166 avmcard_dma_free(card->dma); in t1pci_remove()
/linux-4.19.296/drivers/dma-buf/
DMakefile1 obj-y := dma-buf.o dma-fence.o dma-fence-array.o reservation.o seqno-fence.o
/linux-4.19.296/include/media/
Dvideobuf-dma-sg.h74 struct videobuf_dmabuf dma; member
87 int videobuf_dma_free(struct videobuf_dmabuf *dma);
89 int videobuf_dma_unmap(struct device *dev, struct videobuf_dmabuf *dma);
/linux-4.19.296/drivers/message/fusion/
Dmptlan.c84 dma_addr_t dma; member
522 pci_unmap_single(mpt_dev->pcidev, priv->RcvCtl[i].dma, in mpt_lan_close()
534 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[i].dma, in mpt_lan_close()
588 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[ctx].dma, in mpt_lan_send_turbo()
654 pci_unmap_single(mpt_dev->pcidev, priv->SendCtl[ctx].dma, in mpt_lan_send_reply()
684 dma_addr_t dma; in mpt_lan_sdu_send() local
726 dma = pci_map_single(mpt_dev->pcidev, skb->data, skb->len, in mpt_lan_sdu_send()
730 priv->SendCtl[ctx].dma = dma; in mpt_lan_sdu_send()
778 pSimple->Address.Low = cpu_to_le32((u32) dma); in mpt_lan_sdu_send()
780 pSimple->Address.High = cpu_to_le32((u32) ((u64) dma >> 32)); in mpt_lan_sdu_send()
[all …]
/linux-4.19.296/drivers/media/common/videobuf2/
DMakefile12 obj-$(CONFIG_VIDEOBUF2_DMA_CONTIG) += videobuf2-dma-contig.o
13 obj-$(CONFIG_VIDEOBUF2_DMA_SG) += videobuf2-dma-sg.o
/linux-4.19.296/crypto/async_tx/
Dasync_raid6_recov.c37 struct dma_device *dma = chan ? chan->device : NULL; in async_sum_product() local
43 if (dma) in async_sum_product()
44 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product()
47 struct device *dev = dma->dev; in async_sum_product()
64 tx = dma->device_prep_dma_pq(chan, pq, unmap->addr, 2, coef, in async_sum_product()
102 struct dma_device *dma = chan ? chan->device : NULL; in async_mult() local
107 if (dma) in async_mult()
108 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_mult()
112 struct device *dev = dma->dev; in async_mult()
129 tx = dma->device_prep_dma_pq(chan, dma_dest, unmap->addr, in async_mult()
/linux-4.19.296/include/linux/
Dasync_tx.h105 struct dma_device *dma = chan->device; in async_tx_issue_pending() local
107 dma->device_issue_pending(chan); in async_tx_issue_pending()
155 dma_addr_t dma; member
Ddmaengine.h1081 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) in dma_set_maxpq() argument
1083 dma->max_pq = maxpq; in dma_set_maxpq()
1085 dma->max_pq |= DMA_HAS_PQ_CONTINUE; in dma_set_maxpq()
1100 static inline bool dma_dev_has_pq_continue(struct dma_device *dma) in dma_dev_has_pq_continue() argument
1102 return (dma->max_pq & DMA_HAS_PQ_CONTINUE) == DMA_HAS_PQ_CONTINUE; in dma_dev_has_pq_continue()
1105 static inline unsigned short dma_dev_to_maxpq(struct dma_device *dma) in dma_dev_to_maxpq() argument
1107 return dma->max_pq & ~DMA_HAS_PQ_CONTINUE; in dma_dev_to_maxpq()
1123 static inline int dma_maxpq(struct dma_device *dma, enum dma_ctrl_flags flags) in dma_maxpq() argument
1125 if (dma_dev_has_pq_continue(dma) || !dmaf_continue(flags)) in dma_maxpq()
1126 return dma_dev_to_maxpq(dma); in dma_maxpq()
[all …]

1234