| /Linux-v5.4/include/linux/ |
| D | dmaengine.h | 496 struct dma_async_tx_descriptor { struct 501 dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx); argument 502 int (*desc_free)(struct dma_async_tx_descriptor *tx); argument 508 struct dma_async_tx_descriptor *next; argument 509 struct dma_async_tx_descriptor *parent; argument 515 static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, in dma_set_unmap() argument 526 static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, in dma_set_unmap() 540 static inline void dma_descriptor_unmap(struct dma_async_tx_descriptor *tx) in dma_descriptor_unmap() 549 static inline void txd_lock(struct dma_async_tx_descriptor *txd) in txd_lock() 552 static inline void txd_unlock(struct dma_async_tx_descriptor *txd) in txd_unlock() [all …]
|
| D | async_tx.h | 71 struct dma_async_tx_descriptor *depend_tx; 88 static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx) in async_tx_issue_pending() 112 static inline void async_tx_issue_pending(struct dma_async_tx_descriptor *tx) in async_tx_issue_pending() 147 struct dma_async_tx_descriptor *tx, in init_async_submit() 158 void async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, 161 struct dma_async_tx_descriptor * 165 struct dma_async_tx_descriptor * 170 struct dma_async_tx_descriptor * 175 struct dma_async_tx_descriptor *async_trigger_callback(struct async_submit_ctl *submit); 177 struct dma_async_tx_descriptor * [all …]
|
| /Linux-v5.4/crypto/async_tx/ |
| D | async_tx.c | 46 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in __async_tx_find_channel() 65 async_tx_channel_switch(struct dma_async_tx_descriptor *depend_tx, in async_tx_channel_switch() 66 struct dma_async_tx_descriptor *tx) in async_tx_channel_switch() 70 struct dma_async_tx_descriptor *intr_tx = (void *) ~0; in async_tx_channel_switch() 143 async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, in async_tx_submit() 146 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_tx_submit() 220 struct dma_async_tx_descriptor * 225 struct dma_async_tx_descriptor *tx; in async_trigger_callback() 226 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_trigger_callback() 263 void async_tx_quiesce(struct dma_async_tx_descriptor **tx) in async_tx_quiesce()
|
| D | async_raid6_recov.c | 17 static struct dma_async_tx_descriptor * 35 struct dma_async_tx_descriptor *tx; in async_sum_product() 82 static struct dma_async_tx_descriptor * 99 struct dma_async_tx_descriptor *tx; in async_mult() 145 static struct dma_async_tx_descriptor * 149 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_4() 184 static struct dma_async_tx_descriptor * 188 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_5() 258 static struct dma_async_tx_descriptor * 262 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_n() [all …]
|
| D | async_xor.c | 22 static __async_inline struct dma_async_tx_descriptor * 27 struct dma_async_tx_descriptor *tx = NULL; in do_async_xor() 159 struct dma_async_tx_descriptor * 175 struct dma_async_tx_descriptor *tx; in async_xor() 255 struct dma_async_tx_descriptor * 262 struct dma_async_tx_descriptor *tx = NULL; in async_xor_val()
|
| D | async_pq.c | 34 static __async_inline struct dma_async_tx_descriptor * 41 struct dma_async_tx_descriptor *tx = NULL; in do_async_gen_syndrome() 162 struct dma_async_tx_descriptor * 183 struct dma_async_tx_descriptor *tx; in async_gen_syndrome() 280 struct dma_async_tx_descriptor * 287 struct dma_async_tx_descriptor *tx; in async_syndrome_val()
|
| D | async_memcpy.c | 31 struct dma_async_tx_descriptor * 39 struct dma_async_tx_descriptor *tx = NULL; in async_memcpy()
|
| /Linux-v5.4/drivers/dma/ioat/ |
| D | dma.h | 184 struct dma_async_tx_descriptor txd; 218 struct dma_async_tx_descriptor *tx, int id) in __dump_desc_dbg() 357 struct dma_async_tx_descriptor * 360 struct dma_async_tx_descriptor * 362 struct dma_async_tx_descriptor * 365 struct dma_async_tx_descriptor * 369 struct dma_async_tx_descriptor * 373 struct dma_async_tx_descriptor * 377 struct dma_async_tx_descriptor * 380 struct dma_async_tx_descriptor *
|
| D | prep.c | 100 struct dma_async_tx_descriptor * 151 static struct dma_async_tx_descriptor * 244 struct dma_async_tx_descriptor * 256 struct dma_async_tx_descriptor * 338 static struct dma_async_tx_descriptor * 462 static struct dma_async_tx_descriptor * 573 struct dma_async_tx_descriptor * 618 struct dma_async_tx_descriptor * 646 struct dma_async_tx_descriptor * 672 struct dma_async_tx_descriptor * [all …]
|
| /Linux-v5.4/drivers/dma/ |
| D | dmaengine.h | 29 static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx) in dma_cookie_assign() 52 static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx) in dma_cookie_complete() 106 dmaengine_desc_get_callback(struct dma_async_tx_descriptor *tx, in dmaengine_desc_get_callback() 152 dmaengine_desc_get_callback_invoke(struct dma_async_tx_descriptor *tx, in dmaengine_desc_get_callback_invoke()
|
| D | virt-dma.h | 16 struct dma_async_tx_descriptor tx; 47 extern dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *); 48 extern int vchan_tx_desc_free(struct dma_async_tx_descriptor *); 56 static inline struct dma_async_tx_descriptor *vchan_tx_prep(struct virt_dma_chan *vc, in vchan_tx_prep()
|
| D | virt-dma.c | 14 static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx) in to_virt_desc() 19 dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx) in vchan_tx_submit() 49 int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx) in vchan_tx_desc_free()
|
| D | mic_x100_dma.c | 80 struct dma_async_tx_descriptor *tx; in mic_dma_cleanup() 238 static dma_cookie_t mic_dma_tx_submit_unlock(struct dma_async_tx_descriptor *tx) in mic_dma_tx_submit_unlock() 257 static inline struct dma_async_tx_descriptor * 261 struct dma_async_tx_descriptor *tx = &ch->tx_array[idx]; in allocate_tx() 269 static struct dma_async_tx_descriptor * 300 static struct dma_async_tx_descriptor * 320 static struct dma_async_tx_descriptor * 499 struct dma_async_tx_descriptor *tx; in mic_dma_drain_chan()
|
| D | mv_xor_v2.c | 179 struct dma_async_tx_descriptor async_tx; 299 mv_xor_v2_tx_submit(struct dma_async_tx_descriptor *tx) in mv_xor_v2_tx_submit() 373 static struct dma_async_tx_descriptor * 426 static struct dma_async_tx_descriptor * 485 static struct dma_async_tx_descriptor *
|
| D | iop-adma.c | 55 struct dma_async_tx_descriptor *tx = &desc->async_tx; in iop_adma_run_tx_complete_actions() 361 iop_adma_tx_submit(struct dma_async_tx_descriptor *tx) in iop_adma_tx_submit() 483 static struct dma_async_tx_descriptor * 505 static struct dma_async_tx_descriptor * 536 static struct dma_async_tx_descriptor * 571 static struct dma_async_tx_descriptor * 606 static struct dma_async_tx_descriptor * 670 static struct dma_async_tx_descriptor * 838 struct dma_async_tx_descriptor *tx; in iop_adma_memcpy_self_test() 914 struct dma_async_tx_descriptor *tx; in iop_adma_xor_val_self_test() [all …]
|
| D | fsl_raid.c | 86 static dma_cookie_t fsl_re_tx_submit(struct dma_async_tx_descriptor *tx) in fsl_re_tx_submit() 316 static struct dma_async_tx_descriptor *fsl_re_prep_dma_genq( in fsl_re_prep_dma_genq() 391 static struct dma_async_tx_descriptor *fsl_re_prep_dma_xor( in fsl_re_prep_dma_xor() 403 static struct dma_async_tx_descriptor *fsl_re_prep_dma_pq( in fsl_re_prep_dma_pq() 430 struct dma_async_tx_descriptor *tx; in fsl_re_prep_dma_pq() 526 static struct dma_async_tx_descriptor *fsl_re_prep_dma_memcpy( in fsl_re_prep_dma_memcpy()
|
| D | imx-dma.c | 122 struct dma_async_tx_descriptor desc; 156 struct dma_async_tx_descriptor desc; 744 static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) in imxdma_tx_submit() 773 memset(&desc->desc, 0, sizeof(struct dma_async_tx_descriptor)); in imxdma_alloc_chan_resources() 815 static struct dma_async_tx_descriptor *imxdma_prep_slave_sg( in imxdma_prep_slave_sg() 866 static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic( in imxdma_prep_dma_cyclic() 924 static struct dma_async_tx_descriptor *imxdma_prep_dma_memcpy( in imxdma_prep_dma_memcpy() 955 static struct dma_async_tx_descriptor *imxdma_prep_dma_interleaved( in imxdma_prep_dma_interleaved()
|
| D | mmp_pdma.c | 85 struct dma_async_tx_descriptor async_tx; 93 struct dma_async_tx_descriptor desc; 341 static dma_cookie_t mmp_pdma_tx_submit(struct dma_async_tx_descriptor *tx) in mmp_pdma_tx_submit() 444 static struct dma_async_tx_descriptor * 525 static struct dma_async_tx_descriptor * 605 static struct dma_async_tx_descriptor * 930 struct dma_async_tx_descriptor *txd = &desc->async_tx; in dma_do_tasklet()
|
| /Linux-v5.4/drivers/spi/ |
| D | spi-dw-mid.c | 141 static struct dma_async_tx_descriptor *dw_spi_dma_prepare_tx(struct dw_spi *dws, in dw_spi_dma_prepare_tx() 145 struct dma_async_tx_descriptor *txdesc; in dw_spi_dma_prepare_tx() 187 static struct dma_async_tx_descriptor *dw_spi_dma_prepare_rx(struct dw_spi *dws, in dw_spi_dma_prepare_rx() 191 struct dma_async_tx_descriptor *rxdesc; in dw_spi_dma_prepare_rx() 242 struct dma_async_tx_descriptor *txdesc, *rxdesc; in mid_spi_dma_transfer()
|
| /Linux-v5.4/include/linux/platform_data/ |
| D | dma-ste-dma40.h | 178 dma_async_tx_descriptor *stedma40_slave_mem(struct dma_chan *chan, in stedma40_slave_mem() 199 dma_async_tx_descriptor *stedma40_slave_mem(struct dma_chan *chan, in stedma40_slave_mem()
|
| /Linux-v5.4/drivers/tty/serial/ |
| D | samsung.h | 66 struct dma_async_tx_descriptor *tx_desc; 67 struct dma_async_tx_descriptor *rx_desc;
|
| /Linux-v5.4/drivers/dma/sh/ |
| D | shdma-base.c | 70 static dma_cookie_t shdma_tx_submit(struct dma_async_tx_descriptor *tx) in shdma_tx_submit() 337 struct dma_async_tx_descriptor *tx = &desc->async_tx; in __ld_cleanup() 560 static struct dma_async_tx_descriptor *shdma_prep_sg(struct shdma_chan *schan, in shdma_prep_sg() 639 static struct dma_async_tx_descriptor *shdma_prep_memcpy( in shdma_prep_memcpy() 661 static struct dma_async_tx_descriptor *shdma_prep_slave_sg( in shdma_prep_slave_sg() 691 static struct dma_async_tx_descriptor *shdma_prep_dma_cyclic( in shdma_prep_dma_cyclic() 698 struct dma_async_tx_descriptor *desc; in shdma_prep_dma_cyclic() 884 struct dma_async_tx_descriptor *tx = &sdesc->async_tx; in shdma_reset()
|
| /Linux-v5.4/include/linux/dma/ |
| D | mxs-dma.h | 16 static inline struct dma_async_tx_descriptor *mxs_dmaengine_prep_pio( in mxs_dmaengine_prep_pio()
|
| /Linux-v5.4/drivers/md/ |
| D | raid5-log.h | 37 extern struct dma_async_tx_descriptor * 39 struct dma_async_tx_descriptor *tx);
|
| /Linux-v5.4/drivers/mmc/host/ |
| D | mxs-mmc.c | 214 static struct dma_async_tx_descriptor *mxs_mmc_prep_dma( in mxs_mmc_prep_dma() 218 struct dma_async_tx_descriptor *desc; in mxs_mmc_prep_dma() 253 struct dma_async_tx_descriptor *desc; in mxs_mmc_bc() 287 struct dma_async_tx_descriptor *desc; in mxs_mmc_ac() 348 struct dma_async_tx_descriptor *desc; in mxs_mmc_adtc()
|