Searched refs:DMA_ATTR_WEAK_ORDERING (Results 1 – 16 of 16) sorted by relevance
8 #define STMMAC_RX_DMA_ATTR (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
8 DMA_ATTR_WEAK_ORDERING section in DMA attributes11 DMA_ATTR_WEAK_ORDERING specifies that reads and writes to the mapping14 Since it is optional for platforms to implement DMA_ATTR_WEAK_ORDERING,
241 dma_attr |= DMA_ATTR_WEAK_ORDERING; in ib_umem_get()
139 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
103 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
116 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
201 if (attrs & DMA_ATTR_WEAK_ORDERING) in dma_4v_alloc_coherent()398 if (attrs & DMA_ATTR_WEAK_ORDERING) in dma_4v_map_page()496 if (attrs & DMA_ATTR_WEAK_ORDERING) in dma_4v_map_sg()
317 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
370 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
301 DMA_ATTR_WEAK_ORDERING); in bnxt_rx_xdp()
734 DMA_ATTR_WEAK_ORDERING); in __bnxt_alloc_rx_page()757 DMA_ATTR_WEAK_ORDERING); in __bnxt_alloc_rx_frag()871 DMA_ATTR_WEAK_ORDERING); in bnxt_alloc_rx_page()990 DMA_ATTR_WEAK_ORDERING); in bnxt_rx_multi_page_skb()1025 DMA_ATTR_WEAK_ORDERING); in bnxt_rx_page_skb()1069 bp->rx_dir, DMA_ATTR_WEAK_ORDERING); in bnxt_rx_skb()1147 DMA_ATTR_WEAK_ORDERING); in __bnxt_rx_agg_pages()1714 DMA_ATTR_WEAK_ORDERING); in bnxt_tpa_end()2951 DMA_ATTR_WEAK_ORDERING); in bnxt_free_one_rx_ring_skbs()2975 DMA_ATTR_WEAK_ORDERING); in bnxt_free_one_rx_ring_skbs()[all …]
597 imem->attrs = DMA_ATTR_WEAK_ORDERING | in gk20a_instmem_new()
207 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)
23 #define DMA_ATTR_WEAK_ORDERING (1UL << 1) macro
185 if (unlikely(attrs & DMA_ATTR_WEAK_ORDERING)) in tce_build_cell()
136 (DMA_ATTR_SKIP_CPU_SYNC | DMA_ATTR_WEAK_ORDERING)