Lines Matching refs:dev_addr

285 	dma_addr_t dev_addr;  in xen_swiotlb_alloc_coherent()  local
316 dev_addr = xen_phys_to_dma(hwdev, phys); in xen_swiotlb_alloc_coherent()
317 if (((dev_addr + size - 1 <= dma_mask)) && in xen_swiotlb_alloc_coherent()
319 *dma_handle = dev_addr; in xen_swiotlb_alloc_coherent()
335 dma_addr_t dev_addr, unsigned long attrs) in xen_swiotlb_free_coherent() argument
347 phys = xen_dma_to_phys(hwdev, dev_addr); in xen_swiotlb_free_coherent()
357 if (!WARN_ON((dev_addr + size - 1 > dma_mask) || in xen_swiotlb_free_coherent()
379 dma_addr_t dev_addr = xen_phys_to_dma(dev, phys); in xen_swiotlb_map_page() local
387 if (dma_capable(dev, dev_addr, size, true) && in xen_swiotlb_map_page()
389 !xen_arch_need_swiotlb(dev, phys, dev_addr) && in xen_swiotlb_map_page()
396 trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force); in xen_swiotlb_map_page()
403 dev_addr = xen_phys_to_dma(dev, map); in xen_swiotlb_map_page()
408 if (unlikely(!dma_capable(dev, dev_addr, size, true))) { in xen_swiotlb_map_page()
416 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr)))) in xen_swiotlb_map_page()
419 xen_dma_sync_for_device(dev, dev_addr, size, dir); in xen_swiotlb_map_page()
421 return dev_addr; in xen_swiotlb_map_page()
432 static void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_unmap_page() argument
435 phys_addr_t paddr = xen_dma_to_phys(hwdev, dev_addr); in xen_swiotlb_unmap_page()
440 if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr)))) in xen_swiotlb_unmap_page()
443 xen_dma_sync_for_cpu(hwdev, dev_addr, size, dir); in xen_swiotlb_unmap_page()
447 if (is_xen_swiotlb_buffer(hwdev, dev_addr)) in xen_swiotlb_unmap_page()