Lines Matching refs:dev_addr

270 	dma_addr_t dev_addr;  in xen_swiotlb_alloc_coherent()  local
301 dev_addr = xen_phys_to_dma(hwdev, phys); in xen_swiotlb_alloc_coherent()
302 if (((dev_addr + size - 1 <= dma_mask)) && in xen_swiotlb_alloc_coherent()
304 *dma_handle = dev_addr; in xen_swiotlb_alloc_coherent()
320 dma_addr_t dev_addr, unsigned long attrs) in xen_swiotlb_free_coherent() argument
332 phys = xen_dma_to_phys(hwdev, dev_addr); in xen_swiotlb_free_coherent()
342 if (!WARN_ON((dev_addr + size - 1 > dma_mask) || in xen_swiotlb_free_coherent()
364 dma_addr_t dev_addr = xen_phys_to_dma(dev, phys); in xen_swiotlb_map_page() local
372 if (dma_capable(dev, dev_addr, size, true) && in xen_swiotlb_map_page()
374 !xen_arch_need_swiotlb(dev, phys, dev_addr) && in xen_swiotlb_map_page()
381 trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force); in xen_swiotlb_map_page()
388 dev_addr = xen_phys_to_dma(dev, map); in xen_swiotlb_map_page()
393 if (unlikely(!dma_capable(dev, dev_addr, size, true))) { in xen_swiotlb_map_page()
401 if (pfn_valid(PFN_DOWN(dma_to_phys(dev, dev_addr)))) in xen_swiotlb_map_page()
404 xen_dma_sync_for_device(dev, dev_addr, size, dir); in xen_swiotlb_map_page()
406 return dev_addr; in xen_swiotlb_map_page()
417 static void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_unmap_page() argument
420 phys_addr_t paddr = xen_dma_to_phys(hwdev, dev_addr); in xen_swiotlb_unmap_page()
425 if (pfn_valid(PFN_DOWN(dma_to_phys(hwdev, dev_addr)))) in xen_swiotlb_unmap_page()
428 xen_dma_sync_for_cpu(hwdev, dev_addr, size, dir); in xen_swiotlb_unmap_page()
432 if (is_xen_swiotlb_buffer(hwdev, dev_addr)) in xen_swiotlb_unmap_page()