Lines Matching refs:dev_addr

296 	dma_addr_t dev_addr;  in xen_swiotlb_alloc_coherent()  local
324 dev_addr = xen_phys_to_bus(phys); in xen_swiotlb_alloc_coherent()
325 if (((dev_addr + size - 1 <= dma_mask)) && in xen_swiotlb_alloc_coherent()
327 *dma_handle = dev_addr; in xen_swiotlb_alloc_coherent()
341 dma_addr_t dev_addr, unsigned long attrs) in xen_swiotlb_free_coherent() argument
352 phys = xen_bus_to_phys(dev_addr); in xen_swiotlb_free_coherent()
354 if (((dev_addr + size - 1 <= dma_mask)) || in xen_swiotlb_free_coherent()
374 dma_addr_t dev_addr = xen_phys_to_bus(phys); in xen_swiotlb_map_page() local
382 if (dma_capable(dev, dev_addr, size) && in xen_swiotlb_map_page()
384 !xen_arch_need_swiotlb(dev, phys, dev_addr) && in xen_swiotlb_map_page()
389 xen_dma_map_page(dev, page, dev_addr, offset, size, dir, attrs); in xen_swiotlb_map_page()
390 return dev_addr; in xen_swiotlb_map_page()
396 trace_swiotlb_bounced(dev, dev_addr, size, swiotlb_force); in xen_swiotlb_map_page()
403 dev_addr = xen_phys_to_bus(map); in xen_swiotlb_map_page()
405 dev_addr, map & ~PAGE_MASK, size, dir, attrs); in xen_swiotlb_map_page()
410 if (dma_capable(dev, dev_addr, size)) in xen_swiotlb_map_page()
411 return dev_addr; in xen_swiotlb_map_page()
427 static void xen_unmap_single(struct device *hwdev, dma_addr_t dev_addr, in xen_unmap_single() argument
431 phys_addr_t paddr = xen_bus_to_phys(dev_addr); in xen_unmap_single()
435 xen_dma_unmap_page(hwdev, dev_addr, size, dir, attrs); in xen_unmap_single()
438 if (is_xen_swiotlb_buffer(dev_addr)) { in xen_unmap_single()
455 static void xen_swiotlb_unmap_page(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_unmap_page() argument
459 xen_unmap_single(hwdev, dev_addr, size, dir, attrs); in xen_swiotlb_unmap_page()
473 xen_swiotlb_sync_single(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_sync_single() argument
477 phys_addr_t paddr = xen_bus_to_phys(dev_addr); in xen_swiotlb_sync_single()
482 xen_dma_sync_single_for_cpu(hwdev, dev_addr, size, dir); in xen_swiotlb_sync_single()
485 if (is_xen_swiotlb_buffer(dev_addr)) in xen_swiotlb_sync_single()
489 xen_dma_sync_single_for_device(hwdev, dev_addr, size, dir); in xen_swiotlb_sync_single()
498 xen_swiotlb_sync_single_for_cpu(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_sync_single_for_cpu() argument
501 xen_swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_CPU); in xen_swiotlb_sync_single_for_cpu()
505 xen_swiotlb_sync_single_for_device(struct device *hwdev, dma_addr_t dev_addr, in xen_swiotlb_sync_single_for_device() argument
508 xen_swiotlb_sync_single(hwdev, dev_addr, size, dir, SYNC_FOR_DEVICE); in xen_swiotlb_sync_single_for_device()
558 dma_addr_t dev_addr = xen_phys_to_bus(paddr); in xen_swiotlb_map_sg_attrs() local
561 xen_arch_need_swiotlb(hwdev, paddr, dev_addr) || in xen_swiotlb_map_sg_attrs()
562 !dma_capable(hwdev, dev_addr, sg->length) || in xen_swiotlb_map_sg_attrs()
579 dev_addr = xen_phys_to_bus(map); in xen_swiotlb_map_sg_attrs()
581 dev_addr, in xen_swiotlb_map_sg_attrs()
586 sg->dma_address = dev_addr; in xen_swiotlb_map_sg_attrs()
592 dev_addr, in xen_swiotlb_map_sg_attrs()
597 sg->dma_address = dev_addr; in xen_swiotlb_map_sg_attrs()