Home
last modified time | relevance | path

Searched refs:cpu_addr (Results 1 – 25 of 138) sorted by relevance

123456

/Linux-v5.4/kernel/dma/
Dmapping.c112 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_get_sgtable() argument
125 pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr); in dma_common_get_sgtable()
130 page = virt_to_page(cpu_addr); in dma_common_get_sgtable()
151 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_get_sgtable_attrs() argument
157 return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, in dma_get_sgtable_attrs()
161 return ops->get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); in dma_get_sgtable_attrs()
188 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_common_mmap() argument
200 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in dma_common_mmap()
211 pfn = arch_dma_coherent_to_pfn(dev, cpu_addr, dma_addr); in dma_common_mmap()
215 pfn = page_to_pfn(virt_to_page(cpu_addr)); in dma_common_mmap()
[all …]
Dremap.c14 struct page **dma_common_find_pages(void *cpu_addr) in dma_common_find_pages() argument
16 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_find_pages()
88 void dma_common_free_remap(void *cpu_addr, size_t size) in dma_common_free_remap() argument
90 struct vm_struct *area = find_vm_area(cpu_addr); in dma_common_free_remap()
93 WARN(1, "trying to free invalid coherent area: %p\n", cpu_addr); in dma_common_free_remap()
97 unmap_kernel_range((unsigned long)cpu_addr, PAGE_ALIGN(size)); in dma_common_free_remap()
98 vunmap(cpu_addr); in dma_common_free_remap()
263 long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, in arch_dma_coherent_to_pfn() argument
Ddirect.c183 void dma_direct_free_pages(struct device *dev, size_t size, void *cpu_addr, in dma_direct_free_pages() argument
191 __dma_direct_free_pages(dev, size, cpu_addr); in dma_direct_free_pages()
196 set_memory_encrypted((unsigned long)cpu_addr, 1 << page_order); in dma_direct_free_pages()
200 cpu_addr = cached_kernel_address(cpu_addr); in dma_direct_free_pages()
201 __dma_direct_free_pages(dev, size, virt_to_page(cpu_addr)); in dma_direct_free_pages()
214 void *cpu_addr, dma_addr_t dma_addr, unsigned long attrs) in dma_direct_free() argument
218 arch_dma_free(dev, size, cpu_addr, dma_addr, attrs); in dma_direct_free()
220 dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs); in dma_direct_free()
/Linux-v5.4/drivers/pci/controller/dwc/
Dpcie-designware.c242 int type, u64 cpu_addr, in dw_pcie_prog_outbound_atu_unroll() argument
248 lower_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu_unroll()
250 upper_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu_unroll()
252 lower_32_bits(cpu_addr + size - 1)); in dw_pcie_prog_outbound_atu_unroll()
278 u64 cpu_addr, u64 pci_addr, u32 size) in dw_pcie_prog_outbound_atu() argument
283 cpu_addr = pci->ops->cpu_addr_fixup(pci, cpu_addr); in dw_pcie_prog_outbound_atu()
286 dw_pcie_prog_outbound_atu_unroll(pci, index, type, cpu_addr, in dw_pcie_prog_outbound_atu()
294 lower_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu()
296 upper_32_bits(cpu_addr)); in dw_pcie_prog_outbound_atu()
298 lower_32_bits(cpu_addr + size - 1)); in dw_pcie_prog_outbound_atu()
[all …]
/Linux-v5.4/arch/ia64/kernel/
Ddma-mapping.c17 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, in arch_dma_free() argument
20 dma_direct_free_pages(dev, size, cpu_addr, dma_addr, attrs); in arch_dma_free()
23 long arch_dma_coherent_to_pfn(struct device *dev, void *cpu_addr, in arch_dma_coherent_to_pfn() argument
26 return page_to_pfn(virt_to_page(cpu_addr)); in arch_dma_coherent_to_pfn()
/Linux-v5.4/drivers/iommu/
Ddma-iommu.c923 static void __iommu_dma_free(struct device *dev, size_t size, void *cpu_addr) in __iommu_dma_free() argument
931 dma_free_from_pool(cpu_addr, alloc_size)) in __iommu_dma_free()
934 if (IS_ENABLED(CONFIG_DMA_REMAP) && is_vmalloc_addr(cpu_addr)) { in __iommu_dma_free()
939 pages = dma_common_find_pages(cpu_addr); in __iommu_dma_free()
941 page = vmalloc_to_page(cpu_addr); in __iommu_dma_free()
942 dma_common_free_remap(cpu_addr, alloc_size); in __iommu_dma_free()
945 page = virt_to_page(cpu_addr); in __iommu_dma_free()
954 static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr, in iommu_dma_free() argument
958 __iommu_dma_free(dev, size, cpu_addr); in iommu_dma_free()
968 void *cpu_addr; in iommu_dma_alloc_pages() local
[all …]
/Linux-v5.4/arch/s390/kvm/
Dtrace.h220 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr, \
222 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr, parameter),
227 __field(__u16, cpu_addr)
234 __entry->cpu_addr = cpu_addr;
242 __entry->cpu_addr, __entry->parameter)
246 TP_PROTO(VCPU_PROTO_COMMON, __u8 order_code, __u16 cpu_addr),
247 TP_ARGS(VCPU_ARGS_COMMON, order_code, cpu_addr),
252 __field(__u16, cpu_addr)
258 __entry->cpu_addr = cpu_addr;
265 __entry->cpu_addr)
Dsigp.c283 u16 cpu_addr, u32 parameter, u64 *status_reg) in handle_sigp_dst() argument
286 struct kvm_vcpu *dst_vcpu = kvm_get_vcpu_by_id(vcpu->kvm, cpu_addr); in handle_sigp_dst()
360 u16 cpu_addr) in handle_sigp_order_in_user_space() argument
404 order_code, cpu_addr); in handle_sigp_order_in_user_space()
414 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp() local
423 if (handle_sigp_order_in_user_space(vcpu, order_code, cpu_addr)) in kvm_s390_handle_sigp()
431 trace_kvm_s390_handle_sigp(vcpu, order_code, cpu_addr, parameter); in kvm_s390_handle_sigp()
439 rc = handle_sigp_dst(vcpu, order_code, cpu_addr, in kvm_s390_handle_sigp()
463 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp_pei() local
467 trace_kvm_s390_handle_sigp_pei(vcpu, order_code, cpu_addr); in kvm_s390_handle_sigp_pei()
[all …]
/Linux-v5.4/arch/arm/mm/
Ddma-mapping.c57 void *cpu_addr; member
203 static void arm_coherent_dma_free(struct device *dev, size_t size, void *cpu_addr,
206 void *cpu_addr, dma_addr_t dma_addr, size_t size,
581 void *cpu_addr, size_t size, bool want_vaddr) in __free_from_contiguous() argument
585 dma_common_free_remap(cpu_addr, size); in __free_from_contiguous()
641 __free_from_contiguous(args->dev, args->page, args->cpu_addr, in cma_allocator_free()
658 __free_from_pool(args->cpu_addr, args->size); in pool_allocator_free()
677 dma_common_free_remap(args->cpu_addr, args->size); in remap_allocator_free()
788 void *cpu_addr, dma_addr_t dma_addr, size_t size, in __arm_dma_mmap() argument
797 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in __arm_dma_mmap()
[all …]
Ddma-mapping-nommu.c55 void *cpu_addr, dma_addr_t dma_addr, in arm_nommu_dma_free() argument
58 int ret = dma_release_from_global_coherent(get_order(size), cpu_addr); in arm_nommu_dma_free()
64 void *cpu_addr, dma_addr_t dma_addr, size_t size, in arm_nommu_dma_mmap() argument
69 if (dma_mmap_from_global_coherent(vma, cpu_addr, size, &ret)) in arm_nommu_dma_mmap()
71 if (dma_mmap_from_dev_coherent(dev, vma, cpu_addr, size, &ret)) in arm_nommu_dma_mmap()
/Linux-v5.4/drivers/pci/controller/
Dpcie-cadence.c12 u64 cpu_addr, u64 pci_addr, size_t size) in cdns_pcie_set_outbound_region() argument
76 cpu_addr -= pcie->mem_res->start; in cdns_pcie_set_outbound_region()
78 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region()
79 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region()
86 u32 r, u64 cpu_addr) in cdns_pcie_set_outbound_region_for_normal_msg() argument
103 cpu_addr -= pcie->mem_res->start; in cdns_pcie_set_outbound_region_for_normal_msg()
105 (lower_32_bits(cpu_addr) & GENMASK(31, 8)); in cdns_pcie_set_outbound_region_for_normal_msg()
106 addr1 = upper_32_bits(cpu_addr); in cdns_pcie_set_outbound_region_for_normal_msg()
Dpci-xgene.c369 u64 cpu_addr, u64 pci_addr) in xgene_pcie_setup_ob_reg() argument
391 xgene_pcie_writel(port, offset, lower_32_bits(cpu_addr)); in xgene_pcie_setup_ob_reg()
392 xgene_pcie_writel(port, offset + 0x04, upper_32_bits(cpu_addr)); in xgene_pcie_setup_ob_reg()
494 u64 cpu_addr = range->cpu_addr; in xgene_pcie_setup_ib_reg() local
511 bar_low = pcie_bar_low_val((u32)cpu_addr, flags); in xgene_pcie_setup_ib_reg()
517 writel(upper_32_bits(cpu_addr), bar_addr + 0x4); in xgene_pcie_setup_ib_reg()
527 xgene_pcie_writel(port, IBAR3L + 0x4, upper_32_bits(cpu_addr)); in xgene_pcie_setup_ib_reg()
552 u64 end = range.cpu_addr + range.size - 1; in xgene_pcie_parse_map_dma_ranges()
555 range.flags, range.cpu_addr, end, range.pci_addr); in xgene_pcie_parse_map_dma_ranges()
/Linux-v5.4/drivers/media/common/saa7146/
Dsaa7146_core.c415 dev->d_rps0.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, in saa7146_init_one()
417 if (!dev->d_rps0.cpu_addr) in saa7146_init_one()
420 dev->d_rps1.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, in saa7146_init_one()
422 if (!dev->d_rps1.cpu_addr) in saa7146_init_one()
425 dev->d_i2c.cpu_addr = pci_zalloc_consistent(pci, SAA7146_RPS_MEM, in saa7146_init_one()
427 if (!dev->d_i2c.cpu_addr) in saa7146_init_one()
474 pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_i2c.cpu_addr, in saa7146_init_one()
477 pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps1.cpu_addr, in saa7146_init_one()
480 pci_free_consistent(pci, SAA7146_RPS_MEM, dev->d_rps0.cpu_addr, in saa7146_init_one()
503 { dev->d_i2c.cpu_addr, dev->d_i2c.dma_handle }, in saa7146_remove_one()
[all …]
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_fence.c102 if (drv->cpu_addr) in amdgpu_fence_write()
103 *drv->cpu_addr = cpu_to_le32(seq); in amdgpu_fence_write()
119 if (drv->cpu_addr) in amdgpu_fence_read()
120 seq = le32_to_cpu(*drv->cpu_addr); in amdgpu_fence_read()
398 ring->fence_drv.cpu_addr = &adev->wb.wb[ring->fence_offs]; in amdgpu_fence_driver_start_ring()
403 ring->fence_drv.cpu_addr = adev->uvd.inst[ring->me].cpu_addr + index; in amdgpu_fence_driver_start_ring()
415 ring->fence_drv.gpu_addr, ring->fence_drv.cpu_addr); in amdgpu_fence_driver_start_ring()
443 ring->fence_drv.cpu_addr = NULL; in amdgpu_fence_driver_init_ring()
730 le32_to_cpu(*(ring->fence_drv.cpu_addr + 2))); in amdgpu_debugfs_fence_info()
733 le32_to_cpu(*(ring->fence_drv.cpu_addr + 4))); in amdgpu_debugfs_fence_info()
[all …]
/Linux-v5.4/include/linux/
Ddma-mapping.h163 void *cpu_addr, size_t size, int *ret);
167 int dma_mmap_from_global_coherent(struct vm_area_struct *vma, void *cpu_addr,
187 void *cpu_addr, size_t size, in dma_mmap_from_global_coherent() argument
447 void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr,
456 void *cpu_addr, dma_addr_t dma_addr, size_t size,
459 void *cpu_addr, dma_addr_t dma_addr, size_t size,
524 static void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, in dma_free_attrs() argument
542 struct sg_table *sgt, void *cpu_addr, dma_addr_t dma_addr, in dma_get_sgtable_attrs() argument
548 void *cpu_addr, dma_addr_t dma_addr, size_t size, in dma_mmap_attrs() argument
621 void *cpu_addr, dma_addr_t dma_addr, size_t size,
[all …]
/Linux-v5.4/arch/sparc/kernel/
Dioport.c267 bool sparc_dma_free_resource(void *cpu_addr, size_t size) in sparc_dma_free_resource() argument
269 unsigned long addr = (unsigned long)cpu_addr; in sparc_dma_free_resource()
274 printk("%s: cannot free %p\n", __func__, cpu_addr); in sparc_dma_free_resource()
279 printk("%s: unaligned va %p\n", __func__, cpu_addr); in sparc_dma_free_resource()
356 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr, in arch_dma_free() argument
359 if (!sparc_dma_free_resource(cpu_addr, PAGE_ALIGN(size))) in arch_dma_free()
363 srmmu_unmapiorange((unsigned long)cpu_addr, size); in arch_dma_free()
/Linux-v5.4/drivers/dma/
Dst_fdma.h185 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
190 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
197 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
207 readl((fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
212 writel((val), (fchan)->fdev->slim_rproc->mem[ST_SLIM_DMEM].cpu_addr \
/Linux-v5.4/arch/alpha/kernel/
Dpci_iommu.c264 pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size, in pci_map_single_1() argument
276 paddr = __pa(cpu_addr); in pci_map_single_1()
285 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
296 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
330 ret += (unsigned long)cpu_addr & ~PAGE_MASK; in pci_map_single_1()
333 cpu_addr, size, npages, ret, __builtin_return_address(0)); in pci_map_single_1()
452 void *cpu_addr; in alpha_pci_alloc_coherent() local
458 cpu_addr = (void *)__get_free_pages(gfp | __GFP_ZERO, order); in alpha_pci_alloc_coherent()
459 if (! cpu_addr) { in alpha_pci_alloc_coherent()
467 memset(cpu_addr, 0, size); in alpha_pci_alloc_coherent()
[all …]
/Linux-v5.4/drivers/remoteproc/
Dimx_rproc.c55 void __iomem *cpu_addr; member
233 va = (__force void *)(priv->mem[i].cpu_addr + offset); in imx_rproc_da_to_va()
267 priv->mem[b].cpu_addr = devm_ioremap(&pdev->dev, in imx_rproc_addr_init()
269 if (!priv->mem[b].cpu_addr) { in imx_rproc_addr_init()
298 priv->mem[b].cpu_addr = devm_ioremap_resource(&pdev->dev, &res); in imx_rproc_addr_init()
299 if (IS_ERR(priv->mem[b].cpu_addr)) { in imx_rproc_addr_init()
301 err = PTR_ERR(priv->mem[b].cpu_addr); in imx_rproc_addr_init()
Dwkup_m3_rproc.c35 void __iomem *cpu_addr; member
98 va = (__force void *)(wkupm3->mem[i].cpu_addr + offset); in wkup_m3_rproc_da_to_va()
171 wkupm3->mem[i].cpu_addr = devm_ioremap_resource(dev, res); in wkup_m3_rproc_probe()
172 if (IS_ERR(wkupm3->mem[i].cpu_addr)) { in wkup_m3_rproc_probe()
175 ret = PTR_ERR(wkupm3->mem[i].cpu_addr); in wkup_m3_rproc_probe()
Dkeystone_remoteproc.c35 void __iomem *cpu_addr; member
271 va = ksproc->mem[i].cpu_addr + offset; in keystone_rproc_da_to_va()
279 va = ksproc->mem[i].cpu_addr + offset; in keystone_rproc_da_to_va()
313 ksproc->mem[i].cpu_addr = devm_ioremap_resource(dev, res); in keystone_rproc_of_get_memories()
314 if (IS_ERR(ksproc->mem[i].cpu_addr)) { in keystone_rproc_of_get_memories()
317 return PTR_ERR(ksproc->mem[i].cpu_addr); in keystone_rproc_of_get_memories()
325 memset((__force void *)ksproc->mem[i].cpu_addr, 0, in keystone_rproc_of_get_memories()
/Linux-v5.4/arch/microblaze/pci/
Dpci-common.c439 range.cpu_addr, range.size); in pci_process_bridge_OF_ranges()
446 if (range.cpu_addr == OF_BAD_ADDR || range.size == 0) in pci_process_bridge_OF_ranges()
454 range.cpu_addr, range.cpu_addr + range.size - 1, in pci_process_bridge_OF_ranges()
467 hose->io_base_virt = ioremap(range.cpu_addr, in pci_process_bridge_OF_ranges()
478 hose->io_base_phys = range.cpu_addr - range.pci_addr; in pci_process_bridge_OF_ranges()
482 range.cpu_addr = range.pci_addr; in pci_process_bridge_OF_ranges()
487 range.cpu_addr, range.cpu_addr + range.size - 1, in pci_process_bridge_OF_ranges()
499 isa_mb = range.cpu_addr; in pci_process_bridge_OF_ranges()
502 isa_mem_base = range.cpu_addr; in pci_process_bridge_OF_ranges()
503 hose->isa_mem_phys = range.cpu_addr; in pci_process_bridge_OF_ranges()
[all …]
/Linux-v5.4/arch/mips/pci/
Dpci-legacy.c155 range.cpu_addr, in pci_load_of_ranges()
156 range.cpu_addr + range.size - 1); in pci_load_of_ranges()
158 (unsigned long)ioremap(range.cpu_addr, in pci_load_of_ranges()
164 range.cpu_addr, in pci_load_of_ranges()
165 range.cpu_addr + range.size - 1); in pci_load_of_ranges()
/Linux-v5.4/include/xen/arm/
Dpage-coherent.h15 void *cpu_addr, dma_addr_t dma_handle, unsigned long attrs) in xen_free_coherent_pages() argument
17 dma_direct_free(hwdev, size, cpu_addr, dma_handle, attrs); in xen_free_coherent_pages()
/Linux-v5.4/arch/x86/include/asm/xen/
Dpage-coherent.h18 void *cpu_addr, dma_addr_t dma_handle, in xen_free_coherent_pages() argument
21 free_pages((unsigned long) cpu_addr, get_order(size)); in xen_free_coherent_pages()

123456