/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vf_error.c | 41 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put() 42 index = adev->virt.vf_errors.write_count % AMDGPU_VF_ERROR_ENTRY_SIZE; in amdgpu_vf_error_put() 43 adev->virt.vf_errors.code [index] = error_code; in amdgpu_vf_error_put() 44 adev->virt.vf_errors.flags [index] = error_flags; in amdgpu_vf_error_put() 45 adev->virt.vf_errors.data [index] = error_data; in amdgpu_vf_error_put() 46 adev->virt.vf_errors.write_count ++; in amdgpu_vf_error_put() 47 mutex_unlock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put() 58 (!adev->virt.ops) || (!adev->virt.ops->trans_msg)) { in amdgpu_vf_error_trans_all() 69 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_trans_all() 71 …if (adev->virt.vf_errors.write_count - adev->virt.vf_errors.read_count > AMDGPU_VF_ERROR_ENTRY_SIZ… in amdgpu_vf_error_trans_all() [all …]
|
D | amdgpu_virt.c | 55 AMDGPU_GEM_DOMAIN_VRAM, &adev->virt.csa_obj, in amdgpu_allocate_static_csa() 56 &adev->virt.csa_vmid0_addr, &ptr); in amdgpu_allocate_static_csa() 65 amdgpu_bo_free_kernel(&adev->virt.csa_obj, in amdgpu_free_static_csa() 66 &adev->virt.csa_vmid0_addr, in amdgpu_free_static_csa() 88 csa_tv.bo = &adev->virt.csa_obj->tbo; in amdgpu_map_static_csa() 100 *bo_va = amdgpu_vm_bo_add(adev, vm, adev->virt.csa_obj); in amdgpu_map_static_csa() 181 return adev->wb.wb[adev->virt.reg_val_offs]; in amdgpu_virt_kiq_rreg() 245 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_request_full_gpu() local 248 if (virt->ops && virt->ops->req_full_gpu) { in amdgpu_virt_request_full_gpu() 249 r = virt->ops->req_full_gpu(adev, init); in amdgpu_virt_request_full_gpu() [all …]
|
D | mxgpu_ai.c | 178 adev->virt.fw_reserve.checksum_key = in xgpu_ai_send_access_requests() 237 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_ai_mailbox_flr_work() local 238 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_ai_mailbox_flr_work() 296 schedule_work(&adev->virt.flr_work); in xgpu_ai_mailbox_rcv_irq() 324 adev->virt.ack_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs() 325 adev->virt.ack_irq.funcs = &xgpu_ai_mailbox_ack_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs() 326 adev->virt.rcv_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs() 327 adev->virt.rcv_irq.funcs = &xgpu_ai_mailbox_rcv_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs() 334 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_ai_mailbox_add_irq_id() 338 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 138, &adev->virt.ack_irq); in xgpu_ai_mailbox_add_irq_id() [all …]
|
D | amdgpu_virt.h | 217 ((amdgim_vf2pf_info *)adev->virt.fw_reserve.p_vf2pf)->field = (val); \ 222 (*val) = ((amdgim_vf2pf_info *)adev->virt.fw_reserve.p_vf2pf)->field; \ 227 if (!adev->virt.fw_reserve.p_pf2vf) \ 230 if (adev->virt.fw_reserve.p_pf2vf->version == 1) \ 231 *(val) = ((struct amdgim_pf2vf_info_v1 *)adev->virt.fw_reserve.p_pf2vf)->field; \ 232 if (adev->virt.fw_reserve.p_pf2vf->version == 2) \ 233 *(val) = ((struct amdgim_pf2vf_info_v2 *)adev->virt.fw_reserve.p_pf2vf)->field; \ 257 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_ENABLE_IOV) 260 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_IS_VF) 263 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_SRIOV_VBIOS) [all …]
|
D | mxgpu_vi.c | 514 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_vi_mailbox_flr_work() local 515 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_vi_mailbox_flr_work() 554 schedule_work(&adev->virt.flr_work); in xgpu_vi_mailbox_rcv_irq() 572 adev->virt.ack_irq.num_types = 1; in xgpu_vi_mailbox_set_irq_funcs() 573 adev->virt.ack_irq.funcs = &xgpu_vi_mailbox_ack_irq_funcs; in xgpu_vi_mailbox_set_irq_funcs() 574 adev->virt.rcv_irq.num_types = 1; in xgpu_vi_mailbox_set_irq_funcs() 575 adev->virt.rcv_irq.funcs = &xgpu_vi_mailbox_rcv_irq_funcs; in xgpu_vi_mailbox_set_irq_funcs() 582 r = amdgpu_irq_add_id(adev, AMDGPU_IH_CLIENTID_LEGACY, 135, &adev->virt.rcv_irq); in xgpu_vi_mailbox_add_irq_id() 586 r = amdgpu_irq_add_id(adev, AMDGPU_IH_CLIENTID_LEGACY, 138, &adev->virt.ack_irq); in xgpu_vi_mailbox_add_irq_id() 588 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_vi_mailbox_add_irq_id() [all …]
|
/Linux-v4.19/arch/m68k/mm/ |
D | sun3kmap.c | 30 static inline void do_page_mapin(unsigned long phys, unsigned long virt, in do_page_mapin() argument 40 sun3_put_pte(virt, pte); in do_page_mapin() 44 print_pte_vaddr(virt); in do_page_mapin() 49 static inline void do_pmeg_mapin(unsigned long phys, unsigned long virt, in do_pmeg_mapin() argument 53 if(sun3_get_segmap(virt & ~SUN3_PMEG_MASK) == SUN3_INVALID_PMEG) in do_pmeg_mapin() 54 mmu_emu_map_pmeg(sun3_get_context(), virt); in do_pmeg_mapin() 57 do_page_mapin(phys, virt, type); in do_pmeg_mapin() 59 virt += PAGE_SIZE; in do_pmeg_mapin() 68 unsigned long offset, virt, ret; in sun3_ioremap() local 89 virt = (unsigned long)area->addr; in sun3_ioremap() [all …]
|
/Linux-v4.19/include/asm-generic/ |
D | sections.h | 91 static inline bool memory_contains(void *begin, void *end, void *virt, in memory_contains() argument 94 return virt >= begin && virt + size <= end; in memory_contains() 108 static inline bool memory_intersects(void *begin, void *end, void *virt, in memory_intersects() argument 111 void *vend = virt + size; in memory_intersects() 113 return (virt >= begin && virt < end) || (vend >= begin && vend < end); in memory_intersects() 125 static inline bool init_section_contains(void *virt, size_t size) in init_section_contains() argument 127 return memory_contains(__init_begin, __init_end, virt, size); in init_section_contains() 139 static inline bool init_section_intersects(void *virt, size_t size) in init_section_intersects() argument 141 return memory_intersects(__init_begin, __init_end, virt, size); in init_section_intersects()
|
/Linux-v4.19/drivers/gpio/ |
D | gpio-stp-xway.c | 82 void __iomem *virt; member 103 return (xway_stp_r32(chip->virt, XWAY_STP_CPU0) & BIT(gpio)); in xway_stp_get() 122 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set() 123 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_set() 167 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); in xway_stp_hw_init() 168 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); in xway_stp_hw_init() 169 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); in xway_stp_hw_init() 170 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_hw_init() 171 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); in xway_stp_hw_init() 174 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MASK, in xway_stp_hw_init() [all …]
|
/Linux-v4.19/drivers/scsi/lpfc/ |
D | lpfc_mem.c | 132 pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc() 134 if (!pool->elements[i].virt) in lpfc_mem_alloc() 205 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_alloc() 286 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_free() 335 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all() 345 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all() 359 lpfc_mbuf_free(phba, mp->virt, mp->phys); in lpfc_mem_free_all() 406 ret = pool->elements[pool->current_count].virt; in lpfc_mbuf_alloc() 428 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in __lpfc_mbuf_free() argument 433 pool->elements[pool->current_count].virt = virt; in __lpfc_mbuf_free() [all …]
|
/Linux-v4.19/drivers/media/platform/s5p-mfc/ |
D | s5p_mfc_opr.c | 57 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf() 64 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf() 65 if (!b->virt) in s5p_mfc_alloc_priv_buf() 70 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf() 75 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf() 90 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_generic_buf() 91 if (!b->virt) in s5p_mfc_alloc_generic_buf() 94 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_generic_buf() 112 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_release_priv_buf() 114 b->virt = NULL; in s5p_mfc_release_priv_buf() [all …]
|
/Linux-v4.19/drivers/mtd/maps/ |
D | nettel.c | 219 nettel_amd_map.virt = ioremap_nocache(amdaddr, maxsize); in nettel_init() 220 if (!nettel_amd_map.virt) { in nettel_init() 279 iounmap(nettel_amd_map.virt); in nettel_init() 280 nettel_amd_map.virt = NULL; in nettel_init() 305 nettel_intel_map.virt = ioremap_nocache(intel0addr, maxsize); in nettel_init() 306 if (!nettel_intel_map.virt) { in nettel_init() 336 iounmap(nettel_intel_map.virt); in nettel_init() 339 nettel_intel_map.virt = ioremap_nocache(intel0addr, maxsize); in nettel_init() 340 if (!nettel_intel_map.virt) { in nettel_init() 411 iounmap(nettel_intel_map.virt); in nettel_init() [all …]
|
D | dc21285.c | 57 val.x[0] = *(uint8_t*)(map->virt + ofs); in dc21285_read8() 64 val.x[0] = *(uint16_t*)(map->virt + ofs); in dc21285_read16() 71 val.x[0] = *(uint32_t*)(map->virt + ofs); in dc21285_read32() 77 memcpy(to, (void*)(map->virt + from), len); in dc21285_copy_from() 86 *(uint8_t*)(map->virt + adr) = d.x[0]; in dc21285_write8() 95 *(uint16_t*)(map->virt + adr) = d.x[0]; in dc21285_write16() 102 *(uint32_t*)(map->virt + adr) = d.x[0]; in dc21285_write32() 179 dc21285_map.virt = ioremap(DC21285_FLASH, 16*1024*1024); in init_dc21285() 180 if (!dc21285_map.virt) { in init_dc21285() 192 iounmap(dc21285_map.virt); in init_dc21285() [all …]
|
D | ts5500_flash.c | 72 ts5500_map.virt = ioremap_nocache(ts5500_map.phys, ts5500_map.size); in init_ts5500_map() 74 if (!ts5500_map.virt) { in init_ts5500_map() 97 iounmap(ts5500_map.virt); in init_ts5500_map() 109 if (ts5500_map.virt) { in cleanup_ts5500_map() 110 iounmap(ts5500_map.virt); in cleanup_ts5500_map() 111 ts5500_map.virt = NULL; in cleanup_ts5500_map()
|
D | netsc520.c | 98 netsc520_map.virt = ioremap_nocache(netsc520_map.phys, netsc520_map.size); in init_netsc520() 100 if (!netsc520_map.virt) { in init_netsc520() 114 iounmap(netsc520_map.virt); in init_netsc520() 129 if (netsc520_map.virt) { in cleanup_netsc520() 130 iounmap(netsc520_map.virt); in cleanup_netsc520() 131 netsc520_map.virt = NULL; in cleanup_netsc520()
|
D | cfi_flagadm.c | 97 flagadm_map.virt = ioremap(FLASH_PHYS_ADDR, in init_flagadm() 100 if (!flagadm_map.virt) { in init_flagadm() 115 iounmap((void __iomem *)flagadm_map.virt); in init_flagadm() 125 if (flagadm_map.virt) { in cleanup_flagadm() 126 iounmap((void __iomem *)flagadm_map.virt); in cleanup_flagadm() 127 flagadm_map.virt = NULL; in cleanup_flagadm()
|
/Linux-v4.19/arch/unicore32/mm/ |
D | ioremap.c | 48 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument 51 return ioremap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page() 66 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 68 unsigned long addr = virt, end = virt + (size & ~(SZ_4M - 1)); in unmap_area_sections() 98 flush_tlb_kernel_range(virt, end); in unmap_area_sections() 102 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 105 unsigned long addr = virt, end = virt + size; in remap_area_sections() 112 unmap_area_sections(virt, size); in remap_area_sections()
|
D | mm.h | 20 static inline pmd_t *pmd_off(pgd_t *pgd, unsigned long virt) in pmd_off() argument 22 return pmd_offset((pud_t *)pgd, virt); in pmd_off() 25 static inline pmd_t *pmd_off_k(unsigned long virt) in pmd_off_k() argument 27 return pmd_off(pgd_offset_k(virt), virt); in pmd_off_k()
|
/Linux-v4.19/arch/arm/mach-hisi/ |
D | platsmp.c | 108 void __iomem *virt; in hix5hd2_set_scu_boot_addr() local 110 virt = ioremap(start_addr, PAGE_SIZE); in hix5hd2_set_scu_boot_addr() 112 writel_relaxed(0xe51ff004, virt); /* ldr pc, [pc, #-4] */ in hix5hd2_set_scu_boot_addr() 113 writel_relaxed(jump_addr, virt + 4); /* pc jump phy address */ in hix5hd2_set_scu_boot_addr() 114 iounmap(virt); in hix5hd2_set_scu_boot_addr() 144 void __iomem *virt; in hip01_set_boot_addr() local 146 virt = phys_to_virt(start_addr); in hip01_set_boot_addr() 148 writel_relaxed(0xe51ff004, virt); in hip01_set_boot_addr() 149 writel_relaxed(jump_addr, virt + 4); in hip01_set_boot_addr()
|
/Linux-v4.19/include/crypto/ |
D | cbc.h | 26 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_segment() 27 u8 *dst = walk->dst.virt.addr; in crypto_cbc_encrypt_segment() 48 u8 *src = walk->src.virt.addr; in crypto_cbc_encrypt_inplace() 75 if (walk.src.virt.addr == walk.dst.virt.addr) in crypto_cbc_encrypt_walk() 91 u8 *src = walk->src.virt.addr; in crypto_cbc_decrypt_segment() 92 u8 *dst = walk->dst.virt.addr; in crypto_cbc_decrypt_segment() 115 u8 *src = walk->src.virt.addr; in crypto_cbc_decrypt_inplace() 140 if (walk->src.virt.addr == walk->dst.virt.addr) in crypto_cbc_decrypt_blocks()
|
/Linux-v4.19/arch/powerpc/mm/ |
D | fsl_booke_mmu.c | 110 static void settlbcam(int index, unsigned long virt, phys_addr_t phys, in settlbcam() argument 124 TLBCAM[index].MAS2 = virt & PAGE_MASK; in settlbcam() 143 tlbcam_addrs[index].start = virt; in settlbcam() 144 tlbcam_addrs[index].limit = virt + size - 1; in settlbcam() 148 unsigned long calc_cam_sz(unsigned long ram, unsigned long virt, in calc_cam_sz() argument 152 unsigned int align = __ffs(virt | phys); in calc_cam_sz() 173 static unsigned long map_mem_in_cams_addr(phys_addr_t phys, unsigned long virt, in map_mem_in_cams_addr() argument 184 cam_sz = calc_cam_sz(ram, virt, phys); in map_mem_in_cams_addr() 186 settlbcam(i, virt, phys, cam_sz, in map_mem_in_cams_addr() 191 virt += cam_sz; in map_mem_in_cams_addr() [all …]
|
/Linux-v4.19/arch/arm/mm/ |
D | ioremap.c | 107 int ioremap_page(unsigned long virt, unsigned long phys, in ioremap_page() argument 110 return ioremap_page_range(virt, virt + PAGE_SIZE, phys, in ioremap_page() 140 static void unmap_area_sections(unsigned long virt, unsigned long size) in unmap_area_sections() argument 142 unsigned long addr = virt, end = virt + (size & ~(SZ_1M - 1)); in unmap_area_sections() 183 flush_tlb_kernel_range(virt, end); in unmap_area_sections() 187 remap_area_sections(unsigned long virt, unsigned long pfn, in remap_area_sections() argument 190 unsigned long addr = virt, end = virt + size; in remap_area_sections() 199 unmap_area_sections(virt, size); in remap_area_sections() 219 remap_area_supersections(unsigned long virt, unsigned long pfn, in remap_area_supersections() argument 222 unsigned long addr = virt, end = virt + size; in remap_area_supersections() [all …]
|
/Linux-v4.19/arch/x86/pci/ |
D | mmconfig_64.c | 23 if (cfg && cfg->virt) in pci_dev_base() 24 return cfg->virt + (PCI_MMCFG_BUS_OFFSET(bus) | (devfn << 12)); in pci_dev_base() 139 cfg->virt = mcfg_ioremap(cfg); in pci_mmcfg_arch_map() 140 if (!cfg->virt) { in pci_mmcfg_arch_map() 150 if (cfg && cfg->virt) { in pci_mmcfg_arch_unmap() 151 iounmap(cfg->virt + PCI_MMCFG_BUS_OFFSET(cfg->start_bus)); in pci_mmcfg_arch_unmap() 152 cfg->virt = NULL; in pci_mmcfg_arch_unmap()
|
/Linux-v4.19/arch/xtensa/mm/ |
D | cache.c | 148 unsigned long virt; in flush_dcache_page() local 160 virt = TLBTEMP_BASE_1 + (phys & DCACHE_ALIAS_MASK); in flush_dcache_page() 161 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 163 virt = TLBTEMP_BASE_1 + (temp & DCACHE_ALIAS_MASK); in flush_dcache_page() 166 __flush_invalidate_dcache_page_alias(virt, phys); in flush_dcache_page() 169 __invalidate_icache_page_alias(virt, phys); in flush_dcache_page() 201 unsigned long virt = TLBTEMP_BASE_1 + (address & DCACHE_ALIAS_MASK); in local_flush_cache_page() local 203 __flush_invalidate_dcache_page_alias(virt, phys); in local_flush_cache_page() 204 __invalidate_icache_page_alias(virt, phys); in local_flush_cache_page()
|
/Linux-v4.19/arch/x86/crypto/ |
D | glue_helper.c | 48 const u8 *src = walk.src.virt.addr; in glue_ecb_req_128bit() 49 u8 *dst = walk.dst.virt.addr; in glue_ecb_req_128bit() 92 const u128 *src = (u128 *)walk.src.virt.addr; in glue_cbc_encrypt_req_128bit() 93 u128 *dst = (u128 *)walk.dst.virt.addr; in glue_cbc_encrypt_req_128bit() 125 const u128 *src = walk.src.virt.addr; in glue_cbc_decrypt_req_128bit() 126 u128 *dst = walk.dst.virt.addr; in glue_cbc_decrypt_req_128bit() 185 const u128 *src = walk.src.virt.addr; in glue_ctr_req_128bit() 186 u128 *dst = walk.dst.virt.addr; in glue_ctr_req_128bit() 226 memcpy(&tmp, walk.src.virt.addr, nbytes); in glue_ctr_req_128bit() 229 memcpy(walk.dst.virt.addr, &tmp, nbytes); in glue_ctr_req_128bit() [all …]
|
/Linux-v4.19/arch/powerpc/platforms/4xx/ |
D | ocm.c | 48 void __iomem *virt; member 90 if (!ocm_reg->virt) in ocm_free_region() 95 offset = addr - ocm_reg->virt; in ocm_free_region() 183 ocm->nc.virt = __ioremap(ocm->nc.phys, ocm->nc.memtotal, in ocm_init_node() 186 if (!ocm->nc.virt) { in ocm_init_node() 199 ocm->c.virt = __ioremap(ocm->c.phys, ocm->c.memtotal, in ocm_init_node() 202 if (!ocm->c.virt) { in ocm_init_node() 215 if (ocm->nc.virt) { in ocm_init_node() 220 if (ocm->c.virt) { in ocm_init_node() 253 seq_printf(m, "NC.VirtAddr : 0x%p\n", ocm->nc.virt); in ocm_debugfs_show() [all …]
|