/Linux-v5.15/mm/ |
D | truncate.c | 34 static inline void __clear_shadow_entry(struct address_space *mapping, in __clear_shadow_entry() argument 37 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry() 45 static void clear_shadow_entry(struct address_space *mapping, pgoff_t index, in clear_shadow_entry() argument 48 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry() 49 __clear_shadow_entry(mapping, index, entry); in clear_shadow_entry() 50 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry() 58 static void truncate_exceptional_pvec_entries(struct address_space *mapping, in truncate_exceptional_pvec_entries() argument 65 if (shmem_mapping(mapping)) in truncate_exceptional_pvec_entries() 75 dax = dax_mapping(mapping); in truncate_exceptional_pvec_entries() 77 xa_lock_irq(&mapping->i_pages); in truncate_exceptional_pvec_entries() [all …]
|
D | filemap.c | 124 static void page_cache_delete(struct address_space *mapping, in page_cache_delete() argument 127 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete() 130 mapping_set_update(&xas, mapping); in page_cache_delete() 145 page->mapping = NULL; in page_cache_delete() 147 mapping->nrpages -= nr; in page_cache_delete() 150 static void unaccount_page_cache_page(struct address_space *mapping, in unaccount_page_cache_page() argument 163 cleancache_invalidate_page(mapping, page); in unaccount_page_cache_page() 177 if (mapping_exiting(mapping) && in unaccount_page_cache_page() 203 filemap_nr_thps_dec(mapping); in unaccount_page_cache_page() 217 account_page_cleaned(page, mapping, inode_to_wb(mapping->host)); in unaccount_page_cache_page() [all …]
|
D | readahead.c | 34 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping) in file_ra_state_init() argument 36 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages; in file_ra_state_init() 48 static void read_cache_pages_invalidate_page(struct address_space *mapping, in read_cache_pages_invalidate_page() argument 54 page->mapping = mapping; in read_cache_pages_invalidate_page() 56 page->mapping = NULL; in read_cache_pages_invalidate_page() 65 static void read_cache_pages_invalidate_pages(struct address_space *mapping, in read_cache_pages_invalidate_pages() argument 73 read_cache_pages_invalidate_page(mapping, victim); in read_cache_pages_invalidate_pages() 89 int read_cache_pages(struct address_space *mapping, struct list_head *pages, in read_cache_pages() argument 98 if (add_to_page_cache_lru(page, mapping, page->index, in read_cache_pages() 99 readahead_gfp_mask(mapping))) { in read_cache_pages() [all …]
|
D | page-writeback.c | 1878 void balance_dirty_pages_ratelimited(struct address_space *mapping) in balance_dirty_pages_ratelimited() argument 1880 struct inode *inode = mapping->host; in balance_dirty_pages_ratelimited() 2124 void tag_pages_for_writeback(struct address_space *mapping, in tag_pages_for_writeback() argument 2127 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback() 2177 int write_cache_pages(struct address_space *mapping, in write_cache_pages() argument 2194 index = mapping->writeback_index; /* prev offset */ in write_cache_pages() 2203 tag_pages_for_writeback(mapping, index, end); in write_cache_pages() 2212 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2232 if (unlikely(page->mapping != mapping)) { in write_cache_pages() 2254 trace_wbc_writepage(wbc, inode_to_bdi(mapping->host)); in write_cache_pages() [all …]
|
/Linux-v5.15/drivers/gpu/drm/tegra/ |
D | uapi.c | 17 struct tegra_drm_mapping *mapping = in tegra_drm_mapping_release() local 20 if (mapping->sgt) in tegra_drm_mapping_release() 21 dma_unmap_sgtable(mapping->dev, mapping->sgt, mapping->direction, in tegra_drm_mapping_release() 24 host1x_bo_unpin(mapping->dev, mapping->bo, mapping->sgt); in tegra_drm_mapping_release() 25 host1x_bo_put(mapping->bo); in tegra_drm_mapping_release() 27 kfree(mapping); in tegra_drm_mapping_release() 30 void tegra_drm_mapping_put(struct tegra_drm_mapping *mapping) in tegra_drm_mapping_put() argument 32 kref_put(&mapping->ref, tegra_drm_mapping_release); in tegra_drm_mapping_put() 37 struct tegra_drm_mapping *mapping; in tegra_drm_channel_context_close() local 40 xa_for_each(&context->mappings, id, mapping) in tegra_drm_channel_context_close() [all …]
|
/Linux-v5.15/drivers/gpu/drm/panfrost/ |
D | panfrost_gem.c | 59 struct panfrost_gem_mapping *iter, *mapping = NULL; in panfrost_gem_mapping_get() local 65 mapping = iter; in panfrost_gem_mapping_get() 71 return mapping; in panfrost_gem_mapping_get() 75 panfrost_gem_teardown_mapping(struct panfrost_gem_mapping *mapping) in panfrost_gem_teardown_mapping() argument 77 if (mapping->active) in panfrost_gem_teardown_mapping() 78 panfrost_mmu_unmap(mapping); in panfrost_gem_teardown_mapping() 80 spin_lock(&mapping->mmu->mm_lock); in panfrost_gem_teardown_mapping() 81 if (drm_mm_node_allocated(&mapping->mmnode)) in panfrost_gem_teardown_mapping() 82 drm_mm_remove_node(&mapping->mmnode); in panfrost_gem_teardown_mapping() 83 spin_unlock(&mapping->mmu->mm_lock); in panfrost_gem_teardown_mapping() [all …]
|
/Linux-v5.15/include/linux/ |
D | pagemap.h | 21 static inline bool mapping_empty(struct address_space *mapping) in mapping_empty() argument 23 return xa_empty(&mapping->i_pages); in mapping_empty() 54 static inline void mapping_set_error(struct address_space *mapping, int error) in mapping_set_error() argument 60 __filemap_set_wb_err(mapping, error); in mapping_set_error() 63 if (mapping->host) in mapping_set_error() 64 errseq_set(&mapping->host->i_sb->s_wb_err, error); in mapping_set_error() 68 set_bit(AS_ENOSPC, &mapping->flags); in mapping_set_error() 70 set_bit(AS_EIO, &mapping->flags); in mapping_set_error() 73 static inline void mapping_set_unevictable(struct address_space *mapping) in mapping_set_unevictable() argument 75 set_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_set_unevictable() [all …]
|
D | io-mapping.h | 58 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument 60 iomap_free(mapping->base, mapping->size); in io_mapping_fini() 65 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument 70 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc() 71 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc() 74 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_atomic_wc() 86 io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) in io_mapping_map_local_wc() argument 90 BUG_ON(offset >= mapping->size); in io_mapping_map_local_wc() 91 phys_addr = mapping->base + offset; in io_mapping_map_local_wc() 92 return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); in io_mapping_map_local_wc() [all …]
|
D | tpm_eventlog.h | 166 void *mapping = NULL; in __calc_tpm2_event_size() local 186 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size() 188 if (!mapping) { in __calc_tpm2_event_size() 193 mapping = marker_start; in __calc_tpm2_event_size() 196 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size() 233 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size() 235 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size() 237 if (!mapping) { in __calc_tpm2_event_size() 242 mapping = marker; in __calc_tpm2_event_size() 245 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size() [all …]
|
D | secretmem.h | 11 struct address_space *mapping; in page_is_secretmem() local 23 mapping = (struct address_space *) in page_is_secretmem() 24 ((unsigned long)page->mapping & ~PAGE_MAPPING_FLAGS); in page_is_secretmem() 26 if (!mapping || mapping != page->mapping) in page_is_secretmem() 29 return mapping->a_ops == &secretmem_aops; in page_is_secretmem()
|
D | cleancache.h | 52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument 54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping() 58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled() 104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument 108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page() 109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page() 112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument 114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode() 115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
|
/Linux-v5.15/drivers/gpu/drm/exynos/ |
D | exynos_drm_dma.c | 67 ret = arm_iommu_attach_device(subdrv_dev, priv->mapping); in drm_iommu_attach_device() 69 ret = iommu_attach_device(priv->mapping, subdrv_dev); in drm_iommu_attach_device() 93 iommu_detach_device(priv->mapping, subdrv_dev); in drm_iommu_detach_device() 110 if (!priv->mapping) { in exynos_drm_register_dma() 111 void *mapping; in exynos_drm_register_dma() local 114 mapping = arm_iommu_create_mapping(&platform_bus_type, in exynos_drm_register_dma() 117 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma() 119 mapping = ERR_PTR(-ENODEV); in exynos_drm_register_dma() 121 if (IS_ERR(mapping)) in exynos_drm_register_dma() 122 return PTR_ERR(mapping); in exynos_drm_register_dma() [all …]
|
/Linux-v5.15/drivers/media/usb/uvc/ |
D | uvc_ctrl.c | 373 static s32 uvc_ctrl_get_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_get_zoom() argument 391 static void uvc_ctrl_set_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_set_zoom() argument 398 static s32 uvc_ctrl_get_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_get_rel_speed() argument 401 unsigned int first = mapping->offset / 8; in uvc_ctrl_get_rel_speed() 418 static void uvc_ctrl_set_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_set_rel_speed() argument 421 unsigned int first = mapping->offset / 8; in uvc_ctrl_set_rel_speed() 782 static s32 uvc_get_le_value(struct uvc_control_mapping *mapping, in uvc_get_le_value() argument 785 int bits = mapping->size; in uvc_get_le_value() 786 int offset = mapping->offset; in uvc_get_le_value() 807 if (mapping->data_type == UVC_CTRL_DATA_TYPE_SIGNED) in uvc_get_le_value() [all …]
|
/Linux-v5.15/fs/ |
D | dax.c | 342 static void dax_associate_entry(void *entry, struct address_space *mapping, in dax_associate_entry() argument 355 WARN_ON_ONCE(page->mapping); in dax_associate_entry() 356 page->mapping = mapping; in dax_associate_entry() 361 static void dax_disassociate_entry(void *entry, struct address_space *mapping, in dax_disassociate_entry() argument 373 WARN_ON_ONCE(page->mapping && page->mapping != mapping); in dax_disassociate_entry() 374 page->mapping = NULL; in dax_disassociate_entry() 408 struct address_space *mapping = READ_ONCE(page->mapping); in dax_lock_page() local 411 if (!mapping || !dax_mapping(mapping)) in dax_lock_page() 422 if (S_ISCHR(mapping->host->i_mode)) in dax_lock_page() 425 xas.xa = &mapping->i_pages; in dax_lock_page() [all …]
|
/Linux-v5.15/arch/arm/mm/ |
D | dma-mapping.c | 1101 static int extend_iommu_mapping(struct dma_iommu_mapping *mapping); 1103 static inline dma_addr_t __alloc_iova(struct dma_iommu_mapping *mapping, in __alloc_iova() argument 1109 size_t mapping_size = mapping->bits << PAGE_SHIFT; in __alloc_iova() 1120 spin_lock_irqsave(&mapping->lock, flags); in __alloc_iova() 1121 for (i = 0; i < mapping->nr_bitmaps; i++) { in __alloc_iova() 1122 start = bitmap_find_next_zero_area(mapping->bitmaps[i], in __alloc_iova() 1123 mapping->bits, 0, count, align); in __alloc_iova() 1125 if (start > mapping->bits) in __alloc_iova() 1128 bitmap_set(mapping->bitmaps[i], start, count); in __alloc_iova() 1137 if (i == mapping->nr_bitmaps) { in __alloc_iova() [all …]
|
D | flush.c | 199 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 232 if (mapping && cache_is_vipt_aliasing()) in __flush_dcache_page() 237 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument 251 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases() 252 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in __flush_dcache_aliases() 265 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases() 273 struct address_space *mapping; in __sync_icache_dcache() local 284 mapping = page_mapping_file(page); in __sync_icache_dcache() 286 mapping = NULL; in __sync_icache_dcache() 289 __flush_dcache_page(mapping, page); in __sync_icache_dcache() [all …]
|
/Linux-v5.15/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 218 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_get_vram_mapping() local 220 list_for_each_entry(mapping, &obj->vram_list, obj_node) { in etnaviv_gem_get_vram_mapping() 221 if (mapping->context == context) in etnaviv_gem_get_vram_mapping() 222 return mapping; in etnaviv_gem_get_vram_mapping() 228 void etnaviv_gem_mapping_unreference(struct etnaviv_vram_mapping *mapping) in etnaviv_gem_mapping_unreference() argument 230 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_gem_mapping_unreference() 233 WARN_ON(mapping->use == 0); in etnaviv_gem_mapping_unreference() 234 mapping->use -= 1; in etnaviv_gem_mapping_unreference() 245 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_mapping_get() local 250 mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, mmu_context); in etnaviv_gem_mapping_get() [all …]
|
D | etnaviv_mmu.c | 123 struct etnaviv_vram_mapping *mapping) in etnaviv_iommu_remove_mapping() argument 125 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_iommu_remove_mapping() 129 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping() 131 drm_mm_remove_node(&mapping->vram_node); in etnaviv_iommu_remove_mapping() 230 struct etnaviv_vram_mapping *mapping, u64 va) in etnaviv_iommu_map_gem() argument 247 mapping->iova = iova; in etnaviv_iommu_map_gem() 248 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() 254 node = &mapping->vram_node; in etnaviv_iommu_map_gem() 265 mapping->iova = node->start; in etnaviv_iommu_map_gem() 274 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() [all …]
|
/Linux-v5.15/arch/nios2/mm/ |
D | cacheflush.c | 74 static void flush_aliases(struct address_space *mapping, struct page *page) in flush_aliases() argument 82 flush_dcache_mmap_lock(mapping); in flush_aliases() 83 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in flush_aliases() 95 flush_dcache_mmap_unlock(mapping); in flush_aliases() 161 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 175 struct address_space *mapping; in flush_dcache_page() local 184 mapping = page_mapping_file(page); in flush_dcache_page() 187 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_page() 190 __flush_dcache_page(mapping, page); in flush_dcache_page() 191 if (mapping) { in flush_dcache_page() [all …]
|
/Linux-v5.15/drivers/net/wireless/marvell/mwifiex/ |
D | util.h | 69 struct mwifiex_dma_mapping *mapping) in mwifiex_store_mapping() argument 73 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 77 struct mwifiex_dma_mapping *mapping) in mwifiex_get_mapping() argument 81 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping() 86 struct mwifiex_dma_mapping mapping; in MWIFIEX_SKB_DMA_ADDR() local 88 mwifiex_get_mapping(skb, &mapping); in MWIFIEX_SKB_DMA_ADDR() 90 return mapping.addr; in MWIFIEX_SKB_DMA_ADDR()
|
/Linux-v5.15/fs/gfs2/ |
D | aops.c | 91 struct inode *inode = page->mapping->host; in gfs2_writepage() 120 struct inode * const inode = page->mapping->host; in gfs2_write_jdata_page() 153 struct inode *inode = page->mapping->host; in __gfs2_jdata_writepage() 179 struct inode *inode = page->mapping->host; in gfs2_jdata_writepage() 203 static int gfs2_writepages(struct address_space *mapping, in gfs2_writepages() argument 206 struct gfs2_sbd *sdp = gfs2_mapping2sbd(mapping); in gfs2_writepages() 216 ret = iomap_writepages(mapping, wbc, &wpc, &gfs2_writeback_ops); in gfs2_writepages() 233 static int gfs2_write_jdata_pagevec(struct address_space *mapping, in gfs2_write_jdata_pagevec() argument 239 struct inode *inode = mapping->host; in gfs2_write_jdata_pagevec() 256 if (unlikely(page->mapping != mapping)) { in gfs2_write_jdata_pagevec() [all …]
|
/Linux-v5.15/drivers/sh/clk/ |
D | core.c | 340 struct clk_mapping *mapping = clk->mapping; in clk_establish_mapping() local 345 if (!mapping) { in clk_establish_mapping() 352 clk->mapping = &dummy_mapping; in clk_establish_mapping() 361 mapping = clkp->mapping; in clk_establish_mapping() 362 BUG_ON(!mapping); in clk_establish_mapping() 368 if (!mapping->base && mapping->phys) { in clk_establish_mapping() 369 kref_init(&mapping->ref); in clk_establish_mapping() 371 mapping->base = ioremap(mapping->phys, mapping->len); in clk_establish_mapping() 372 if (unlikely(!mapping->base)) in clk_establish_mapping() 374 } else if (mapping->base) { in clk_establish_mapping() [all …]
|
/Linux-v5.15/Documentation/driver-api/ |
D | io-mapping.rst | 8 The io_mapping functions in linux/io-mapping.h provide an abstraction for 9 efficiently mapping small regions of an I/O device to the CPU. The initial 14 A mapping object is created during driver initialization using:: 20 mappable, while 'size' indicates how large a mapping region to 23 This _wc variant provides a mapping which may only be used with 27 With this mapping object, individual pages can be mapped either temporarily 31 void *io_mapping_map_local_wc(struct io_mapping *mapping, 34 void *io_mapping_map_atomic_wc(struct io_mapping *mapping, 37 'offset' is the offset within the defined mapping region. Accessing 46 Temporary mappings are only valid in the context of the caller. The mapping [all …]
|
/Linux-v5.15/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_xdp.c | 25 dma_addr_t mapping, u32 len) in bnxt_xmit_bd() argument 40 txbd->tx_bd_haddr = cpu_to_le64(mapping); in bnxt_xmit_bd() 48 dma_addr_t mapping, u32 len, u16 rx_prod) in __bnxt_xmit_xdp() argument 52 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp() 59 dma_addr_t mapping, u32 len, in __bnxt_xmit_xdp_redirect() argument 64 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp_redirect() 67 dma_unmap_addr_set(tx_buf, mapping, mapping); in __bnxt_xmit_xdp_redirect() 88 dma_unmap_addr(tx_buf, mapping), in bnxt_tx_int_xdp() 119 dma_addr_t mapping; in bnxt_rx_xdp() local 132 mapping = rx_buf->mapping - bp->rx_dma_offset; in bnxt_rx_xdp() [all …]
|
/Linux-v5.15/fs/afs/ |
D | write.c | 30 int afs_write_begin(struct file *file, struct address_space *mapping, in afs_write_begin() argument 49 ret = netfs_write_begin(file, mapping, pos, len, flags, &page, fsdata, in afs_write_begin() 108 int afs_write_end(struct file *file, struct address_space *mapping, in afs_write_end() argument 173 static void afs_kill_pages(struct address_space *mapping, in afs_kill_pages() argument 176 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_kill_pages() 188 pv.nr = find_get_pages_contig(mapping, start / PAGE_SIZE, in afs_kill_pages() 205 generic_error_remove_page(mapping, page); in afs_kill_pages() 219 struct address_space *mapping, in afs_redirty_pages() argument 222 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_redirty_pages() 234 pv.nr = find_get_pages_contig(mapping, start / PAGE_SIZE, in afs_redirty_pages() [all …]
|