/Linux-v5.4/drivers/staging/gasket/ |
D | gasket_sysfs.c | 77 static void put_mapping(struct gasket_sysfs_mapping *mapping) in put_mapping() argument 84 if (!mapping) { in put_mapping() 89 mutex_lock(&mapping->mutex); in put_mapping() 90 if (kref_put(&mapping->refcount, release_entry)) { in put_mapping() 91 dev_dbg(mapping->device, "Removing Gasket sysfs mapping\n"); in put_mapping() 99 device = mapping->device; in put_mapping() 100 num_files_to_remove = mapping->attribute_count; in put_mapping() 107 mapping->attributes[i].attr; in put_mapping() 111 kfree(mapping->attributes); in put_mapping() 112 mapping->attributes = NULL; in put_mapping() [all …]
|
/Linux-v5.4/mm/ |
D | truncate.c | 34 static inline void __clear_shadow_entry(struct address_space *mapping, in __clear_shadow_entry() argument 37 XA_STATE(xas, &mapping->i_pages, index); in __clear_shadow_entry() 43 mapping->nrexceptional--; in __clear_shadow_entry() 46 static void clear_shadow_entry(struct address_space *mapping, pgoff_t index, in clear_shadow_entry() argument 49 xa_lock_irq(&mapping->i_pages); in clear_shadow_entry() 50 __clear_shadow_entry(mapping, index, entry); in clear_shadow_entry() 51 xa_unlock_irq(&mapping->i_pages); in clear_shadow_entry() 59 static void truncate_exceptional_pvec_entries(struct address_space *mapping, in truncate_exceptional_pvec_entries() argument 67 if (shmem_mapping(mapping)) in truncate_exceptional_pvec_entries() 77 dax = dax_mapping(mapping); in truncate_exceptional_pvec_entries() [all …]
|
D | filemap.c | 119 static void page_cache_delete(struct address_space *mapping, in page_cache_delete() argument 122 XA_STATE(xas, &mapping->i_pages, page->index); in page_cache_delete() 125 mapping_set_update(&xas, mapping); in page_cache_delete() 140 page->mapping = NULL; in page_cache_delete() 144 mapping->nrexceptional += nr; in page_cache_delete() 153 mapping->nrpages -= nr; in page_cache_delete() 156 static void unaccount_page_cache_page(struct address_space *mapping, in unaccount_page_cache_page() argument 169 cleancache_invalidate_page(mapping, page); in unaccount_page_cache_page() 183 if (mapping_exiting(mapping) && in unaccount_page_cache_page() 209 filemap_nr_thps_dec(mapping); in unaccount_page_cache_page() [all …]
|
D | readahead.c | 33 file_ra_state_init(struct file_ra_state *ra, struct address_space *mapping) in file_ra_state_init() argument 35 ra->ra_pages = inode_to_bdi(mapping->host)->ra_pages; in file_ra_state_init() 47 static void read_cache_pages_invalidate_page(struct address_space *mapping, in read_cache_pages_invalidate_page() argument 53 page->mapping = mapping; in read_cache_pages_invalidate_page() 55 page->mapping = NULL; in read_cache_pages_invalidate_page() 64 static void read_cache_pages_invalidate_pages(struct address_space *mapping, in read_cache_pages_invalidate_pages() argument 72 read_cache_pages_invalidate_page(mapping, victim); in read_cache_pages_invalidate_pages() 88 int read_cache_pages(struct address_space *mapping, struct list_head *pages, in read_cache_pages() argument 97 if (add_to_page_cache_lru(page, mapping, page->index, in read_cache_pages() 98 readahead_gfp_mask(mapping))) { in read_cache_pages() [all …]
|
D | page-writeback.c | 1864 void balance_dirty_pages_ratelimited(struct address_space *mapping) in balance_dirty_pages_ratelimited() argument 1866 struct inode *inode = mapping->host; in balance_dirty_pages_ratelimited() 2105 void tag_pages_for_writeback(struct address_space *mapping, in tag_pages_for_writeback() argument 2108 XA_STATE(xas, &mapping->i_pages, start); in tag_pages_for_writeback() 2158 int write_cache_pages(struct address_space *mapping, in write_cache_pages() argument 2176 writeback_index = mapping->writeback_index; /* prev offset */ in write_cache_pages() 2190 tag_pages_for_writeback(mapping, index, end); in write_cache_pages() 2195 nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, in write_cache_pages() 2215 if (unlikely(page->mapping != mapping)) { in write_cache_pages() 2237 trace_wbc_writepage(wbc, inode_to_bdi(mapping->host)); in write_cache_pages() [all …]
|
/Linux-v5.4/include/linux/ |
D | pagemap.h | 48 static inline void mapping_set_error(struct address_space *mapping, int error) in mapping_set_error() argument 54 filemap_set_wb_err(mapping, error); in mapping_set_error() 58 set_bit(AS_ENOSPC, &mapping->flags); in mapping_set_error() 60 set_bit(AS_EIO, &mapping->flags); in mapping_set_error() 63 static inline void mapping_set_unevictable(struct address_space *mapping) in mapping_set_unevictable() argument 65 set_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_set_unevictable() 68 static inline void mapping_clear_unevictable(struct address_space *mapping) in mapping_clear_unevictable() argument 70 clear_bit(AS_UNEVICTABLE, &mapping->flags); in mapping_clear_unevictable() 73 static inline int mapping_unevictable(struct address_space *mapping) in mapping_unevictable() argument 75 if (mapping) in mapping_unevictable() [all …]
|
D | io-mapping.h | 56 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument 58 iomap_free(mapping->base, mapping->size); in io_mapping_fini() 63 io_mapping_map_atomic_wc(struct io_mapping *mapping, in io_mapping_map_atomic_wc() argument 69 BUG_ON(offset >= mapping->size); in io_mapping_map_atomic_wc() 70 phys_addr = mapping->base + offset; in io_mapping_map_atomic_wc() 72 return iomap_atomic_prot_pfn(pfn, mapping->prot); in io_mapping_map_atomic_wc() 82 io_mapping_map_wc(struct io_mapping *mapping, in io_mapping_map_wc() argument 88 BUG_ON(offset >= mapping->size); in io_mapping_map_wc() 89 phys_addr = mapping->base + offset; in io_mapping_map_wc() 126 io_mapping_fini(struct io_mapping *mapping) in io_mapping_fini() argument [all …]
|
D | tpm_eventlog.h | 164 void *mapping = NULL; in __calc_tpm2_event_size() local 183 mapping = TPM_MEMREMAP((unsigned long)marker_start, in __calc_tpm2_event_size() 185 if (!mapping) { in __calc_tpm2_event_size() 190 mapping = marker_start; in __calc_tpm2_event_size() 193 event = (struct tcg_pcr_event2_head *)mapping; in __calc_tpm2_event_size() 214 TPM_MEMUNMAP(mapping, mapping_size); in __calc_tpm2_event_size() 216 mapping = TPM_MEMREMAP((unsigned long)marker, in __calc_tpm2_event_size() 218 if (!mapping) { in __calc_tpm2_event_size() 223 mapping = marker; in __calc_tpm2_event_size() 226 memcpy(&halg, mapping, halg_size); in __calc_tpm2_event_size() [all …]
|
D | cleancache.h | 52 static inline bool cleancache_fs_enabled_mapping(struct address_space *mapping) in cleancache_fs_enabled_mapping() argument 54 return mapping->host->i_sb->cleancache_poolid >= 0; in cleancache_fs_enabled_mapping() 58 return cleancache_fs_enabled_mapping(page->mapping); in cleancache_fs_enabled() 104 static inline void cleancache_invalidate_page(struct address_space *mapping, in cleancache_invalidate_page() argument 108 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_page() 109 __cleancache_invalidate_page(mapping, page); in cleancache_invalidate_page() 112 static inline void cleancache_invalidate_inode(struct address_space *mapping) in cleancache_invalidate_inode() argument 114 if (cleancache_enabled && cleancache_fs_enabled_mapping(mapping)) in cleancache_invalidate_inode() 115 __cleancache_invalidate_inode(mapping); in cleancache_invalidate_inode()
|
/Linux-v5.4/drivers/media/usb/uvc/ |
D | uvc_ctrl.c | 365 static s32 uvc_ctrl_get_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_get_zoom() argument 383 static void uvc_ctrl_set_zoom(struct uvc_control_mapping *mapping, in uvc_ctrl_set_zoom() argument 390 static s32 uvc_ctrl_get_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_get_rel_speed() argument 393 unsigned int first = mapping->offset / 8; in uvc_ctrl_get_rel_speed() 410 static void uvc_ctrl_set_rel_speed(struct uvc_control_mapping *mapping, in uvc_ctrl_set_rel_speed() argument 413 unsigned int first = mapping->offset / 8; in uvc_ctrl_set_rel_speed() 764 static s32 uvc_get_le_value(struct uvc_control_mapping *mapping, in uvc_get_le_value() argument 767 int bits = mapping->size; in uvc_get_le_value() 768 int offset = mapping->offset; in uvc_get_le_value() 785 if (mapping->data_type == UVC_CTRL_DATA_TYPE_SIGNED) in uvc_get_le_value() [all …]
|
/Linux-v5.4/drivers/gpu/drm/exynos/ |
D | exynos_drm_dma.c | 80 ret = arm_iommu_attach_device(subdrv_dev, priv->mapping); in drm_iommu_attach_device() 82 ret = iommu_attach_device(priv->mapping, subdrv_dev); in drm_iommu_attach_device() 108 iommu_detach_device(priv->mapping, subdrv_dev); in drm_iommu_detach_device() 126 if (!priv->mapping) { in exynos_drm_register_dma() 127 void *mapping; in exynos_drm_register_dma() local 130 mapping = arm_iommu_create_mapping(&platform_bus_type, in exynos_drm_register_dma() 133 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma() 135 if (IS_ERR(mapping)) in exynos_drm_register_dma() 136 return PTR_ERR(mapping); in exynos_drm_register_dma() 137 priv->mapping = mapping; in exynos_drm_register_dma() [all …]
|
/Linux-v5.4/fs/ |
D | dax.c | 331 static void dax_associate_entry(void *entry, struct address_space *mapping, in dax_associate_entry() argument 344 WARN_ON_ONCE(page->mapping); in dax_associate_entry() 345 page->mapping = mapping; in dax_associate_entry() 350 static void dax_disassociate_entry(void *entry, struct address_space *mapping, in dax_disassociate_entry() argument 362 WARN_ON_ONCE(page->mapping && page->mapping != mapping); in dax_disassociate_entry() 363 page->mapping = NULL; in dax_disassociate_entry() 397 struct address_space *mapping = READ_ONCE(page->mapping); in dax_lock_page() local 400 if (!mapping || !dax_mapping(mapping)) in dax_lock_page() 411 if (S_ISCHR(mapping->host->i_mode)) in dax_lock_page() 414 xas.xa = &mapping->i_pages; in dax_lock_page() [all …]
|
/Linux-v5.4/arch/arm/mm/ |
D | flush.c | 199 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 232 if (mapping && cache_is_vipt_aliasing()) in __flush_dcache_page() 237 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) in __flush_dcache_aliases() argument 251 flush_dcache_mmap_lock(mapping); in __flush_dcache_aliases() 252 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in __flush_dcache_aliases() 265 flush_dcache_mmap_unlock(mapping); in __flush_dcache_aliases() 273 struct address_space *mapping; in __sync_icache_dcache() local 284 mapping = page_mapping_file(page); in __sync_icache_dcache() 286 mapping = NULL; in __sync_icache_dcache() 289 __flush_dcache_page(mapping, page); in __sync_icache_dcache() [all …]
|
D | dma-mapping.c | 1138 static int extend_iommu_mapping(struct dma_iommu_mapping *mapping); 1140 static inline dma_addr_t __alloc_iova(struct dma_iommu_mapping *mapping, in __alloc_iova() argument 1146 size_t mapping_size = mapping->bits << PAGE_SHIFT; in __alloc_iova() 1157 spin_lock_irqsave(&mapping->lock, flags); in __alloc_iova() 1158 for (i = 0; i < mapping->nr_bitmaps; i++) { in __alloc_iova() 1159 start = bitmap_find_next_zero_area(mapping->bitmaps[i], in __alloc_iova() 1160 mapping->bits, 0, count, align); in __alloc_iova() 1162 if (start > mapping->bits) in __alloc_iova() 1165 bitmap_set(mapping->bitmaps[i], start, count); in __alloc_iova() 1174 if (i == mapping->nr_bitmaps) { in __alloc_iova() [all …]
|
/Linux-v5.4/fs/gfs2/ |
D | aops.c | 91 struct inode *inode = page->mapping->host; in gfs2_writepage() 105 page->mapping->a_ops->invalidatepage(page, 0, PAGE_SIZE); in gfs2_writepage() 124 struct inode * const inode = page->mapping->host; in gfs2_write_full_page() 157 struct inode *inode = page->mapping->host; in __gfs2_jdata_writepage() 183 struct inode *inode = page->mapping->host; in gfs2_jdata_writepage() 209 static int gfs2_writepages(struct address_space *mapping, in gfs2_writepages() argument 212 struct gfs2_sbd *sdp = gfs2_mapping2sbd(mapping); in gfs2_writepages() 213 int ret = mpage_writepages(mapping, wbc, gfs2_get_block_noalloc); in gfs2_writepages() 238 static int gfs2_write_jdata_pagevec(struct address_space *mapping, in gfs2_write_jdata_pagevec() argument 244 struct inode *inode = mapping->host; in gfs2_write_jdata_pagevec() [all …]
|
/Linux-v5.4/arch/nios2/mm/ |
D | cacheflush.c | 73 static void flush_aliases(struct address_space *mapping, struct page *page) in flush_aliases() argument 81 flush_dcache_mmap_lock(mapping); in flush_aliases() 82 vma_interval_tree_foreach(mpnt, &mapping->i_mmap, pgoff, pgoff) { in flush_aliases() 94 flush_dcache_mmap_unlock(mapping); in flush_aliases() 160 void __flush_dcache_page(struct address_space *mapping, struct page *page) in __flush_dcache_page() argument 174 struct address_space *mapping; in flush_dcache_page() local 183 mapping = page_mapping_file(page); in flush_dcache_page() 186 if (mapping && !mapping_mapped(mapping)) { in flush_dcache_page() 189 __flush_dcache_page(mapping, page); in flush_dcache_page() 190 if (mapping) { in flush_dcache_page() [all …]
|
/Linux-v5.4/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 228 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_get_vram_mapping() local 230 list_for_each_entry(mapping, &obj->vram_list, obj_node) { in etnaviv_gem_get_vram_mapping() 231 if (mapping->context == context) in etnaviv_gem_get_vram_mapping() 232 return mapping; in etnaviv_gem_get_vram_mapping() 238 void etnaviv_gem_mapping_unreference(struct etnaviv_vram_mapping *mapping) in etnaviv_gem_mapping_unreference() argument 240 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_gem_mapping_unreference() 243 WARN_ON(mapping->use == 0); in etnaviv_gem_mapping_unreference() 244 mapping->use -= 1; in etnaviv_gem_mapping_unreference() 255 struct etnaviv_vram_mapping *mapping; in etnaviv_gem_mapping_get() local 260 mapping = etnaviv_gem_get_vram_mapping(etnaviv_obj, mmu_context); in etnaviv_gem_mapping_get() [all …]
|
D | etnaviv_mmu.c | 130 struct etnaviv_vram_mapping *mapping) in etnaviv_iommu_remove_mapping() argument 132 struct etnaviv_gem_object *etnaviv_obj = mapping->object; in etnaviv_iommu_remove_mapping() 134 etnaviv_iommu_unmap(context, mapping->vram_node.start, in etnaviv_iommu_remove_mapping() 136 drm_mm_remove_node(&mapping->vram_node); in etnaviv_iommu_remove_mapping() 232 struct etnaviv_vram_mapping *mapping, u64 va) in etnaviv_iommu_map_gem() argument 249 mapping->iova = iova; in etnaviv_iommu_map_gem() 250 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() 256 node = &mapping->vram_node; in etnaviv_iommu_map_gem() 267 mapping->iova = node->start; in etnaviv_iommu_map_gem() 276 list_add_tail(&mapping->mmu_node, &context->mappings); in etnaviv_iommu_map_gem() [all …]
|
/Linux-v5.4/arch/csky/abiv1/ |
D | cacheflush.c | 18 struct address_space *mapping; in flush_dcache_page() local 23 mapping = page_mapping_file(page); in flush_dcache_page() 25 if (mapping && !page_mapcount(page)) in flush_dcache_page() 29 if (mapping) in flush_dcache_page() 60 struct address_space *mapping; in flush_kernel_dcache_page() local 62 mapping = page_mapping_file(page); in flush_kernel_dcache_page() 64 if (!mapping || mapping_mapped(mapping)) in flush_kernel_dcache_page()
|
/Linux-v5.4/drivers/net/wireless/marvell/mwifiex/ |
D | util.h | 69 struct mwifiex_dma_mapping *mapping) in mwifiex_store_mapping() argument 73 memcpy(&cb->dma_mapping, mapping, sizeof(*mapping)); in mwifiex_store_mapping() 77 struct mwifiex_dma_mapping *mapping) in mwifiex_get_mapping() argument 81 memcpy(mapping, &cb->dma_mapping, sizeof(*mapping)); in mwifiex_get_mapping() 86 struct mwifiex_dma_mapping mapping; in MWIFIEX_SKB_DMA_ADDR() local 88 mwifiex_get_mapping(skb, &mapping); in MWIFIEX_SKB_DMA_ADDR() 90 return mapping.addr; in MWIFIEX_SKB_DMA_ADDR()
|
/Linux-v5.4/drivers/sh/clk/ |
D | core.c | 340 struct clk_mapping *mapping = clk->mapping; in clk_establish_mapping() local 345 if (!mapping) { in clk_establish_mapping() 352 clk->mapping = &dummy_mapping; in clk_establish_mapping() 361 mapping = clkp->mapping; in clk_establish_mapping() 362 BUG_ON(!mapping); in clk_establish_mapping() 368 if (!mapping->base && mapping->phys) { in clk_establish_mapping() 369 kref_init(&mapping->ref); in clk_establish_mapping() 371 mapping->base = ioremap_nocache(mapping->phys, mapping->len); in clk_establish_mapping() 372 if (unlikely(!mapping->base)) in clk_establish_mapping() 374 } else if (mapping->base) { in clk_establish_mapping() [all …]
|
/Linux-v5.4/fs/afs/ |
D | write.c | 77 int afs_write_begin(struct file *file, struct address_space *mapping, in afs_write_begin() argument 98 page = grab_cache_page_write_begin(mapping, index, flags); in afs_write_begin() 181 int afs_write_end(struct file *file, struct address_space *mapping, in afs_write_end() argument 232 static void afs_kill_pages(struct address_space *mapping, in afs_kill_pages() argument 235 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_kill_pages() 250 pv.nr = find_get_pages_contig(mapping, first, count, pv.pages); in afs_kill_pages() 261 generic_error_remove_page(mapping, page); in afs_kill_pages() 275 struct address_space *mapping, in afs_redirty_pages() argument 278 struct afs_vnode *vnode = AFS_FS_I(mapping->host); in afs_redirty_pages() 293 pv.nr = find_get_pages_contig(mapping, first, count, pv.pages); in afs_redirty_pages() [all …]
|
/Linux-v5.4/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_xdp.c | 25 dma_addr_t mapping, u32 len) in bnxt_xmit_bd() argument 40 txbd->tx_bd_haddr = cpu_to_le64(mapping); in bnxt_xmit_bd() 48 dma_addr_t mapping, u32 len, u16 rx_prod) in __bnxt_xmit_xdp() argument 52 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp() 59 dma_addr_t mapping, u32 len, in __bnxt_xmit_xdp_redirect() argument 64 tx_buf = bnxt_xmit_bd(bp, txr, mapping, len); in __bnxt_xmit_xdp_redirect() 67 dma_unmap_addr_set(tx_buf, mapping, mapping); in __bnxt_xmit_xdp_redirect() 88 dma_unmap_addr(tx_buf, mapping), in bnxt_tx_int_xdp() 119 dma_addr_t mapping; in bnxt_rx_xdp() local 132 mapping = rx_buf->mapping - bp->rx_dma_offset; in bnxt_rx_xdp() [all …]
|
/Linux-v5.4/fs/hpfs/ |
D | file.c | 128 static int hpfs_readpages(struct file *file, struct address_space *mapping, in hpfs_readpages() argument 131 return mpage_readpages(mapping, pages, nr_pages, hpfs_get_block); in hpfs_readpages() 134 static int hpfs_writepages(struct address_space *mapping, in hpfs_writepages() argument 137 return mpage_writepages(mapping, wbc, hpfs_get_block); in hpfs_writepages() 140 static void hpfs_write_failed(struct address_space *mapping, loff_t to) in hpfs_write_failed() argument 142 struct inode *inode = mapping->host; in hpfs_write_failed() 154 static int hpfs_write_begin(struct file *file, struct address_space *mapping, in hpfs_write_begin() argument 161 ret = cont_write_begin(file, mapping, pos, len, flags, pagep, fsdata, in hpfs_write_begin() 163 &hpfs_i(mapping->host)->mmu_private); in hpfs_write_begin() 165 hpfs_write_failed(mapping, pos + len); in hpfs_write_begin() [all …]
|
/Linux-v5.4/fs/9p/ |
D | vfs_addr.c | 41 struct inode *inode = page->mapping->host; in v9fs_fid_readpage() 98 static int v9fs_vfs_readpages(struct file *filp, struct address_space *mapping, in v9fs_vfs_readpages() argument 104 inode = mapping->host; in v9fs_vfs_readpages() 107 ret = v9fs_readpages_from_fscache(inode, mapping, pages, &nr_pages); in v9fs_vfs_readpages() 111 ret = read_cache_pages(mapping, pages, v9fs_fid_readpage, in v9fs_vfs_readpages() 150 struct inode *inode = page->mapping->host; in v9fs_vfs_writepage_locked() 191 mapping_set_error(page->mapping, retval); in v9fs_vfs_writepage() 208 struct inode *inode = page->mapping->host; in v9fs_launder_page() 255 static int v9fs_write_begin(struct file *filp, struct address_space *mapping, in v9fs_write_begin() argument 263 struct inode *inode = mapping->host; in v9fs_write_begin() [all …]
|