Lines Matching refs:size_t

26 	void *(*alloc)(struct device *dev, size_t size,
29 void (*free)(struct device *dev, size_t size, void *vaddr,
31 struct page *(*alloc_pages)(struct device *dev, size_t size,
34 void (*free_pages)(struct device *dev, size_t size, struct page *vaddr,
36 struct sg_table *(*alloc_noncontiguous)(struct device *dev, size_t size,
39 void (*free_noncontiguous)(struct device *dev, size_t size,
42 void *, dma_addr_t, size_t, unsigned long attrs);
45 void *cpu_addr, dma_addr_t dma_addr, size_t size,
49 unsigned long offset, size_t size,
52 size_t size, enum dma_data_direction dir,
64 size_t size, enum dma_data_direction dir,
67 size_t size, enum dma_data_direction dir,
70 size_t size, enum dma_data_direction dir);
72 dma_addr_t dma_handle, size_t size,
78 void (*cache_sync)(struct device *dev, void *vaddr, size_t size,
82 size_t (*max_mapping_size)(struct device *dev);
83 size_t (*opt_mapping_size)(void);
127 struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
131 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
132 void dma_free_contiguous(struct device *dev, struct page *page, size_t size);
150 size_t count, unsigned int order, bool no_warn) in dma_alloc_from_contiguous()
160 static inline struct page *dma_alloc_contiguous(struct device *dev, size_t size, in dma_alloc_contiguous()
166 size_t size) in dma_free_contiguous()
174 dma_addr_t device_addr, size_t size);
180 void *cpu_addr, size_t size, int *ret);
183 phys_addr_t phys_addr, dma_addr_t device_addr, size_t size) in dma_declare_coherent_memory()
199 size_t size, int *ret);
200 int dma_init_global_coherent(phys_addr_t phys_addr, size_t size);
212 void *cpu_addr, size_t size, int *ret) in dma_mmap_from_global_coherent()
233 void *cpu_addr, dma_addr_t dma_addr, size_t size,
236 void *cpu_addr, dma_addr_t dma_addr, size_t size,
238 struct page *dma_common_alloc_pages(struct device *dev, size_t size,
240 void dma_common_free_pages(struct device *dev, size_t size, struct page *vaddr,
244 void *dma_common_contiguous_remap(struct page *page, size_t size, pgprot_t prot,
246 void *dma_common_pages_remap(struct page **pages, size_t size, pgprot_t prot,
248 void dma_common_free_remap(void *cpu_addr, size_t size);
250 struct page *dma_alloc_from_pool(struct device *dev, size_t size,
252 bool (*phys_addr_ok)(struct device *, phys_addr_t, size_t));
253 bool dma_free_from_pool(struct device *dev, void *start, size_t size);
303 static inline bool dma_kmalloc_size_aligned(size_t size) in dma_kmalloc_size_aligned()
329 static inline bool dma_kmalloc_needs_bounce(struct device *dev, size_t size, in dma_kmalloc_needs_bounce()
335 void *arch_dma_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
337 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
367 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size,
370 static inline void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device()
377 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size,
380 static inline void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu()
395 void arch_dma_prep_coherent(struct page *page, size_t size);
397 static inline void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent()
403 void arch_dma_mark_clean(phys_addr_t paddr, size_t size);
405 static inline void arch_dma_mark_clean(phys_addr_t paddr, size_t size) in arch_dma_mark_clean()
410 void *arch_dma_set_uncached(void *addr, size_t size);
411 void arch_dma_clear_uncached(void *addr, size_t size);