Lines Matching refs:size_t

15 	void *(*alloc)(struct device *dev, size_t size,
18 void (*free)(struct device *dev, size_t size, void *vaddr,
20 struct page *(*alloc_pages)(struct device *dev, size_t size,
23 void (*free_pages)(struct device *dev, size_t size, struct page *vaddr,
25 void *(*alloc_noncoherent)(struct device *dev, size_t size,
28 void (*free_noncoherent)(struct device *dev, size_t size, void *vaddr,
31 void *, dma_addr_t, size_t, unsigned long attrs);
34 void *cpu_addr, dma_addr_t dma_addr, size_t size,
38 unsigned long offset, size_t size,
41 size_t size, enum dma_data_direction dir,
52 size_t size, enum dma_data_direction dir,
55 size_t size, enum dma_data_direction dir,
58 size_t size, enum dma_data_direction dir);
60 dma_addr_t dma_handle, size_t size,
66 void (*cache_sync)(struct device *dev, void *vaddr, size_t size,
70 size_t (*max_mapping_size)(struct device *dev);
114 struct page *dma_alloc_from_contiguous(struct device *dev, size_t count,
118 struct page *dma_alloc_contiguous(struct device *dev, size_t size, gfp_t gfp);
119 void dma_free_contiguous(struct device *dev, struct page *page, size_t size);
137 size_t count, unsigned int order, bool no_warn) in dma_alloc_from_contiguous()
147 static inline struct page *dma_alloc_contiguous(struct device *dev, size_t size, in dma_alloc_contiguous()
153 size_t size) in dma_free_contiguous()
167 dma_addr_t device_addr, size_t size);
172 void *cpu_addr, size_t size, int *ret);
178 size_t size, int *ret);
182 phys_addr_t phys_addr, dma_addr_t device_addr, size_t size) in dma_declare_coherent_memory()
200 void *cpu_addr, size_t size, int *ret) in dma_mmap_from_global_coherent()
207 void *cpu_addr, dma_addr_t dma_addr, size_t size,
210 void *cpu_addr, dma_addr_t dma_addr, size_t size,
212 struct page *dma_common_alloc_pages(struct device *dev, size_t size,
214 void dma_common_free_pages(struct device *dev, size_t size, struct page *vaddr,
218 void *dma_common_contiguous_remap(struct page *page, size_t size, pgprot_t prot,
220 void *dma_common_pages_remap(struct page **pages, size_t size, pgprot_t prot,
222 void dma_common_free_remap(void *cpu_addr, size_t size);
224 struct page *dma_alloc_from_pool(struct device *dev, size_t size,
226 bool (*phys_addr_ok)(struct device *, phys_addr_t, size_t));
227 bool dma_free_from_pool(struct device *dev, void *start, size_t size);
245 void *arch_dma_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
247 void arch_dma_free(struct device *dev, size_t size, void *cpu_addr,
271 void arch_sync_dma_for_device(phys_addr_t paddr, size_t size,
274 static inline void arch_sync_dma_for_device(phys_addr_t paddr, size_t size, in arch_sync_dma_for_device()
281 void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size,
284 static inline void arch_sync_dma_for_cpu(phys_addr_t paddr, size_t size, in arch_sync_dma_for_cpu()
299 void arch_dma_prep_coherent(struct page *page, size_t size);
301 static inline void arch_dma_prep_coherent(struct page *page, size_t size) in arch_dma_prep_coherent()
307 void arch_dma_mark_clean(phys_addr_t paddr, size_t size);
309 static inline void arch_dma_mark_clean(phys_addr_t paddr, size_t size) in arch_dma_mark_clean()
314 void *arch_dma_set_uncached(void *addr, size_t size);
315 void arch_dma_clear_uncached(void *addr, size_t size);