/Linux-v5.4/drivers/gpu/drm/i915/ |
D | i915_scatterlist.c | 18 if (sg_alloc_table(&new_st, orig_st->nents, GFP_KERNEL | __GFP_NOWARN)) in i915_sg_trim()
|
/Linux-v5.4/drivers/gpu/drm/armada/ |
D | armada_gem.c | 393 if (sg_alloc_table(sgt, count, GFP_KERNEL)) in armada_gem_prime_map_dma_buf() 416 if (sg_alloc_table(sgt, 1, GFP_KERNEL)) in armada_gem_prime_map_dma_buf() 425 if (sg_alloc_table(sgt, 1, GFP_KERNEL)) in armada_gem_prime_map_dma_buf()
|
/Linux-v5.4/drivers/hwtracing/intel_th/ |
D | msu-sink.c | 64 ret = sg_alloc_table(*sgt, nents, GFP_KERNEL); in msu_sink_alloc_window()
|
/Linux-v5.4/drivers/staging/android/ion/ |
D | ion_system_heap.c | 131 if (sg_alloc_table(table, i, GFP_KERNEL)) in ion_system_heap_allocate() 311 ret = sg_alloc_table(table, 1, GFP_KERNEL); in ion_system_contig_heap_allocate()
|
D | ion_cma_heap.c | 66 ret = sg_alloc_table(table, 1, GFP_KERNEL); in ion_cma_allocate()
|
/Linux-v5.4/lib/ |
D | scatterlist.c | 355 int sg_alloc_table(struct sg_table *table, unsigned int nents, gfp_t gfp_mask) in sg_alloc_table() function 366 EXPORT_SYMBOL(sg_alloc_table); 413 ret = sg_alloc_table(sgt, chunks, gfp_mask); in __sg_alloc_table_from_pages()
|
/Linux-v5.4/drivers/gpu/drm/i915/gem/selftests/ |
D | huge_gem_object.c | 37 if (sg_alloc_table(pages, npages, GFP)) { in huge_get_pages()
|
D | mock_dmabuf.c | 21 err = sg_alloc_table(st, mock->npages, GFP_KERNEL); in mock_map_dma_buf()
|
D | huge_pages.c | 69 if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) { in get_huge_pages() 193 if (sg_alloc_table(st, obj->base.size >> PAGE_SHIFT, GFP)) { in fake_get_huge_pages() 248 if (sg_alloc_table(st, 1, GFP)) { in fake_get_huge_pages_single()
|
/Linux-v5.4/drivers/gpu/drm/i915/gem/ |
D | i915_gem_internal.c | 72 if (sg_alloc_table(st, npages, GFP_KERNEL)) { in i915_gem_object_get_pages_internal()
|
D | i915_gem_phys.c | 72 if (sg_alloc_table(st, 1, GFP_KERNEL)) { in i915_gem_object_get_pages_phys()
|
D | i915_gem_dmabuf.c | 39 ret = sg_alloc_table(st, obj->mm.pages->nents, GFP_KERNEL); in i915_gem_map_dma_buf()
|
D | i915_gem_shmem.c | 63 if (sg_alloc_table(st, page_count, GFP_KERNEL)) { in shmem_get_pages()
|
/Linux-v5.4/drivers/gpu/drm/omapdrm/ |
D | omap_gem_dmabuf.c | 38 ret = sg_alloc_table(sg, 1, GFP_KERNEL); in omap_gem_map_dma_buf()
|
/Linux-v5.4/drivers/mmc/core/ |
D | sdio_ops.c | 153 if (sg_alloc_table(&sgtable, nents, GFP_KERNEL)) in mmc_io_rw_extended()
|
/Linux-v5.4/drivers/gpu/drm/udl/ |
D | udl_dmabuf.c | 99 ret = sg_alloc_table(sgt, obj->sg->orig_nents, GFP_KERNEL); in udl_map_dma_buf()
|
/Linux-v5.4/drivers/gpu/drm/tegra/ |
D | gem.c | 514 if (sg_alloc_table(sgt, bo->num_pages, GFP_KERNEL)) in tegra_gem_prime_map_dma_buf() 523 if (sg_alloc_table(sgt, 1, GFP_KERNEL)) in tegra_gem_prime_map_dma_buf()
|
/Linux-v5.4/include/linux/ |
D | scatterlist.h | 274 int sg_alloc_table(struct sg_table *, unsigned int, gfp_t);
|
/Linux-v5.4/drivers/infiniband/core/ |
D | umem.c | 264 ret = sg_alloc_table(&umem->sg_head, npages, GFP_KERNEL); in ib_umem_get()
|
/Linux-v5.4/drivers/gpu/drm/i915/selftests/ |
D | scatterlist.c | 223 if (sg_alloc_table(&pt->st, max, in alloc_table()
|
/Linux-v5.4/net/ceph/ |
D | crypto.c | 182 ret = sg_alloc_table(sgt, chunk_cnt, GFP_NOFS); in setup_sgtable()
|
/Linux-v5.4/kernel/dma/ |
D | mapping.c | 133 ret = sg_alloc_table(sgt, 1, GFP_KERNEL); in dma_common_get_sgtable()
|
/Linux-v5.4/drivers/media/common/videobuf2/ |
D | videobuf2-vmalloc.c | 228 ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL); in vb2_vmalloc_dmabuf_ops_attach()
|
/Linux-v5.4/drivers/crypto/ccp/ |
D | ccp-crypto-aes-cmac.c | 110 ret = sg_alloc_table(&rctx->data_sg, sg_count, gfp); in ccp_do_cmac_update()
|
/Linux-v5.4/drivers/crypto/qce/ |
D | ablkcipher.c | 92 ret = sg_alloc_table(&rctx->dst_tbl, rctx->dst_nents, gfp); in qce_ablkcipher_async_req_handle()
|