Lines Matching refs:ttm
441 vsgt->pages = vmw_tt->dma_ttm.ttm.pages; in vmw_ttm_map_dma()
442 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma()
552 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_sg_table()
559 struct ttm_tt *ttm, struct ttm_resource *bo_mem) in vmw_ttm_bind() argument
562 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_bind()
581 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind()
586 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind()
592 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()
603 struct ttm_tt *ttm) in vmw_ttm_unbind() argument
606 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_unbind()
628 static void vmw_ttm_destroy(struct ttm_bo_device *bdev, struct ttm_tt *ttm) in vmw_ttm_destroy() argument
631 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_destroy()
633 vmw_ttm_unbind(bdev, ttm); in vmw_ttm_destroy()
634 ttm_tt_destroy_common(bdev, ttm); in vmw_ttm_destroy()
639 ttm_tt_fini(ttm); in vmw_ttm_destroy()
649 struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) in vmw_ttm_populate() argument
652 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_populate()
657 if (ttm_tt_is_populated(ttm)) in vmw_ttm_populate()
662 ttm_round_pot(ttm->num_pages * sizeof(dma_addr_t)); in vmw_ttm_populate()
672 ret = ttm_pool_populate(ttm, ctx); in vmw_ttm_populate()
678 struct ttm_tt *ttm) in vmw_ttm_unpopulate() argument
680 struct vmw_ttm_tt *vmw_tt = container_of(ttm, struct vmw_ttm_tt, in vmw_ttm_unpopulate()
681 dma_ttm.ttm); in vmw_ttm_unpopulate()
694 ttm_round_pot(ttm->num_pages * sizeof(dma_addr_t)); in vmw_ttm_unpopulate()
699 ttm_pool_unpopulate(ttm); in vmw_ttm_unpopulate()
718 ret = ttm_tt_init(&vmw_be->dma_ttm.ttm, bo, page_flags); in vmw_ttm_tt_create()
722 return &vmw_be->dma_ttm.ttm; in vmw_ttm_tt_create()
830 ret = vmw_ttm_populate(bo->bdev, bo->ttm, &ctx); in vmw_bo_create_and_populate()
833 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_create_and_populate()