Lines Matching refs:vsgt
267 struct vmw_sg_table vsgt; member
346 void vmw_piter_start(struct vmw_piter *viter, const struct vmw_sg_table *vsgt, in vmw_piter_start() argument
350 viter->num_pages = vsgt->num_pages; in vmw_piter_start()
352 viter->pages = vsgt->pages; in vmw_piter_start()
353 switch (vsgt->mode) { in vmw_piter_start()
361 viter->addrs = vsgt->addrs; in vmw_piter_start()
367 __sg_page_iter_start(&viter->iter.base, vsgt->sgt->sgl, in vmw_piter_start()
368 vsgt->sgt->orig_nents, p_offset); in vmw_piter_start()
425 struct vmw_sg_table *vsgt = &vmw_tt->vsgt; in vmw_ttm_map_dma() local
440 vsgt->mode = dev_priv->map_mode; in vmw_ttm_map_dma()
441 vsgt->pages = vmw_tt->dma_ttm.ttm.pages; in vmw_ttm_map_dma()
442 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma()
443 vsgt->addrs = vmw_tt->dma_ttm.dma_address; in vmw_ttm_map_dma()
444 vsgt->sgt = &vmw_tt->sgt; in vmw_ttm_map_dma()
453 vmw_tt->sg_alloc_size = sgt_size + sgl_size * vsgt->num_pages; in vmw_ttm_map_dma()
458 sg = __sg_alloc_table_from_pages(&vmw_tt->sgt, vsgt->pages, in vmw_ttm_map_dma()
459 vsgt->num_pages, 0, in vmw_ttm_map_dma()
460 (unsigned long) vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma()
468 if (vsgt->num_pages > vmw_tt->sgt.orig_nents) { in vmw_ttm_map_dma()
470 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma()
487 vmw_tt->vsgt.num_regions = 0; in vmw_ttm_map_dma()
488 for (vmw_piter_start(&iter, vsgt, 0); vmw_piter_next(&iter);) { in vmw_ttm_map_dma()
492 vmw_tt->vsgt.num_regions++; in vmw_ttm_map_dma()
500 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_map_dma()
501 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_map_dma()
520 if (!vmw_tt->vsgt.sgt) in vmw_ttm_unmap_dma()
527 sg_free_table(vmw_tt->vsgt.sgt); in vmw_ttm_unmap_dma()
528 vmw_tt->vsgt.sgt = NULL; in vmw_ttm_unmap_dma()
554 return &vmw_tt->vsgt; in vmw_bo_sg_table()
580 ret = vmw_gmr_bind(vmw_be->dev_priv, &vmw_be->vsgt, in vmw_ttm_bind()
592 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()