/Linux-v4.19/drivers/gpu/drm/ttm/ |
D | ttm_tt.c | 70 bo->ttm = NULL; in ttm_tt_create() 75 bo->ttm = bdev->driver->ttm_tt_create(bo, page_flags); in ttm_tt_create() 76 if (unlikely(bo->ttm == NULL)) in ttm_tt_create() 85 static int ttm_tt_alloc_page_directory(struct ttm_tt *ttm) in ttm_tt_alloc_page_directory() argument 87 ttm->pages = kvmalloc_array(ttm->num_pages, sizeof(void*), in ttm_tt_alloc_page_directory() 89 if (!ttm->pages) in ttm_tt_alloc_page_directory() 94 static int ttm_dma_tt_alloc_page_directory(struct ttm_dma_tt *ttm) in ttm_dma_tt_alloc_page_directory() argument 96 ttm->ttm.pages = kvmalloc_array(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory() 97 sizeof(*ttm->ttm.pages) + in ttm_dma_tt_alloc_page_directory() 98 sizeof(*ttm->dma_address), in ttm_dma_tt_alloc_page_directory() [all …]
|
D | ttm_agp_backend.c | 46 struct ttm_tt ttm; member 51 static int ttm_agp_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem) in ttm_agp_bind() argument 53 struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); in ttm_agp_bind() 54 struct page *dummy_read_page = ttm->bdev->glob->dummy_read_page; in ttm_agp_bind() 60 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); in ttm_agp_bind() 65 for (i = 0; i < ttm->num_pages; i++) { in ttm_agp_bind() 66 struct page *page = ttm->pages[i]; in ttm_agp_bind() 85 static int ttm_agp_unbind(struct ttm_tt *ttm) in ttm_agp_unbind() argument 87 struct ttm_agp_backend *agp_be = container_of(ttm, struct ttm_agp_backend, ttm); in ttm_agp_unbind() 98 static void ttm_agp_destroy(struct ttm_tt *ttm) in ttm_agp_destroy() argument [all …]
|
D | ttm_page_alloc.c | 1028 ttm_pool_unpopulate_helper(struct ttm_tt *ttm, unsigned mem_count_update) in ttm_pool_unpopulate_helper() argument 1030 struct ttm_mem_global *mem_glob = ttm->bdev->glob->mem_glob; in ttm_pool_unpopulate_helper() 1037 if (!ttm->pages[i]) in ttm_pool_unpopulate_helper() 1040 ttm_mem_global_free_page(mem_glob, ttm->pages[i], PAGE_SIZE); in ttm_pool_unpopulate_helper() 1044 ttm_put_pages(ttm->pages, ttm->num_pages, ttm->page_flags, in ttm_pool_unpopulate_helper() 1045 ttm->caching_state); in ttm_pool_unpopulate_helper() 1046 ttm->state = tt_unpopulated; in ttm_pool_unpopulate_helper() 1049 int ttm_pool_populate(struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) in ttm_pool_populate() argument 1051 struct ttm_mem_global *mem_glob = ttm->bdev->glob->mem_glob; in ttm_pool_populate() 1055 if (ttm->state != tt_unpopulated) in ttm_pool_populate() [all …]
|
D | ttm_bo_util.c | 57 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_ttm() local 70 ttm_tt_unbind(ttm); in ttm_bo_move_ttm() 77 ret = ttm_tt_set_placement_caching(ttm, new_mem->placement); in ttm_bo_move_ttm() 82 ret = ttm_tt_bind(ttm, new_mem, ctx); in ttm_bo_move_ttm() 312 static int ttm_copy_io_ttm_page(struct ttm_tt *ttm, void *src, in ttm_copy_io_ttm_page() argument 316 struct page *d = ttm->pages[page]; in ttm_copy_io_ttm_page() 334 static int ttm_copy_ttm_io_page(struct ttm_tt *ttm, void *dst, in ttm_copy_ttm_io_page() argument 338 struct page *s = ttm->pages[page]; in ttm_copy_ttm_io_page() 362 struct ttm_tt *ttm = bo->ttm; in ttm_bo_move_memcpy() local 394 (ttm == NULL || (ttm->state == tt_unpopulated && in ttm_bo_move_memcpy() [all …]
|
D | ttm_page_alloc_dma.c | 841 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_pool_get_pages() local 849 ttm->pages[index] = d_page->p; in ttm_dma_pool_get_pages() 861 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_pool_gfp_flags() local 864 if (ttm->page_flags & TTM_PAGE_FLAG_DMA32) in ttm_dma_pool_gfp_flags() 868 if (ttm->page_flags & TTM_PAGE_FLAG_ZERO_ALLOC) in ttm_dma_pool_gfp_flags() 878 if (ttm->page_flags & TTM_PAGE_FLAG_NO_RETRY) in ttm_dma_pool_gfp_flags() 891 struct ttm_tt *ttm = &ttm_dma->ttm; in ttm_dma_populate() local 892 struct ttm_mem_global *mem_glob = ttm->bdev->glob->mem_glob; in ttm_dma_populate() 893 unsigned long num_pages = ttm->num_pages; in ttm_dma_populate() 900 if (ttm->state != tt_unpopulated) in ttm_dma_populate() [all …]
|
D | Makefile | 5 ttm-y := ttm_memory.o ttm_tt.o ttm_bo.o \ 9 ttm-$(CONFIG_AGP) += ttm_agp_backend.o 11 obj-$(CONFIG_DRM_TTM) += ttm.o
|
D | ttm_bo_vm.c | 118 struct ttm_tt *ttm = NULL; in ttm_bo_vm_fault() local 162 if (bo->ttm && (bo->ttm->page_flags & TTM_PAGE_FLAG_SG)) { in ttm_bo_vm_fault() 237 ttm = bo->ttm; in ttm_bo_vm_fault() 242 if (ttm_tt_populate(ttm, &ctx)) { in ttm_bo_vm_fault() 258 page = ttm->pages[page_offset]; in ttm_bo_vm_fault() 371 if (unlikely(bo->ttm->page_flags & TTM_PAGE_FLAG_SWAPPED)) { in ttm_bo_vm_access() 372 ret = ttm_tt_swapin(bo->ttm); in ttm_bo_vm_access()
|
/Linux-v4.19/drivers/gpu/drm/nouveau/ |
D | nouveau_sgdma.c | 13 struct ttm_dma_tt ttm; member 18 nouveau_sgdma_destroy(struct ttm_tt *ttm) in nouveau_sgdma_destroy() argument 20 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; in nouveau_sgdma_destroy() 22 if (ttm) { in nouveau_sgdma_destroy() 23 ttm_dma_tt_fini(&nvbe->ttm); in nouveau_sgdma_destroy() 29 nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *reg) in nv04_sgdma_bind() argument 31 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; in nv04_sgdma_bind() 35 ret = nouveau_mem_host(reg, &nvbe->ttm); in nv04_sgdma_bind() 50 nv04_sgdma_unbind(struct ttm_tt *ttm) in nv04_sgdma_unbind() argument 52 struct nouveau_sgdma_be *nvbe = (struct nouveau_sgdma_be *)ttm; in nv04_sgdma_unbind() [all …]
|
D | nouveau_ttm.c | 174 return ttm_bo_mmap(filp, vma, &drm->ttm.bdev); in nouveau_ttm_mmap() 195 global_ref = &drm->ttm.mem_global_ref; in nouveau_ttm_global_init() 204 drm->ttm.mem_global_ref.release = NULL; in nouveau_ttm_global_init() 208 drm->ttm.bo_global_ref.mem_glob = global_ref->object; in nouveau_ttm_global_init() 209 global_ref = &drm->ttm.bo_global_ref.ref; in nouveau_ttm_global_init() 218 drm_global_item_unref(&drm->ttm.mem_global_ref); in nouveau_ttm_global_init() 219 drm->ttm.mem_global_ref.release = NULL; in nouveau_ttm_global_init() 229 if (drm->ttm.mem_global_ref.release == NULL) in nouveau_ttm_global_release() 232 drm_global_item_unref(&drm->ttm.bo_global_ref.ref); in nouveau_ttm_global_release() 233 drm_global_item_unref(&drm->ttm.mem_global_ref); in nouveau_ttm_global_release() [all …]
|
D | nouveau_bo.c | 216 nvbo->bo.bdev = &drm->ttm.bdev; in nouveau_bo_new() 295 acc_size = ttm_bo_dma_acc_size(&drm->ttm.bdev, size, in nouveau_bo_new() 298 ret = ttm_bo_init(&drm->ttm.bdev, &nvbo->bo, size, in nouveau_bo_new() 512 struct ttm_dma_tt *ttm_dma = (struct ttm_dma_tt *)nvbo->bo.ttm; in nouveau_bo_sync_for_device() 522 for (i = 0; i < ttm_dma->ttm.num_pages; i++) in nouveau_bo_sync_for_device() 532 struct ttm_dma_tt *ttm_dma = (struct ttm_dma_tt *)nvbo->bo.ttm; in nouveau_bo_sync_for_cpu() 542 for (i = 0; i < ttm_dma->ttm.num_pages; i++) in nouveau_bo_sync_for_cpu() 648 const u8 type = mmu->type[drm->ttm.type_vram].type; in nouveau_bo_init_mem_type() 1098 struct nouveau_channel *chan = drm->ttm.chan; in nouveau_bo_move_m2mf() 1116 ret = drm->ttm.move(chan, bo, &bo->mem, new_reg); in nouveau_bo_move_m2mf() [all …]
|
D | nouveau_mem.c | 107 type = drm->ttm.type_ncoh[!!mem->kind]; in nouveau_mem_host() 109 type = drm->ttm.type_host[0]; in nouveau_mem_host() 119 if (tt->ttm.sg) args.sgl = tt->ttm.sg->sgl; in nouveau_mem_host() 148 drm->ttm.type_vram, page, size, in nouveau_mem_vram() 156 drm->ttm.type_vram, page, size, in nouveau_mem_vram()
|
/Linux-v4.19/include/drm/ttm/ |
D | ttm_tt.h | 63 int (*bind) (struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem); 73 int (*unbind) (struct ttm_tt *ttm); 83 void (*destroy) (struct ttm_tt *ttm); 133 struct ttm_tt ttm; member 161 int ttm_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, 175 void ttm_tt_fini(struct ttm_tt *ttm); 186 int ttm_tt_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem, 196 void ttm_tt_destroy(struct ttm_tt *ttm); 205 void ttm_tt_unbind(struct ttm_tt *ttm); 214 int ttm_tt_swapin(struct ttm_tt *ttm); [all …]
|
/Linux-v4.19/drivers/gpu/drm/radeon/ |
D | radeon_ttm.c | 238 if (radeon_ttm_tt_has_userptr(bo->ttm)) in radeon_verify_access() 337 r = ttm_tt_set_placement_caching(bo->ttm, tmp_mem.placement); in radeon_move_vram_ram() 342 r = ttm_tt_bind(bo->ttm, &tmp_mem, &ctx); in radeon_move_vram_ram() 415 if (old_mem->mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { in radeon_bo_move() 532 struct ttm_dma_tt ttm; member 542 static int radeon_ttm_tt_pin_userptr(struct ttm_tt *ttm) in radeon_ttm_tt_pin_userptr() argument 544 struct radeon_device *rdev = radeon_get_rdev(ttm->bdev); in radeon_ttm_tt_pin_userptr() 545 struct radeon_ttm_tt *gtt = (void *)ttm; in radeon_ttm_tt_pin_userptr() 559 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE; in radeon_ttm_tt_pin_userptr() 567 unsigned num_pages = ttm->num_pages - pinned; in radeon_ttm_tt_pin_userptr() [all …]
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_ttm.c | 309 if (amdgpu_ttm_tt_get_usermm(bo->ttm)) in amdgpu_verify_access() 573 r = ttm_tt_set_placement_caching(bo->ttm, tmp_mem.placement); in amdgpu_move_vram_ram() 579 r = ttm_tt_bind(bo->ttm, &tmp_mem, ctx); in amdgpu_move_vram_ram() 667 if (old_mem->mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { in amdgpu_bo_move() 787 struct ttm_dma_tt ttm; member 806 int amdgpu_ttm_tt_get_user_pages(struct ttm_tt *ttm, struct page **pages) in amdgpu_ttm_tt_get_user_pages() argument 808 struct amdgpu_ttm_tt *gtt = (void *)ttm; in amdgpu_ttm_tt_get_user_pages() 827 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE; in amdgpu_ttm_tt_get_user_pages() 839 unsigned num_pages = ttm->num_pages - pinned; in amdgpu_ttm_tt_get_user_pages() 865 } while (pinned < ttm->num_pages); in amdgpu_ttm_tt_get_user_pages() [all …]
|
D | amdgpu_ttm.h | 106 int amdgpu_ttm_tt_get_user_pages(struct ttm_tt *ttm, struct page **pages); 107 void amdgpu_ttm_tt_set_user_pages(struct ttm_tt *ttm, struct page **pages); 108 void amdgpu_ttm_tt_mark_user_pages(struct ttm_tt *ttm); 109 int amdgpu_ttm_tt_set_userptr(struct ttm_tt *ttm, uint64_t addr, 111 bool amdgpu_ttm_tt_has_userptr(struct ttm_tt *ttm); 112 struct mm_struct *amdgpu_ttm_tt_get_usermm(struct ttm_tt *ttm); 113 bool amdgpu_ttm_tt_affect_userptr(struct ttm_tt *ttm, unsigned long start, 115 bool amdgpu_ttm_tt_userptr_invalidated(struct ttm_tt *ttm, 117 bool amdgpu_ttm_tt_userptr_needs_pages(struct ttm_tt *ttm); 118 bool amdgpu_ttm_tt_is_readonly(struct ttm_tt *ttm); [all …]
|
/Linux-v4.19/drivers/gpu/drm/vmwgfx/ |
D | vmwgfx_ttm_buffer.c | 434 vsgt->pages = vmw_tt->dma_ttm.ttm.pages; in vmw_ttm_map_dma() 435 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma() 543 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_map_dma() 560 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_unmap_dma() 580 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_bo_sg_table() 586 static int vmw_ttm_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem) in vmw_ttm_bind() argument 589 container_of(ttm, struct vmw_ttm_tt, dma_ttm.ttm); in vmw_ttm_bind() 602 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind() 606 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind() 612 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind() [all …]
|
D | vmwgfx_blit.c | 466 if (dst->ttm->state == tt_unpopulated) { in vmw_bo_cpu_blit() 467 ret = dst->ttm->bdev->driver->ttm_tt_populate(dst->ttm, &ctx); in vmw_bo_cpu_blit() 472 if (src->ttm->state == tt_unpopulated) { in vmw_bo_cpu_blit() 473 ret = src->ttm->bdev->driver->ttm_tt_populate(src->ttm, &ctx); in vmw_bo_cpu_blit() 482 d.dst_pages = dst->ttm->pages; in vmw_bo_cpu_blit() 483 d.src_pages = src->ttm->pages; in vmw_bo_cpu_blit()
|
/Linux-v4.19/drivers/gpu/drm/mgag200/ |
D | mgag200_ttm.c | 36 return container_of(bd, struct mga_device, ttm.bdev); in mgag200_bdev() 56 global_ref = &ast->ttm.mem_global_ref; in mgag200_ttm_global_init() 68 ast->ttm.bo_global_ref.mem_glob = in mgag200_ttm_global_init() 69 ast->ttm.mem_global_ref.object; in mgag200_ttm_global_init() 70 global_ref = &ast->ttm.bo_global_ref.ref; in mgag200_ttm_global_init() 78 drm_global_item_unref(&ast->ttm.mem_global_ref); in mgag200_ttm_global_init() 87 if (ast->ttm.mem_global_ref.release == NULL) in mgag200_ttm_global_release() 90 drm_global_item_unref(&ast->ttm.bo_global_ref.ref); in mgag200_ttm_global_release() 91 drm_global_item_unref(&ast->ttm.mem_global_ref); in mgag200_ttm_global_release() 92 ast->ttm.mem_global_ref.release = NULL; in mgag200_ttm_global_release() [all …]
|
/Linux-v4.19/drivers/gpu/drm/ast/ |
D | ast_ttm.c | 36 return container_of(bd, struct ast_private, ttm.bdev); in ast_bdev() 56 global_ref = &ast->ttm.mem_global_ref; in ast_ttm_global_init() 68 ast->ttm.bo_global_ref.mem_glob = in ast_ttm_global_init() 69 ast->ttm.mem_global_ref.object; in ast_ttm_global_init() 70 global_ref = &ast->ttm.bo_global_ref.ref; in ast_ttm_global_init() 78 drm_global_item_unref(&ast->ttm.mem_global_ref); in ast_ttm_global_init() 87 if (ast->ttm.mem_global_ref.release == NULL) in ast_ttm_global_release() 90 drm_global_item_unref(&ast->ttm.bo_global_ref.ref); in ast_ttm_global_release() 91 drm_global_item_unref(&ast->ttm.mem_global_ref); in ast_ttm_global_release() 92 ast->ttm.mem_global_ref.release = NULL; in ast_ttm_global_release() [all …]
|
/Linux-v4.19/drivers/gpu/drm/cirrus/ |
D | cirrus_ttm.c | 36 return container_of(bd, struct cirrus_device, ttm.bdev); in cirrus_bdev() 56 global_ref = &cirrus->ttm.mem_global_ref; in cirrus_ttm_global_init() 68 cirrus->ttm.bo_global_ref.mem_glob = in cirrus_ttm_global_init() 69 cirrus->ttm.mem_global_ref.object; in cirrus_ttm_global_init() 70 global_ref = &cirrus->ttm.bo_global_ref.ref; in cirrus_ttm_global_init() 78 drm_global_item_unref(&cirrus->ttm.mem_global_ref); in cirrus_ttm_global_init() 87 if (cirrus->ttm.mem_global_ref.release == NULL) in cirrus_ttm_global_release() 90 drm_global_item_unref(&cirrus->ttm.bo_global_ref.ref); in cirrus_ttm_global_release() 91 drm_global_item_unref(&cirrus->ttm.mem_global_ref); in cirrus_ttm_global_release() 92 cirrus->ttm.mem_global_ref.release = NULL; in cirrus_ttm_global_release() [all …]
|
/Linux-v4.19/drivers/gpu/drm/bochs/ |
D | bochs_mm.c | 16 return container_of(bd, struct bochs_device, ttm.bdev); in bochs_bdev() 34 global_ref = &bochs->ttm.mem_global_ref; in bochs_ttm_global_init() 46 bochs->ttm.bo_global_ref.mem_glob = in bochs_ttm_global_init() 47 bochs->ttm.mem_global_ref.object; in bochs_ttm_global_init() 48 global_ref = &bochs->ttm.bo_global_ref.ref; in bochs_ttm_global_init() 56 drm_global_item_unref(&bochs->ttm.mem_global_ref); in bochs_ttm_global_init() 65 if (bochs->ttm.mem_global_ref.release == NULL) in bochs_ttm_global_release() 68 drm_global_item_unref(&bochs->ttm.bo_global_ref.ref); in bochs_ttm_global_release() 69 drm_global_item_unref(&bochs->ttm.mem_global_ref); in bochs_ttm_global_release() 70 bochs->ttm.mem_global_ref.release = NULL; in bochs_ttm_global_release() [all …]
|
/Linux-v4.19/drivers/staging/vboxvideo/ |
D | vbox_ttm.c | 35 return container_of(bd, struct vbox_private, ttm.bdev); in vbox_bdev() 56 global_ref = &vbox->ttm.mem_global_ref; in vbox_ttm_global_init() 67 vbox->ttm.bo_global_ref.mem_glob = vbox->ttm.mem_global_ref.object; in vbox_ttm_global_init() 68 global_ref = &vbox->ttm.bo_global_ref.ref; in vbox_ttm_global_init() 77 drm_global_item_unref(&vbox->ttm.mem_global_ref); in vbox_ttm_global_init() 89 drm_global_item_unref(&vbox->ttm.bo_global_ref.ref); in vbox_ttm_global_release() 90 drm_global_item_unref(&vbox->ttm.mem_global_ref); in vbox_ttm_global_release() 228 struct ttm_bo_device *bdev = &vbox->ttm.bdev; in vbox_mm_init() 234 ret = ttm_bo_device_init(&vbox->ttm.bdev, in vbox_mm_init() 235 vbox->ttm.bo_global_ref.ref.object, in vbox_mm_init() [all …]
|
/Linux-v4.19/drivers/gpu/drm/virtio/ |
D | virtgpu_ttm.c | 286 struct ttm_dma_tt ttm; member 291 static int virtio_gpu_ttm_backend_bind(struct ttm_tt *ttm, in virtio_gpu_ttm_backend_bind() argument 294 struct virtio_gpu_ttm_tt *gtt = (void *)ttm; in virtio_gpu_ttm_backend_bind() 297 if (!ttm->num_pages) in virtio_gpu_ttm_backend_bind() 299 ttm->num_pages, bo_mem, ttm); in virtio_gpu_ttm_backend_bind() 305 static int virtio_gpu_ttm_backend_unbind(struct ttm_tt *ttm) in virtio_gpu_ttm_backend_unbind() argument 311 static void virtio_gpu_ttm_backend_destroy(struct ttm_tt *ttm) in virtio_gpu_ttm_backend_destroy() argument 313 struct virtio_gpu_ttm_tt *gtt = (void *)ttm; in virtio_gpu_ttm_backend_destroy() 315 ttm_dma_tt_fini(>t->ttm); in virtio_gpu_ttm_backend_destroy() 335 gtt->ttm.ttm.func = &virtio_gpu_backend_func; in virtio_gpu_ttm_tt_create() [all …]
|
/Linux-v4.19/drivers/gpu/drm/qxl/ |
D | qxl_ttm.c | 255 struct ttm_dma_tt ttm; member 260 static int qxl_ttm_backend_bind(struct ttm_tt *ttm, in qxl_ttm_backend_bind() argument 263 struct qxl_ttm_tt *gtt = (void *)ttm; in qxl_ttm_backend_bind() 266 if (!ttm->num_pages) { in qxl_ttm_backend_bind() 268 ttm->num_pages, bo_mem, ttm); in qxl_ttm_backend_bind() 274 static int qxl_ttm_backend_unbind(struct ttm_tt *ttm) in qxl_ttm_backend_unbind() argument 280 static void qxl_ttm_backend_destroy(struct ttm_tt *ttm) in qxl_ttm_backend_destroy() argument 282 struct qxl_ttm_tt *gtt = (void *)ttm; in qxl_ttm_backend_destroy() 284 ttm_dma_tt_fini(>t->ttm); in qxl_ttm_backend_destroy() 304 gtt->ttm.ttm.func = &qxl_backend_func; in qxl_ttm_tt_create() [all …]
|
/Linux-v4.19/arch/powerpc/perf/ |
D | ppc970-pmu.c | 264 unsigned int ttm, grp; in p970_compute_mmcr() local 321 ttm = unitmap[i]; in p970_compute_mmcr() 322 ++ttmuse[(ttm >> 2) & 1]; in p970_compute_mmcr() 323 mmcr1 |= (unsigned long)(ttm & ~4) << MMCR1_TTM1SEL_SH; in p970_compute_mmcr() 335 ttm = (unitmap[unit] >> 2) & 1; in p970_compute_mmcr() 337 ttm = 2; in p970_compute_mmcr() 339 ttm = 3; in p970_compute_mmcr() 343 mmcr1 |= (unsigned long)ttm in p970_compute_mmcr()
|