Lines Matching refs:qdev
60 struct qxl_device *qdev; in qxl_fence_wait() local
63 qdev = container_of(fence->lock, struct qxl_device, release_lock); in qxl_fence_wait()
65 if (!wait_event_timeout(qdev->release_event, in qxl_fence_wait()
67 (qxl_io_notify_oom(qdev), 0)), in qxl_fence_wait()
84 qxl_release_alloc(struct qxl_device *qdev, int type, in qxl_release_alloc() argument
103 spin_lock(&qdev->release_idr_lock); in qxl_release_alloc()
104 handle = idr_alloc(&qdev->release_idr, release, 1, 0, GFP_NOWAIT); in qxl_release_alloc()
105 release->base.seqno = ++qdev->release_seqno; in qxl_release_alloc()
106 spin_unlock(&qdev->release_idr_lock); in qxl_release_alloc()
137 qxl_release_free(struct qxl_device *qdev, in qxl_release_free() argument
143 qxl_surface_id_dealloc(qdev, release->surface_release_id); in qxl_release_free()
145 spin_lock(&qdev->release_idr_lock); in qxl_release_free()
146 idr_remove(&qdev->release_idr, release->id); in qxl_release_free()
147 spin_unlock(&qdev->release_idr_lock); in qxl_release_free()
159 atomic_dec(&qdev->release_count); in qxl_release_free()
162 static int qxl_release_bo_alloc(struct qxl_device *qdev, in qxl_release_bo_alloc() argument
167 return qxl_bo_create(qdev, PAGE_SIZE, false, true, in qxl_release_bo_alloc()
251 int qxl_alloc_surface_release_reserved(struct qxl_device *qdev, in qxl_alloc_surface_release_reserved() argument
262 idr_ret = qxl_release_alloc(qdev, QXL_RELEASE_SURFACE_CMD, release); in qxl_alloc_surface_release_reserved()
272 info = qxl_release_map(qdev, *release); in qxl_alloc_surface_release_reserved()
274 qxl_release_unmap(qdev, *release, info); in qxl_alloc_surface_release_reserved()
278 return qxl_alloc_release_reserved(qdev, sizeof(struct qxl_surface_cmd), in qxl_alloc_surface_release_reserved()
282 int qxl_alloc_release_reserved(struct qxl_device *qdev, unsigned long size, in qxl_alloc_release_reserved() argument
308 idr_ret = qxl_release_alloc(qdev, type, release); in qxl_alloc_release_reserved()
314 atomic_inc(&qdev->release_count); in qxl_alloc_release_reserved()
316 mutex_lock(&qdev->release_mutex); in qxl_alloc_release_reserved()
317 if (qdev->current_release_bo_offset[cur_idx] + 1 >= releases_per_bo[cur_idx]) { in qxl_alloc_release_reserved()
318 free_bo = qdev->current_release_bo[cur_idx]; in qxl_alloc_release_reserved()
319 qdev->current_release_bo_offset[cur_idx] = 0; in qxl_alloc_release_reserved()
320 qdev->current_release_bo[cur_idx] = NULL; in qxl_alloc_release_reserved()
322 if (!qdev->current_release_bo[cur_idx]) { in qxl_alloc_release_reserved()
323 ret = qxl_release_bo_alloc(qdev, &qdev->current_release_bo[cur_idx], priority); in qxl_alloc_release_reserved()
325 mutex_unlock(&qdev->release_mutex); in qxl_alloc_release_reserved()
330 qxl_release_free(qdev, *release); in qxl_alloc_release_reserved()
335 bo = qxl_bo_ref(qdev->current_release_bo[cur_idx]); in qxl_alloc_release_reserved()
338 …(*release)->release_offset = qdev->current_release_bo_offset[cur_idx] * release_size_per_bo[cur_id… in qxl_alloc_release_reserved()
339 qdev->current_release_bo_offset[cur_idx]++; in qxl_alloc_release_reserved()
344 mutex_unlock(&qdev->release_mutex); in qxl_alloc_release_reserved()
353 qxl_release_free(qdev, *release); in qxl_alloc_release_reserved()
357 info = qxl_release_map(qdev, *release); in qxl_alloc_release_reserved()
359 qxl_release_unmap(qdev, *release, info); in qxl_alloc_release_reserved()
364 struct qxl_release *qxl_release_from_id_locked(struct qxl_device *qdev, in qxl_release_from_id_locked() argument
369 spin_lock(&qdev->release_idr_lock); in qxl_release_from_id_locked()
370 release = idr_find(&qdev->release_idr, id); in qxl_release_from_id_locked()
371 spin_unlock(&qdev->release_idr_lock); in qxl_release_from_id_locked()
380 union qxl_release_info *qxl_release_map(struct qxl_device *qdev, in qxl_release_map() argument
387 ptr = qxl_bo_kmap_atomic_page(qdev, bo, release->release_offset & PAGE_MASK); in qxl_release_map()
394 void qxl_release_unmap(struct qxl_device *qdev, in qxl_release_unmap() argument
402 qxl_bo_kunmap_atomic_page(qdev, bo, ptr); in qxl_release_unmap()
410 struct qxl_device *qdev; in qxl_release_fence_buffer_objects() local
419 qdev = container_of(bdev, struct qxl_device, mman.bdev); in qxl_release_fence_buffer_objects()
425 dma_fence_init(&release->base, &qxl_fence_ops, &qdev->release_lock, in qxl_release_fence_buffer_objects()