/Linux-v4.19/drivers/gpu/drm/msm/ |
D | msm_fence.c | 27 struct msm_fence_context *fctx; in msm_fence_context_alloc() local 29 fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in msm_fence_context_alloc() 30 if (!fctx) in msm_fence_context_alloc() 33 fctx->dev = dev; in msm_fence_context_alloc() 34 strncpy(fctx->name, name, sizeof(fctx->name)); in msm_fence_context_alloc() 35 fctx->context = dma_fence_context_alloc(1); in msm_fence_context_alloc() 36 init_waitqueue_head(&fctx->event); in msm_fence_context_alloc() 37 spin_lock_init(&fctx->spinlock); in msm_fence_context_alloc() 39 return fctx; in msm_fence_context_alloc() 42 void msm_fence_context_free(struct msm_fence_context *fctx) in msm_fence_context_free() argument [all …]
|
D | msm_fence.h | 36 void msm_fence_context_free(struct msm_fence_context *fctx); 38 int msm_wait_fence(struct msm_fence_context *fctx, uint32_t fence, 40 void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence); 42 struct dma_fence * msm_fence_alloc(struct msm_fence_context *fctx);
|
D | msm_ringbuffer.c | 60 ring->fctx = msm_fence_context_alloc(gpu->dev, name); in msm_ringbuffer_new() 74 msm_fence_context_free(ring->fctx); in msm_ringbuffer_destroy()
|
D | msm_ringbuffer.h | 42 struct msm_fence_context *fctx; member
|
D | msm_gem_submit.c | 252 ret = msm_gem_sync_object(&msm_obj->base, submit->ring->fctx, in submit_fence_sync() 455 if (!dma_fence_match_context(in_fence, ring->fctx->context)) { in msm_ioctl_gem_submit() 560 submit->fence = msm_fence_alloc(ring->fctx); in msm_ioctl_gem_submit()
|
D | msm_gem.c | 619 struct msm_fence_context *fctx, bool exclusive) in msm_gem_sync_object() argument 630 if (fence && (fence->context != fctx->context)) { in msm_gem_sync_object() 643 if (fence->context != fctx->context) { in msm_gem_sync_object()
|
D | msm_drv.h | 294 struct msm_fence_context *fctx, bool exclusive);
|
D | msm_gpu.c | 389 msm_update_fence(submit->ring->fctx, in update_fences()
|
/Linux-v4.19/drivers/gpu/drm/nouveau/ |
D | nouveau_fence.c | 66 struct nouveau_fence_chan *fctx = nouveau_fctx(fence); in nouveau_fence_signal() local 68 if (!--fctx->notify_ref) in nouveau_fence_signal() 91 nouveau_fence_context_del(struct nouveau_fence_chan *fctx) in nouveau_fence_context_del() argument 95 spin_lock_irq(&fctx->lock); in nouveau_fence_context_del() 96 while (!list_empty(&fctx->pending)) { in nouveau_fence_context_del() 97 fence = list_entry(fctx->pending.next, typeof(*fence), head); in nouveau_fence_context_del() 100 nvif_notify_put(&fctx->notify); in nouveau_fence_context_del() 102 spin_unlock_irq(&fctx->lock); in nouveau_fence_context_del() 104 nvif_notify_fini(&fctx->notify); in nouveau_fence_context_del() 105 fctx->dead = 1; in nouveau_fence_context_del() [all …]
|
D | nv84_fence.c | 71 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_emit() local 72 u64 addr = fctx->vma->addr + chan->chid * 16; in nv84_fence_emit() 74 return fctx->base.emit32(chan, addr, fence->base.seqno); in nv84_fence_emit() 81 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_sync() local 82 u64 addr = fctx->vma->addr + prev->chid * 16; in nv84_fence_sync() 84 return fctx->base.sync32(chan, addr, fence->base.seqno); in nv84_fence_sync() 98 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_context_del() local 100 nouveau_bo_wr32(priv->bo, chan->chid * 16 / 4, fctx->base.sequence); in nv84_fence_context_del() 102 nouveau_vma_del(&fctx->vma); in nv84_fence_context_del() 104 nouveau_fence_context_del(&fctx->base); in nv84_fence_context_del() [all …]
|
D | nv10_fence.c | 59 struct nv10_fence_chan *fctx = chan->fence; in nv10_fence_context_del() local 60 nouveau_fence_context_del(&fctx->base); in nv10_fence_context_del() 61 nvif_object_fini(&fctx->sema); in nv10_fence_context_del() 63 nouveau_fence_context_free(&fctx->base); in nv10_fence_context_del() 69 struct nv10_fence_chan *fctx; in nv10_fence_context_new() local 71 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv10_fence_context_new() 72 if (!fctx) in nv10_fence_context_new() 75 nouveau_fence_context_new(chan, &fctx->base); in nv10_fence_context_new() 76 fctx->base.emit = nv10_fence_emit; in nv10_fence_context_new() 77 fctx->base.read = nv10_fence_read; in nv10_fence_context_new() [all …]
|
D | nv04_fence.c | 71 struct nv04_fence_chan *fctx = chan->fence; in nv04_fence_context_del() local 72 nouveau_fence_context_del(&fctx->base); in nv04_fence_context_del() 74 nouveau_fence_context_free(&fctx->base); in nv04_fence_context_del() 80 struct nv04_fence_chan *fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv04_fence_context_new() local 81 if (fctx) { in nv04_fence_context_new() 82 nouveau_fence_context_new(chan, &fctx->base); in nv04_fence_context_new() 83 fctx->base.emit = nv04_fence_emit; in nv04_fence_context_new() 84 fctx->base.sync = nv04_fence_sync; in nv04_fence_context_new() 85 fctx->base.read = nv04_fence_read; in nv04_fence_context_new() 86 chan->fence = fctx; in nv04_fence_context_new()
|
D | nv17_fence.c | 39 struct nv10_fence_chan *fctx = chan->fence; in nv17_fence_sync() local 54 OUT_RING (prev, fctx->sema.handle); in nv17_fence_sync() 63 OUT_RING (chan, fctx->sema.handle); in nv17_fence_sync() 78 struct nv10_fence_chan *fctx; in nv17_fence_context_new() local 84 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv17_fence_context_new() 85 if (!fctx) in nv17_fence_context_new() 88 nouveau_fence_context_new(chan, &fctx->base); in nv17_fence_context_new() 89 fctx->base.emit = nv10_fence_emit; in nv17_fence_context_new() 90 fctx->base.read = nv10_fence_read; in nv17_fence_context_new() 91 fctx->base.sync = nv17_fence_sync; in nv17_fence_context_new() [all …]
|
D | nv50_fence.c | 39 struct nv10_fence_chan *fctx; in nv50_fence_context_new() local 45 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv50_fence_context_new() 46 if (!fctx) in nv50_fence_context_new() 49 nouveau_fence_context_new(chan, &fctx->base); in nv50_fence_context_new() 50 fctx->base.emit = nv10_fence_emit; in nv50_fence_context_new() 51 fctx->base.read = nv10_fence_read; in nv50_fence_context_new() 52 fctx->base.sync = nv17_fence_sync; in nv50_fence_context_new() 61 &fctx->sema); in nv50_fence_context_new()
|
D | nvc0_fence.c | 68 struct nv84_fence_chan *fctx = chan->fence; in nvc0_fence_context_new() local 69 fctx->base.emit32 = nvc0_fence_emit32; in nvc0_fence_context_new() 70 fctx->base.sync32 = nvc0_fence_sync32; in nvc0_fence_context_new()
|
D | nouveau_display.c | 749 struct nouveau_fence_chan *fctx = chan->fence; in nouveau_page_flip_emit() local 757 list_add_tail(&s->head, &fctx->flip); in nouveau_page_flip_emit() 896 struct nouveau_fence_chan *fctx = chan->fence; in nouveau_finish_page_flip() local 904 if (list_empty(&fctx->flip)) { in nouveau_finish_page_flip() 910 s = list_first_entry(&fctx->flip, struct nouveau_page_flip_state, head); in nouveau_finish_page_flip()
|
/Linux-v4.19/drivers/crypto/cavium/nitrox/ |
D | nitrox_algs.c | 76 void *fctx; in nitrox_skcipher_init() local 84 fctx = crypto_alloc_context(nctx->ndev); in nitrox_skcipher_init() 85 if (!fctx) { in nitrox_skcipher_init() 89 nctx->u.ctx_handle = (uintptr_t)fctx; in nitrox_skcipher_init() 101 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_skcipher_exit() local 103 memset(&fctx->crypto, 0, sizeof(struct crypto_keys)); in nitrox_skcipher_exit() 104 memset(&fctx->auth, 0, sizeof(struct auth_keys)); in nitrox_skcipher_exit() 105 crypto_free_context((void *)fctx); in nitrox_skcipher_exit() 119 struct flexi_crypto_context *fctx; in nitrox_skcipher_setkey() local 131 fctx = nctx->u.fctx; in nitrox_skcipher_setkey() [all …]
|
D | nitrox_req.h | 188 struct flexi_crypto_context *fctx; member
|
/Linux-v4.19/drivers/crypto/cavium/cpt/ |
D | cptvf_algs.c | 106 struct fc_context *fctx = &rctx->fctx; in create_ctx_hdr() local 126 fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type; in create_ctx_hdr() 127 fctx->enc.enc_ctrl.e.aes_key = ctx->key_type; in create_ctx_hdr() 128 fctx->enc.enc_ctrl.e.iv_source = FROM_DPTR; in create_ctx_hdr() 131 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len * 2); in create_ctx_hdr() 133 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len); in create_ctx_hdr() 134 ctrl_flags = (u64 *)&fctx->enc.enc_ctrl.flags; in create_ctx_hdr() 146 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr() 202 struct fc_context *fctx = &rctx->fctx; in cvm_enc_dec() local 208 memset(fctx, 0, sizeof(struct fc_context)); in cvm_enc_dec()
|
D | cptvf_algs.h | 116 struct fc_context fctx; member
|
/Linux-v4.19/security/apparmor/ |
D | file.c | 491 static void update_file_ctx(struct aa_file_ctx *fctx, struct aa_label *label, in update_file_ctx() argument 497 spin_lock(&fctx->lock); in update_file_ctx() 498 old = rcu_dereference_protected(fctx->label, in update_file_ctx() 499 spin_is_locked(&fctx->lock)); in update_file_ctx() 503 rcu_assign_pointer(fctx->label, l); in update_file_ctx() 507 fctx->allow |= request; in update_file_ctx() 509 spin_unlock(&fctx->lock); in update_file_ctx() 603 struct aa_file_ctx *fctx; in aa_file_perm() local 611 fctx = file_ctx(file); in aa_file_perm() 614 flabel = rcu_dereference(fctx->label); in aa_file_perm() [all …]
|
D | lsm.c | 400 struct aa_file_ctx *fctx = file_ctx(file); in apparmor_file_open() local 413 fctx->allow = MAY_EXEC | MAY_READ | AA_EXEC_MMAP; in apparmor_file_open() 425 fctx->allow = aa_map_file_to_perms(file); in apparmor_file_open()
|
/Linux-v4.19/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | dmanv04.c | 79 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini() local 98 nvkm_kmap(fctx); in nv04_fifo_dma_fini() 103 u32 cv = (nvkm_ro32(fctx, c->ctxp + data) & ~cm); in nv04_fifo_dma_fini() 104 nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs)); in nv04_fifo_dma_fini() 106 nvkm_done(fctx); in nv04_fifo_dma_fini()
|
/Linux-v4.19/fs/ |
D | userfaultfd.c | 663 struct userfaultfd_fork_ctx *fctx; in dup_userfaultfd() local 672 list_for_each_entry(fctx, fcs, list) in dup_userfaultfd() 673 if (fctx->orig == octx) { in dup_userfaultfd() 674 ctx = fctx->new; in dup_userfaultfd() 679 fctx = kmalloc(sizeof(*fctx), GFP_KERNEL); in dup_userfaultfd() 680 if (!fctx) in dup_userfaultfd() 685 kfree(fctx); in dup_userfaultfd() 700 fctx->orig = octx; in dup_userfaultfd() 701 fctx->new = ctx; in dup_userfaultfd() 702 list_add_tail(&fctx->list, fcs); in dup_userfaultfd() [all …]
|
/Linux-v4.19/drivers/block/ |
D | rbd.c | 1995 struct rbd_img_fill_ctx *fctx) in rbd_img_fill_request_nocopy() argument 2000 img_req->data_type = fctx->pos_type; in rbd_img_fill_request_nocopy() 2006 fctx->iter = *fctx->pos; in rbd_img_fill_request_nocopy() 2013 fctx->set_pos_fn, &fctx->iter); in rbd_img_fill_request_nocopy() 2037 struct rbd_img_fill_ctx *fctx) in rbd_img_fill_request() argument 2044 if (fctx->pos_type == OBJ_REQUEST_NODATA || in rbd_img_fill_request() 2047 num_img_extents, fctx); in rbd_img_fill_request() 2058 fctx->iter = *fctx->pos; in rbd_img_fill_request() 2065 fctx->count_fn, &fctx->iter); in rbd_img_fill_request() 2082 fctx->iter = *fctx->pos; in rbd_img_fill_request() [all …]
|