Home
last modified time | relevance | path

Searched refs:fctx (Results 1 – 25 of 29) sorted by relevance

12

/Linux-v5.4/drivers/gpu/drm/msm/
Dmsm_fence.c16 struct msm_fence_context *fctx; in msm_fence_context_alloc() local
18 fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in msm_fence_context_alloc()
19 if (!fctx) in msm_fence_context_alloc()
22 fctx->dev = dev; in msm_fence_context_alloc()
23 strncpy(fctx->name, name, sizeof(fctx->name)); in msm_fence_context_alloc()
24 fctx->context = dma_fence_context_alloc(1); in msm_fence_context_alloc()
25 init_waitqueue_head(&fctx->event); in msm_fence_context_alloc()
26 spin_lock_init(&fctx->spinlock); in msm_fence_context_alloc()
28 return fctx; in msm_fence_context_alloc()
31 void msm_fence_context_free(struct msm_fence_context *fctx) in msm_fence_context_free() argument
[all …]
Dmsm_fence.h25 void msm_fence_context_free(struct msm_fence_context *fctx);
27 int msm_wait_fence(struct msm_fence_context *fctx, uint32_t fence,
29 void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence);
31 struct dma_fence * msm_fence_alloc(struct msm_fence_context *fctx);
Dmsm_ringbuffer.c52 ring->fctx = msm_fence_context_alloc(gpu->dev, name); in msm_ringbuffer_new()
66 msm_fence_context_free(ring->fctx); in msm_ringbuffer_destroy()
Dmsm_ringbuffer.h47 struct msm_fence_context *fctx; member
Dmsm_gem_submit.c241 ret = msm_gem_sync_object(&msm_obj->base, submit->ring->fctx, in submit_fence_sync()
455 if (!dma_fence_match_context(in_fence, ring->fctx->context)) in msm_ioctl_gem_submit()
565 submit->fence = msm_fence_alloc(ring->fctx); in msm_ioctl_gem_submit()
Dmsm_gem.c703 struct msm_fence_context *fctx, bool exclusive) in msm_gem_sync_object() argument
713 if (fence && (fence->context != fctx->context)) { in msm_gem_sync_object()
726 if (fence->context != fctx->context) { in msm_gem_sync_object()
/Linux-v5.4/drivers/gpu/drm/nouveau/
Dnouveau_fence.c65 struct nouveau_fence_chan *fctx = nouveau_fctx(fence); in nouveau_fence_signal() local
67 if (!--fctx->notify_ref) in nouveau_fence_signal()
90 nouveau_fence_context_del(struct nouveau_fence_chan *fctx) in nouveau_fence_context_del() argument
94 spin_lock_irq(&fctx->lock); in nouveau_fence_context_del()
95 while (!list_empty(&fctx->pending)) { in nouveau_fence_context_del()
96 fence = list_entry(fctx->pending.next, typeof(*fence), head); in nouveau_fence_context_del()
99 nvif_notify_put(&fctx->notify); in nouveau_fence_context_del()
101 spin_unlock_irq(&fctx->lock); in nouveau_fence_context_del()
103 nvif_notify_fini(&fctx->notify); in nouveau_fence_context_del()
104 fctx->dead = 1; in nouveau_fence_context_del()
[all …]
Dnv84_fence.c71 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_emit() local
72 u64 addr = fctx->vma->addr + chan->chid * 16; in nv84_fence_emit()
74 return fctx->base.emit32(chan, addr, fence->base.seqno); in nv84_fence_emit()
81 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_sync() local
82 u64 addr = fctx->vma->addr + prev->chid * 16; in nv84_fence_sync()
84 return fctx->base.sync32(chan, addr, fence->base.seqno); in nv84_fence_sync()
98 struct nv84_fence_chan *fctx = chan->fence; in nv84_fence_context_del() local
100 nouveau_bo_wr32(priv->bo, chan->chid * 16 / 4, fctx->base.sequence); in nv84_fence_context_del()
102 nouveau_vma_del(&fctx->vma); in nv84_fence_context_del()
104 nouveau_fence_context_del(&fctx->base); in nv84_fence_context_del()
[all …]
Dnv10_fence.c59 struct nv10_fence_chan *fctx = chan->fence; in nv10_fence_context_del() local
60 nouveau_fence_context_del(&fctx->base); in nv10_fence_context_del()
61 nvif_object_fini(&fctx->sema); in nv10_fence_context_del()
63 nouveau_fence_context_free(&fctx->base); in nv10_fence_context_del()
69 struct nv10_fence_chan *fctx; in nv10_fence_context_new() local
71 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv10_fence_context_new()
72 if (!fctx) in nv10_fence_context_new()
75 nouveau_fence_context_new(chan, &fctx->base); in nv10_fence_context_new()
76 fctx->base.emit = nv10_fence_emit; in nv10_fence_context_new()
77 fctx->base.read = nv10_fence_read; in nv10_fence_context_new()
[all …]
Dnv04_fence.c71 struct nv04_fence_chan *fctx = chan->fence; in nv04_fence_context_del() local
72 nouveau_fence_context_del(&fctx->base); in nv04_fence_context_del()
74 nouveau_fence_context_free(&fctx->base); in nv04_fence_context_del()
80 struct nv04_fence_chan *fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv04_fence_context_new() local
81 if (fctx) { in nv04_fence_context_new()
82 nouveau_fence_context_new(chan, &fctx->base); in nv04_fence_context_new()
83 fctx->base.emit = nv04_fence_emit; in nv04_fence_context_new()
84 fctx->base.sync = nv04_fence_sync; in nv04_fence_context_new()
85 fctx->base.read = nv04_fence_read; in nv04_fence_context_new()
86 chan->fence = fctx; in nv04_fence_context_new()
Dnv17_fence.c39 struct nv10_fence_chan *fctx = chan->fence; in nv17_fence_sync() local
54 OUT_RING (prev, fctx->sema.handle); in nv17_fence_sync()
63 OUT_RING (chan, fctx->sema.handle); in nv17_fence_sync()
78 struct nv10_fence_chan *fctx; in nv17_fence_context_new() local
84 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv17_fence_context_new()
85 if (!fctx) in nv17_fence_context_new()
88 nouveau_fence_context_new(chan, &fctx->base); in nv17_fence_context_new()
89 fctx->base.emit = nv10_fence_emit; in nv17_fence_context_new()
90 fctx->base.read = nv10_fence_read; in nv17_fence_context_new()
91 fctx->base.sync = nv17_fence_sync; in nv17_fence_context_new()
[all …]
Dnv50_fence.c39 struct nv10_fence_chan *fctx; in nv50_fence_context_new() local
45 fctx = chan->fence = kzalloc(sizeof(*fctx), GFP_KERNEL); in nv50_fence_context_new()
46 if (!fctx) in nv50_fence_context_new()
49 nouveau_fence_context_new(chan, &fctx->base); in nv50_fence_context_new()
50 fctx->base.emit = nv10_fence_emit; in nv50_fence_context_new()
51 fctx->base.read = nv10_fence_read; in nv50_fence_context_new()
52 fctx->base.sync = nv17_fence_sync; in nv50_fence_context_new()
61 &fctx->sema); in nv50_fence_context_new()
Dnvc0_fence.c68 struct nv84_fence_chan *fctx = chan->fence; in nvc0_fence_context_new() local
69 fctx->base.emit32 = nvc0_fence_emit32; in nvc0_fence_context_new()
70 fctx->base.sync32 = nvc0_fence_sync32; in nvc0_fence_context_new()
/Linux-v5.4/drivers/crypto/cavium/nitrox/
Dnitrox_aead.c39 struct flexi_crypto_context *fctx; in nitrox_aes_gcm_setkey() local
49 fctx = nctx->u.fctx; in nitrox_aes_gcm_setkey()
50 flags.f = be64_to_cpu(fctx->flags.f); in nitrox_aes_gcm_setkey()
52 fctx->flags.f = cpu_to_be64(flags.f); in nitrox_aes_gcm_setkey()
55 memset(&fctx->crypto, 0, sizeof(fctx->crypto)); in nitrox_aes_gcm_setkey()
56 memcpy(fctx->crypto.u.key, key, keylen); in nitrox_aes_gcm_setkey()
65 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aead_setauthsize() local
68 flags.f = be64_to_cpu(fctx->flags.f); in nitrox_aead_setauthsize()
70 fctx->flags.f = cpu_to_be64(flags.f); in nitrox_aead_setauthsize()
195 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_aes_gcm_enc() local
[all …]
Dnitrox_skcipher.c80 struct flexi_crypto_context *fctx = nctx->u.fctx; in nitrox_skcipher_exit() local
82 memzero_explicit(&fctx->crypto, sizeof(struct crypto_keys)); in nitrox_skcipher_exit()
83 memzero_explicit(&fctx->auth, sizeof(struct auth_keys)); in nitrox_skcipher_exit()
98 struct flexi_crypto_context *fctx; in nitrox_skcipher_setkey() local
111 fctx = nctx->u.fctx; in nitrox_skcipher_setkey()
112 flags = &fctx->flags; in nitrox_skcipher_setkey()
119 memcpy(fctx->crypto.u.key, key, keylen); in nitrox_skcipher_setkey()
279 struct flexi_crypto_context *fctx; in nitrox_aes_xts_setkey() local
294 fctx = nctx->u.fctx; in nitrox_aes_xts_setkey()
296 memcpy(fctx->auth.u.key2, (key + keylen), keylen); in nitrox_aes_xts_setkey()
[all …]
Dnitrox_req.h203 struct flexi_crypto_context *fctx; member
/Linux-v5.4/drivers/crypto/cavium/cpt/
Dcptvf_algs.c101 struct fc_context *fctx = &rctx->fctx; in create_ctx_hdr() local
121 fctx->enc.enc_ctrl.e.enc_cipher = ctx->cipher_type; in create_ctx_hdr()
122 fctx->enc.enc_ctrl.e.aes_key = ctx->key_type; in create_ctx_hdr()
123 fctx->enc.enc_ctrl.e.iv_source = FROM_DPTR; in create_ctx_hdr()
126 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len * 2); in create_ctx_hdr()
128 memcpy(fctx->enc.encr_key, ctx->enc_key, ctx->key_len); in create_ctx_hdr()
129 ctrl_flags = (u64 *)&fctx->enc.enc_ctrl.flags; in create_ctx_hdr()
141 req_info->in[*argcnt].vptr = (u8 *)fctx; in create_ctx_hdr()
197 struct fc_context *fctx = &rctx->fctx; in cvm_enc_dec() local
203 memset(fctx, 0, sizeof(struct fc_context)); in cvm_enc_dec()
Dcptvf_algs.h113 struct fc_context fctx; member
/Linux-v5.4/security/apparmor/
Dfile.c487 static void update_file_ctx(struct aa_file_ctx *fctx, struct aa_label *label, in update_file_ctx() argument
493 spin_lock(&fctx->lock); in update_file_ctx()
494 old = rcu_dereference_protected(fctx->label, in update_file_ctx()
495 lockdep_is_held(&fctx->lock)); in update_file_ctx()
499 rcu_assign_pointer(fctx->label, l); in update_file_ctx()
503 fctx->allow |= request; in update_file_ctx()
505 spin_unlock(&fctx->lock); in update_file_ctx()
599 struct aa_file_ctx *fctx; in aa_file_perm() local
607 fctx = file_ctx(file); in aa_file_perm()
610 flabel = rcu_dereference(fctx->label); in aa_file_perm()
[all …]
Dlsm.c394 struct aa_file_ctx *fctx = file_ctx(file); in apparmor_file_open() local
407 fctx->allow = MAY_EXEC | MAY_READ | AA_EXEC_MMAP; in apparmor_file_open()
419 fctx->allow = aa_map_file_to_perms(file); in apparmor_file_open()
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Ddmanv04.c79 struct nvkm_memory *fctx = device->imem->ramfc; in nv04_fifo_dma_fini() local
98 nvkm_kmap(fctx); in nv04_fifo_dma_fini()
103 u32 cv = (nvkm_ro32(fctx, c->ctxp + data) & ~cm); in nv04_fifo_dma_fini()
104 nvkm_wo32(fctx, c->ctxp + data, cv | (rv << c->ctxs)); in nv04_fifo_dma_fini()
106 nvkm_done(fctx); in nv04_fifo_dma_fini()
/Linux-v5.4/fs/
Duserfaultfd.c674 struct userfaultfd_fork_ctx *fctx; in dup_userfaultfd() local
683 list_for_each_entry(fctx, fcs, list) in dup_userfaultfd()
684 if (fctx->orig == octx) { in dup_userfaultfd()
685 ctx = fctx->new; in dup_userfaultfd()
690 fctx = kmalloc(sizeof(*fctx), GFP_KERNEL); in dup_userfaultfd()
691 if (!fctx) in dup_userfaultfd()
696 kfree(fctx); in dup_userfaultfd()
711 fctx->orig = octx; in dup_userfaultfd()
712 fctx->new = ctx; in dup_userfaultfd()
713 list_add_tail(&fctx->list, fcs); in dup_userfaultfd()
[all …]
/Linux-v5.4/drivers/block/
Drbd.c2637 struct rbd_img_fill_ctx *fctx) in rbd_img_fill_request_nocopy() argument
2642 img_req->data_type = fctx->pos_type; in rbd_img_fill_request_nocopy()
2648 fctx->iter = *fctx->pos; in rbd_img_fill_request_nocopy()
2655 fctx->set_pos_fn, &fctx->iter); in rbd_img_fill_request_nocopy()
2679 struct rbd_img_fill_ctx *fctx) in rbd_img_fill_request() argument
2686 if (fctx->pos_type == OBJ_REQUEST_NODATA || in rbd_img_fill_request()
2689 num_img_extents, fctx); in rbd_img_fill_request()
2700 fctx->iter = *fctx->pos; in rbd_img_fill_request()
2707 fctx->count_fn, &fctx->iter); in rbd_img_fill_request()
2724 fctx->iter = *fctx->pos; in rbd_img_fill_request()
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/dispnv04/
Dcrtc.c1045 struct nouveau_fence_chan *fctx = chan->fence; in nv04_finish_page_flip() local
1053 if (list_empty(&fctx->flip)) { in nv04_finish_page_flip()
1059 s = list_first_entry(&fctx->flip, struct nv04_page_flip_state, head); in nv04_finish_page_flip()
1101 struct nouveau_fence_chan *fctx = chan->fence; in nv04_page_flip_emit() local
1109 list_add_tail(&s->head, &fctx->flip); in nv04_page_flip_emit()
/Linux-v5.4/fs/fuse/
Dcontrol.c321 static int fuse_ctl_fill_super(struct super_block *sb, struct fs_context *fctx) in fuse_ctl_fill_super() argument

12