Home
last modified time | relevance | path

Searched refs:gctx (Results 1 – 9 of 9) sorted by relevance

/Linux-v4.19/arch/x86/crypto/
Dglue_helper.c35 int glue_ecb_req_128bit(const struct common_glue_ctx *gctx, in glue_ecb_req_128bit() argument
53 fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, in glue_ecb_req_128bit()
55 for (i = 0; i < gctx->num_funcs; i++) { in glue_ecb_req_128bit()
56 func_bytes = bsize * gctx->funcs[i].num_blocks; in glue_ecb_req_128bit()
63 gctx->funcs[i].fn_u.ecb(ctx, dst, src); in glue_ecb_req_128bit()
112 int glue_cbc_decrypt_req_128bit(const struct common_glue_ctx *gctx, in glue_cbc_decrypt_req_128bit() argument
131 fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, in glue_cbc_decrypt_req_128bit()
139 for (i = 0; i < gctx->num_funcs; i++) { in glue_cbc_decrypt_req_128bit()
140 num_blocks = gctx->funcs[i].num_blocks; in glue_cbc_decrypt_req_128bit()
151 gctx->funcs[i].fn_u.cbc(ctx, dst, src); in glue_cbc_decrypt_req_128bit()
[all …]
/Linux-v4.19/drivers/gpu/drm/radeon/
Datom.c182 struct atom_context *gctx = ctx->ctx; in atom_get_src_int() local
191 idx += gctx->reg_block; in atom_get_src_int()
192 switch (gctx->io_mode) { in atom_get_src_int()
194 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int()
203 if (!(gctx->io_mode & 0x80)) { in atom_get_src_int()
207 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int()
209 gctx->io_mode & 0x7F); in atom_get_src_int()
213 atom_iio_execute(gctx, in atom_get_src_int()
214 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int()
234 val = gctx->divmul[0]; in atom_get_src_int()
[all …]
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Datom.c178 struct atom_context *gctx = ctx->ctx; in atom_get_src_int() local
187 idx += gctx->reg_block; in atom_get_src_int()
188 switch (gctx->io_mode) { in atom_get_src_int()
190 val = gctx->card->reg_read(gctx->card, idx); in atom_get_src_int()
199 if (!(gctx->io_mode & 0x80)) { in atom_get_src_int()
203 if (!gctx->iio[gctx->io_mode & 0x7F]) { in atom_get_src_int()
205 gctx->io_mode & 0x7F); in atom_get_src_int()
209 atom_iio_execute(gctx, in atom_get_src_int()
210 gctx->iio[gctx->io_mode & 0x7F], in atom_get_src_int()
230 val = gctx->divmul[0]; in atom_get_src_int()
[all …]
/Linux-v4.19/arch/x86/include/asm/crypto/
Dglue_helper.h102 extern int glue_ecb_req_128bit(const struct common_glue_ctx *gctx,
108 extern int glue_cbc_decrypt_req_128bit(const struct common_glue_ctx *gctx,
111 extern int glue_ctr_req_128bit(const struct common_glue_ctx *gctx,
114 extern int glue_xts_req_128bit(const struct common_glue_ctx *gctx,
/Linux-v4.19/crypto/
Dgcm.c249 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_len() local
253 lengths.b = cpu_to_be64(gctx->cryptlen * 8); in gcm_hash_len()
266 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_len_continue() local
268 return gctx->complete(req, flags); in gcm_hash_len_continue()
311 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_crypt_continue() local
314 remain = gcm_remain(gctx->cryptlen); in gcm_hash_crypt_continue()
341 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_hash_assoc_remain_continue() local
343 if (gctx->cryptlen) in gcm_hash_assoc_remain_continue()
345 gctx->src, gctx->cryptlen, flags) ?: in gcm_hash_assoc_remain_continue()
449 struct crypto_gcm_ghash_ctx *gctx = &pctx->ghash_ctx; in gcm_encrypt_continue() local
[all …]
/Linux-v4.19/drivers/crypto/mediatek/
Dmtk-aes.c821 struct mtk_aes_gcm_ctx *gctx = mtk_aes_gcm_ctx_cast(ctx); in mtk_aes_gcm_info_init() local
831 info->cmd[cnt++] = AES_GCM_CMD3 | cpu_to_le32(gctx->textlen); in mtk_aes_gcm_info_init()
834 info->cmd[cnt++] = AES_GCM_CMD4 | cpu_to_le32(gctx->authsize); in mtk_aes_gcm_info_init()
837 info->cmd[cnt++] = AES_GCM_CMD5 | cpu_to_le32(gctx->authsize); in mtk_aes_gcm_info_init()
838 info->cmd[cnt++] = AES_GCM_CMD6 | cpu_to_le32(gctx->authsize); in mtk_aes_gcm_info_init()
898 struct mtk_aes_gcm_ctx *gctx = mtk_aes_gcm_ctx_cast(aes->ctx); in mtk_aes_gcm_start() local
910 aes->total = len + gctx->authsize; in mtk_aes_gcm_start()
912 gctx->textlen = req->cryptlen; in mtk_aes_gcm_start()
914 scatterwalk_map_and_copy(tag, req->dst, len, gctx->authsize, 1); in mtk_aes_gcm_start()
918 gctx->textlen = req->cryptlen - gctx->authsize; in mtk_aes_gcm_start()
[all …]
/Linux-v4.19/net/sunrpc/auth_gss/
Dgss_krb5_wrap.c589 gss_wrap_kerberos(struct gss_ctx *gctx, int offset, in gss_wrap_kerberos() argument
592 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_wrap_kerberos()
608 gss_unwrap_kerberos(struct gss_ctx *gctx, int offset, struct xdr_buf *buf) in gss_unwrap_kerberos() argument
610 struct krb5_ctx *kctx = gctx->internal_ctx_id; in gss_unwrap_kerberos()
/Linux-v4.19/kernel/events/
Dcore.c10360 struct perf_event_context *gctx; in __perf_event_ctx_lock_double() local
10364 gctx = READ_ONCE(group_leader->ctx); in __perf_event_ctx_lock_double()
10365 if (!atomic_inc_not_zero(&gctx->refcount)) { in __perf_event_ctx_lock_double()
10371 mutex_lock_double(&gctx->mutex, &ctx->mutex); in __perf_event_ctx_lock_double()
10373 if (group_leader->ctx != gctx) { in __perf_event_ctx_lock_double()
10375 mutex_unlock(&gctx->mutex); in __perf_event_ctx_lock_double()
10376 put_ctx(gctx); in __perf_event_ctx_lock_double()
10380 return gctx; in __perf_event_ctx_lock_double()
10398 struct perf_event_context *ctx, *uninitialized_var(gctx); in SYSCALL_DEFINE5()
10631 gctx = __perf_event_ctx_lock_double(group_leader, ctx); in SYSCALL_DEFINE5()
[all …]
/Linux-v4.19/drivers/crypto/chelsio/
Dchcr_algo.c114 static inline struct chcr_gcm_ctx *GCM_CTX(struct chcr_aead_ctx *gctx) in GCM_CTX() argument
116 return gctx->ctx->gcm; in GCM_CTX()
119 static inline struct chcr_authenc_ctx *AUTHENC_CTX(struct chcr_aead_ctx *gctx) in AUTHENC_CTX() argument
121 return gctx->ctx->authenc; in AUTHENC_CTX()
3295 struct chcr_gcm_ctx *gctx = GCM_CTX(aeadctx); in chcr_gcm_setkey() local
3353 memset(gctx->ghash_h, 0, AEAD_H_SIZE); in chcr_gcm_setkey()
3354 crypto_cipher_encrypt_one(cipher, gctx->ghash_h, gctx->ghash_h); in chcr_gcm_setkey()