Home
last modified time | relevance | path

Searched refs:rctx (Results 1 – 25 of 60) sorted by relevance

123

/Linux-v5.4/drivers/crypto/ccp/
Dccp-crypto-aes-cmac.c28 struct ccp_aes_cmac_req_ctx *rctx = ahash_request_ctx(req); in ccp_aes_cmac_complete() local
34 if (rctx->hash_rem) { in ccp_aes_cmac_complete()
36 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_aes_cmac_complete()
38 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_aes_cmac_complete()
39 offset, rctx->hash_rem, 0); in ccp_aes_cmac_complete()
40 rctx->buf_count = rctx->hash_rem; in ccp_aes_cmac_complete()
42 rctx->buf_count = 0; in ccp_aes_cmac_complete()
46 if (req->result && rctx->final) in ccp_aes_cmac_complete()
47 memcpy(req->result, rctx->iv, digest_size); in ccp_aes_cmac_complete()
50 sg_free_table(&rctx->data_sg); in ccp_aes_cmac_complete()
[all …]
Dccp-crypto-sha.c29 struct ccp_sha_req_ctx *rctx = ahash_request_ctx(req); in ccp_sha_complete() local
35 if (rctx->hash_rem) { in ccp_sha_complete()
37 unsigned int offset = rctx->nbytes - rctx->hash_rem; in ccp_sha_complete()
39 scatterwalk_map_and_copy(rctx->buf, rctx->src, in ccp_sha_complete()
40 offset, rctx->hash_rem, 0); in ccp_sha_complete()
41 rctx->buf_count = rctx->hash_rem; in ccp_sha_complete()
43 rctx->buf_count = 0; in ccp_sha_complete()
47 if (req->result && rctx->final) in ccp_sha_complete()
48 memcpy(req->result, rctx->ctx, digest_size); in ccp_sha_complete()
51 sg_free_table(&rctx->data_sg); in ccp_sha_complete()
[all …]
Dccp-crypto-aes-galois.c81 struct ccp_aes_req_ctx *rctx = aead_request_ctx(req); in ccp_aes_gcm_crypt() local
106 memcpy(rctx->iv, req->iv, GCM_AES_IV_SIZE); in ccp_aes_gcm_crypt()
108 rctx->iv[i + GCM_AES_IV_SIZE] = 0; in ccp_aes_gcm_crypt()
109 rctx->iv[AES_BLOCK_SIZE - 1] = 1; in ccp_aes_gcm_crypt()
112 iv_sg = &rctx->iv_sg; in ccp_aes_gcm_crypt()
114 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_gcm_crypt()
117 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_gcm_crypt()
118 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_gcm_crypt()
119 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_gcm_crypt()
120 rctx->cmd.u.aes.authsize = crypto_aead_authsize(tfm); in ccp_aes_gcm_crypt()
[all …]
Dccp-crypto-des3.c25 struct ccp_des3_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_des3_complete() local
31 memcpy(req->info, rctx->iv, DES3_EDE_BLOCK_SIZE); in ccp_des3_complete()
64 struct ccp_des3_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_des3_crypt() local
81 memcpy(rctx->iv, req->info, DES3_EDE_BLOCK_SIZE); in ccp_des3_crypt()
82 iv_sg = &rctx->iv_sg; in ccp_des3_crypt()
84 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_des3_crypt()
87 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_des3_crypt()
88 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_des3_crypt()
89 rctx->cmd.engine = CCP_ENGINE_DES3; in ccp_des3_crypt()
90 rctx->cmd.u.des3.type = ctx->u.des3.type; in ccp_des3_crypt()
[all …]
Dccp-crypto-aes.c26 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_complete() local
32 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_complete()
70 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_crypt() local
87 memcpy(rctx->iv, req->info, AES_BLOCK_SIZE); in ccp_aes_crypt()
88 iv_sg = &rctx->iv_sg; in ccp_aes_crypt()
90 sg_init_one(iv_sg, rctx->iv, iv_len); in ccp_aes_crypt()
93 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_crypt()
94 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_crypt()
95 rctx->cmd.engine = CCP_ENGINE_AES; in ccp_aes_crypt()
96 rctx->cmd.u.aes.type = ctx->u.aes.type; in ccp_aes_crypt()
[all …]
Dccp-crypto-aes-xts.c65 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_complete() local
70 memcpy(req->info, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_complete()
109 struct ccp_aes_req_ctx *rctx = ablkcipher_request_ctx(req); in ccp_aes_xts_crypt() local
165 memcpy(rctx->iv, req->info, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
166 sg_init_one(&rctx->iv_sg, rctx->iv, AES_BLOCK_SIZE); in ccp_aes_xts_crypt()
168 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_aes_xts_crypt()
169 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_aes_xts_crypt()
170 rctx->cmd.engine = CCP_ENGINE_XTS_AES_128; in ccp_aes_xts_crypt()
171 rctx->cmd.u.xts.type = CCP_AES_TYPE_128; in ccp_aes_xts_crypt()
172 rctx->cmd.u.xts.action = (encrypt) ? CCP_AES_ACTION_ENCRYPT in ccp_aes_xts_crypt()
[all …]
Dccp-crypto-rsa.c47 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx(req); in ccp_rsa_complete() local
52 req->dst_len = rctx->cmd.u.rsa.key_size >> 3; in ccp_rsa_complete()
68 struct ccp_rsa_req_ctx *rctx = akcipher_request_ctx(req); in ccp_rsa_crypt() local
71 memset(&rctx->cmd, 0, sizeof(rctx->cmd)); in ccp_rsa_crypt()
72 INIT_LIST_HEAD(&rctx->cmd.entry); in ccp_rsa_crypt()
73 rctx->cmd.engine = CCP_ENGINE_RSA; in ccp_rsa_crypt()
75 rctx->cmd.u.rsa.key_size = ctx->u.rsa.key_len; /* in bits */ in ccp_rsa_crypt()
77 rctx->cmd.u.rsa.exp = &ctx->u.rsa.e_sg; in ccp_rsa_crypt()
78 rctx->cmd.u.rsa.exp_len = ctx->u.rsa.e_len; in ccp_rsa_crypt()
80 rctx->cmd.u.rsa.exp = &ctx->u.rsa.d_sg; in ccp_rsa_crypt()
[all …]
/Linux-v5.4/drivers/crypto/qce/
Dsha.c34 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_ahash_done() local
46 dma_unmap_sg(qce->dev, req->src, rctx->src_nents, DMA_TO_DEVICE); in qce_ahash_done()
47 dma_unmap_sg(qce->dev, &rctx->result_sg, 1, DMA_FROM_DEVICE); in qce_ahash_done()
49 memcpy(rctx->digest, result->auth_iv, digestsize); in qce_ahash_done()
53 rctx->byte_count[0] = cpu_to_be32(result->auth_byte_count[0]); in qce_ahash_done()
54 rctx->byte_count[1] = cpu_to_be32(result->auth_byte_count[1]); in qce_ahash_done()
60 req->src = rctx->src_orig; in qce_ahash_done()
61 req->nbytes = rctx->nbytes_orig; in qce_ahash_done()
62 rctx->last_blk = false; in qce_ahash_done()
63 rctx->first_blk = false; in qce_ahash_done()
[all …]
Dablkcipher.c21 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_done() local
39 dma_unmap_sg(qce->dev, rctx->src_sg, rctx->src_nents, dir_src); in qce_ablkcipher_done()
40 dma_unmap_sg(qce->dev, rctx->dst_sg, rctx->dst_nents, dir_dst); in qce_ablkcipher_done()
42 sg_free_table(&rctx->dst_tbl); in qce_ablkcipher_done()
55 struct qce_cipher_reqctx *rctx = ablkcipher_request_ctx(req); in qce_ablkcipher_async_req_handle() local
65 rctx->iv = req->info; in qce_ablkcipher_async_req_handle()
66 rctx->ivsize = crypto_ablkcipher_ivsize(ablkcipher); in qce_ablkcipher_async_req_handle()
67 rctx->cryptlen = req->nbytes; in qce_ablkcipher_async_req_handle()
73 rctx->src_nents = sg_nents_for_len(req->src, req->nbytes); in qce_ablkcipher_async_req_handle()
75 rctx->dst_nents = sg_nents_for_len(req->dst, req->nbytes); in qce_ablkcipher_async_req_handle()
[all …]
Dcommon.c227 struct qce_sha_reqctx *rctx = ahash_request_ctx(req); in qce_setup_regs_ahash() local
238 if (!rctx->last_blk && req->nbytes % blocksize) in qce_setup_regs_ahash()
243 if (IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
251 auth_cfg = qce_auth_cfg(rctx->flags, rctx->authklen); in qce_setup_regs_ahash()
254 if (IS_SHA_HMAC(rctx->flags) || IS_CMAC(rctx->flags)) { in qce_setup_regs_ahash()
255 u32 authkey_words = rctx->authklen / sizeof(u32); in qce_setup_regs_ahash()
257 qce_cpu_to_be32p_array(mackey, rctx->authkey, rctx->authklen); in qce_setup_regs_ahash()
262 if (IS_CMAC(rctx->flags)) in qce_setup_regs_ahash()
265 if (rctx->first_blk) in qce_setup_regs_ahash()
266 memcpy(auth, rctx->digest, digestsize); in qce_setup_regs_ahash()
[all …]
/Linux-v5.4/drivers/crypto/cavium/nitrox/
Dnitrox_aead.c135 static int nitrox_set_creq(struct nitrox_aead_rctx *rctx) in nitrox_set_creq() argument
137 struct se_crypto_request *creq = &rctx->nkreq.creq; in nitrox_set_creq()
141 creq->flags = rctx->flags; in nitrox_set_creq()
142 creq->gfp = (rctx->flags & CRYPTO_TFM_REQ_MAY_SLEEP) ? GFP_KERNEL : in nitrox_set_creq()
147 creq->ctrl.s.arg = rctx->ctrl_arg; in nitrox_set_creq()
149 creq->gph.param0 = cpu_to_be16(rctx->cryptlen); in nitrox_set_creq()
150 creq->gph.param1 = cpu_to_be16(rctx->cryptlen + rctx->assoclen); in nitrox_set_creq()
151 creq->gph.param2 = cpu_to_be16(rctx->ivsize + rctx->assoclen); in nitrox_set_creq()
153 param3.auth_offset = rctx->ivsize; in nitrox_set_creq()
156 creq->ctx_handle = rctx->ctx_handle; in nitrox_set_creq()
[all …]
/Linux-v5.4/drivers/crypto/bcm/
Dcipher.c134 struct iproc_reqctx_s *rctx, in spu_ablkcipher_rx_sg_create() argument
140 struct iproc_ctx_s *ctx = rctx->ctx; in spu_ablkcipher_rx_sg_create()
144 rctx->gfp); in spu_ablkcipher_rx_sg_create()
151 sg_set_buf(sg++, rctx->msg_buf.spu_resp_hdr, ctx->spu_resp_hdr_len); in spu_ablkcipher_rx_sg_create()
156 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, in spu_ablkcipher_rx_sg_create()
160 datalen = spu_msg_sg_add(&sg, &rctx->dst_sg, &rctx->dst_skip, in spu_ablkcipher_rx_sg_create()
161 rctx->dst_nents, chunksize); in spu_ablkcipher_rx_sg_create()
170 sg_set_buf(sg++, rctx->msg_buf.c.supdt_tweak, SPU_SUPDT_LEN); in spu_ablkcipher_rx_sg_create()
173 sg_set_buf(sg++, rctx->msg_buf.rx_stat_pad, stat_pad_len); in spu_ablkcipher_rx_sg_create()
175 memset(rctx->msg_buf.rx_stat, 0, SPU_RX_STATUS_LEN); in spu_ablkcipher_rx_sg_create()
[all …]
/Linux-v5.4/drivers/crypto/stm32/
Dstm32-hash.c252 struct stm32_hash_request_ctx *rctx = ahash_request_ctx(hdev->req); in stm32_hash_write_ctrl() local
259 switch (rctx->flags & HASH_FLAGS_ALGO_MASK) { in stm32_hash_write_ctrl()
276 reg |= (rctx->data_type << HASH_CR_DATATYPE_POS); in stm32_hash_write_ctrl()
278 if (rctx->flags & HASH_FLAGS_HMAC) { in stm32_hash_write_ctrl()
295 static void stm32_hash_append_sg(struct stm32_hash_request_ctx *rctx) in stm32_hash_append_sg() argument
299 while ((rctx->bufcnt < rctx->buflen) && rctx->total) { in stm32_hash_append_sg()
300 count = min(rctx->sg->length - rctx->offset, rctx->total); in stm32_hash_append_sg()
301 count = min(count, rctx->buflen - rctx->bufcnt); in stm32_hash_append_sg()
304 if ((rctx->sg->length == 0) && !sg_is_last(rctx->sg)) { in stm32_hash_append_sg()
305 rctx->sg = sg_next(rctx->sg); in stm32_hash_append_sg()
[all …]
/Linux-v5.4/crypto/
Dchacha20poly1305.c76 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in async_done_continue() local
78 rctx->flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in async_done_continue()
99 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_verify_tag() local
100 u8 tag[sizeof(rctx->tag)]; in poly_verify_tag()
103 req->assoclen + rctx->cryptlen, in poly_verify_tag()
105 if (crypto_memneq(tag, rctx->tag, sizeof(tag))) in poly_verify_tag()
112 struct chachapoly_req_ctx *rctx = aead_request_ctx(req); in poly_copy_tag() local
114 scatterwalk_map_and_copy(rctx->tag, req->dst, in poly_copy_tag()
115 req->assoclen + rctx->cryptlen, in poly_copy_tag()
116 sizeof(rctx->tag), 1); in poly_copy_tag()
[all …]
Dxts.c33 struct rctx { struct
89 struct rctx *rctx = skcipher_request_ctx(req); in xor_tweak() local
94 le128 t = rctx->t; in xor_tweak()
98 req = &rctx->subreq; in xor_tweak()
117 rctx->t = t; in xor_tweak()
122 gf128mul_x_ble(&rctx->t, &t); in xor_tweak()
153 struct rctx *rctx = skcipher_request_ctx(req); in cts_done() local
155 scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 0); in cts_done()
156 le128_xor(&b, &rctx->t, &b); in cts_done()
157 scatterwalk_map_and_copy(&b, rctx->tail, 0, XTS_BLOCK_SIZE, 1); in cts_done()
[all …]
Dlrw.c52 struct rctx { struct
147 struct rctx *rctx = skcipher_request_ctx(req); in xor_tweak() local
148 be128 t = rctx->t; in xor_tweak()
155 req = &rctx->subreq; in xor_tweak()
214 struct rctx *rctx = skcipher_request_ctx(req); in crypt_done() local
216 rctx->subreq.base.flags &= ~CRYPTO_TFM_REQ_MAY_SLEEP; in crypt_done()
226 struct rctx *rctx = skcipher_request_ctx(req); in init_crypt() local
227 struct skcipher_request *subreq = &rctx->subreq; in init_crypt()
236 memcpy(&rctx->t, req->iv, sizeof(rctx->t)); in init_crypt()
239 gf128mul_64k_bbe(&rctx->t, ctx->table); in init_crypt()
[all …]
Drmd256.c230 struct rmd256_ctx *rctx = shash_desc_ctx(desc); in rmd256_init() local
232 rctx->byte_count = 0; in rmd256_init()
234 rctx->state[0] = RMD_H0; in rmd256_init()
235 rctx->state[1] = RMD_H1; in rmd256_init()
236 rctx->state[2] = RMD_H2; in rmd256_init()
237 rctx->state[3] = RMD_H3; in rmd256_init()
238 rctx->state[4] = RMD_H5; in rmd256_init()
239 rctx->state[5] = RMD_H6; in rmd256_init()
240 rctx->state[6] = RMD_H7; in rmd256_init()
241 rctx->state[7] = RMD_H8; in rmd256_init()
[all …]
Dadiantum.c231 struct adiantum_request_ctx *rctx = skcipher_request_ctx(req); in adiantum_hash_header() local
251 poly1305_core_emit(&state, &rctx->header_hash); in adiantum_hash_header()
260 struct adiantum_request_ctx *rctx = skcipher_request_ctx(req); in adiantum_hash_message() local
262 struct shash_desc *hash_desc = &rctx->u.hash_desc; in adiantum_hash_message()
294 struct adiantum_request_ctx *rctx = skcipher_request_ctx(req); in adiantum_finish() local
300 if (!rctx->enc) in adiantum_finish()
301 crypto_cipher_decrypt_one(tctx->blockcipher, rctx->rbuf.bytes, in adiantum_finish()
302 rctx->rbuf.bytes); in adiantum_finish()
312 le128_add(&digest, &digest, &rctx->header_hash); in adiantum_finish()
313 le128_sub(&rctx->rbuf.bignum, &rctx->rbuf.bignum, &digest); in adiantum_finish()
[all …]
Drmd128.c215 struct rmd128_ctx *rctx = shash_desc_ctx(desc); in rmd128_init() local
217 rctx->byte_count = 0; in rmd128_init()
219 rctx->state[0] = RMD_H0; in rmd128_init()
220 rctx->state[1] = RMD_H1; in rmd128_init()
221 rctx->state[2] = RMD_H2; in rmd128_init()
222 rctx->state[3] = RMD_H3; in rmd128_init()
224 memset(rctx->buffer, 0, sizeof(rctx->buffer)); in rmd128_init()
232 struct rmd128_ctx *rctx = shash_desc_ctx(desc); in rmd128_update() local
233 const u32 avail = sizeof(rctx->buffer) - (rctx->byte_count & 0x3f); in rmd128_update()
235 rctx->byte_count += len; in rmd128_update()
[all …]
Drmd320.c277 struct rmd320_ctx *rctx = shash_desc_ctx(desc); in rmd320_init() local
279 rctx->byte_count = 0; in rmd320_init()
281 rctx->state[0] = RMD_H0; in rmd320_init()
282 rctx->state[1] = RMD_H1; in rmd320_init()
283 rctx->state[2] = RMD_H2; in rmd320_init()
284 rctx->state[3] = RMD_H3; in rmd320_init()
285 rctx->state[4] = RMD_H4; in rmd320_init()
286 rctx->state[5] = RMD_H5; in rmd320_init()
287 rctx->state[6] = RMD_H6; in rmd320_init()
288 rctx->state[7] = RMD_H7; in rmd320_init()
[all …]
/Linux-v5.4/drivers/crypto/rockchip/
Drk3288_crypto_ahash.c49 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); in rk_ahash_reg_init() local
69 CRYPTO_WRITE(dev, RK_CRYPTO_HASH_CTRL, rctx->mode | in rk_ahash_reg_init()
81 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); in rk_ahash_init() local
85 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm); in rk_ahash_init()
86 rctx->fallback_req.base.flags = req->base.flags & in rk_ahash_init()
89 return crypto_ahash_init(&rctx->fallback_req); in rk_ahash_init()
94 struct rk_ahash_rctx *rctx = ahash_request_ctx(req); in rk_ahash_update() local
98 ahash_request_set_tfm(&rctx->fallback_req, ctx->fallback_tfm); in rk_ahash_update()
99 rctx->fallback_req.base.flags = req->base.flags & in rk_ahash_update()
101 rctx->fallback_req.nbytes = req->nbytes; in rk_ahash_update()
[all …]
/Linux-v5.4/drivers/crypto/
Dsahara.c554 struct sahara_aes_reqctx *rctx; in sahara_aes_process() local
568 rctx = ablkcipher_request_ctx(req); in sahara_aes_process()
570 rctx->mode &= FLAGS_MODE_MASK; in sahara_aes_process()
571 dev->flags = (dev->flags & ~FLAGS_MODE_MASK) | rctx->mode; in sahara_aes_process()
635 struct sahara_aes_reqctx *rctx = ablkcipher_request_ctx(req); in sahara_aes_crypt() local
648 rctx->mode = mode; in sahara_aes_crypt()
772 struct sahara_sha_reqctx *rctx) in sahara_sha_init_hdr() argument
776 hdr = rctx->mode; in sahara_sha_init_hdr()
778 if (rctx->first) { in sahara_sha_init_hdr()
785 if (rctx->last) in sahara_sha_init_hdr()
[all …]
Domap-aes-gcm.c43 struct omap_aes_reqctx *rctx; in omap_aes_gcm_done_task() local
47 rctx = aead_request_ctx(dd->aead_req); in omap_aes_gcm_done_task()
62 scatterwalk_map_and_copy(rctx->auth_tag, in omap_aes_gcm_done_task()
74 tag = (u8 *)rctx->auth_tag; in omap_aes_gcm_done_task()
225 struct omap_aes_reqctx *rctx; in omap_aes_gcm_dma_out_callback() local
234 rctx = aead_request_ctx(dd->aead_req); in omap_aes_gcm_dma_out_callback()
235 auth_tag = (u32 *)rctx->auth_tag; in omap_aes_gcm_dma_out_callback()
251 struct omap_aes_reqctx *rctx; in omap_aes_gcm_handle_queue() local
276 rctx = aead_request_ctx(req); in omap_aes_gcm_handle_queue()
279 rctx->dd = dd; in omap_aes_gcm_handle_queue()
[all …]
/Linux-v5.4/kernel/events/
Dcallchain.c152 static struct perf_callchain_entry *get_callchain_entry(int *rctx) in get_callchain_entry() argument
157 *rctx = get_recursion_context(this_cpu_ptr(callchain_recursion)); in get_callchain_entry()
158 if (*rctx == -1) in get_callchain_entry()
168 (*rctx * perf_callchain_entry__sizeof())); in get_callchain_entry()
172 put_callchain_entry(int rctx) in put_callchain_entry() argument
174 put_recursion_context(this_cpu_ptr(callchain_recursion), rctx); in put_callchain_entry()
183 int rctx; in get_perf_callchain() local
185 entry = get_callchain_entry(&rctx); in get_perf_callchain()
186 if (rctx == -1) in get_perf_callchain()
229 put_callchain_entry(rctx); in get_perf_callchain()
Dinternal.h207 int rctx; in DEFINE_OUTPUT_COPY() local
210 rctx = 3; in DEFINE_OUTPUT_COPY()
212 rctx = 2; in DEFINE_OUTPUT_COPY()
214 rctx = 1; in DEFINE_OUTPUT_COPY()
216 rctx = 0; in DEFINE_OUTPUT_COPY()
218 if (recursion[rctx]) in DEFINE_OUTPUT_COPY()
221 recursion[rctx]++; in DEFINE_OUTPUT_COPY()
224 return rctx; in DEFINE_OUTPUT_COPY()
227 static inline void put_recursion_context(int *recursion, int rctx) in put_recursion_context() argument
230 recursion[rctx]--; in put_recursion_context()

123