Home
last modified time | relevance | path

Searched refs:SHA256_BLOCK_SIZE (Results 1 – 25 of 41) sorted by relevance

12

/Linux-v4.19/include/crypto/
Dsha256_base.h61 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_update()
65 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in sha256_base_do_update()
69 int p = SHA256_BLOCK_SIZE - partial; in sha256_base_do_update()
78 blocks = len / SHA256_BLOCK_SIZE; in sha256_base_do_update()
79 len %= SHA256_BLOCK_SIZE; in sha256_base_do_update()
83 data += blocks * SHA256_BLOCK_SIZE; in sha256_base_do_update()
96 const int bit_offset = SHA256_BLOCK_SIZE - sizeof(__be64); in sha256_base_do_finalize()
99 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_finalize()
103 memset(sctx->buf + partial, 0x0, SHA256_BLOCK_SIZE - partial); in sha256_base_do_finalize()
Dsha.h18 #define SHA256_BLOCK_SIZE 64 macro
87 u8 buf[SHA256_BLOCK_SIZE];
/Linux-v4.19/arch/sparc/crypto/
Dsha256_glue.c68 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update()
72 if (len - done >= SHA256_BLOCK_SIZE) { in __sha256_sparc64_update()
73 const unsigned int rounds = (len - done) / SHA256_BLOCK_SIZE; in __sha256_sparc64_update()
76 done += rounds * SHA256_BLOCK_SIZE; in __sha256_sparc64_update()
86 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update()
89 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update()
104 static const u8 padding[SHA256_BLOCK_SIZE] = { 0x80, }; in sha256_sparc64_final()
109 index = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_final()
110 padlen = (index < 56) ? (56 - index) : ((SHA256_BLOCK_SIZE+56) - index); in sha256_sparc64_final()
172 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/mips/cavium-octeon/crypto/
Docteon-sha256.c111 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update()
116 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update()
120 done + SHA256_BLOCK_SIZE); in __octeon_sha256_update()
126 done += SHA256_BLOCK_SIZE; in __octeon_sha256_update()
128 } while (done + SHA256_BLOCK_SIZE <= len); in __octeon_sha256_update()
147 if ((sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in octeon_sha256_update()
242 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/drivers/crypto/nx/
Dnx-sha256.c80 u64 buf_len = (sctx->count % SHA256_BLOCK_SIZE); in nx_sha256_update()
88 total = (sctx->count % SHA256_BLOCK_SIZE) + len; in nx_sha256_update()
89 if (total < SHA256_BLOCK_SIZE) { in nx_sha256_update()
141 to_process = to_process & ~(SHA256_BLOCK_SIZE - 1); in nx_sha256_update()
176 } while (leftover >= SHA256_BLOCK_SIZE); in nx_sha256_update()
209 if (sctx->count >= SHA256_BLOCK_SIZE) { in nx_sha256_final()
222 len = sctx->count & (SHA256_BLOCK_SIZE - 1); in nx_sha256_final()
226 if (len != (sctx->count & (SHA256_BLOCK_SIZE - 1))) { in nx_sha256_final()
291 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/arm64/crypto/
Dsha256-glue.c71 .base.cra_blocksize = SHA256_BLOCK_SIZE,
105 chunk + sctx->count % SHA256_BLOCK_SIZE > SHA256_BLOCK_SIZE) in sha256_update_neon()
106 chunk = SHA256_BLOCK_SIZE - in sha256_update_neon()
107 sctx->count % SHA256_BLOCK_SIZE; in sha256_update_neon()
154 .base.cra_blocksize = SHA256_BLOCK_SIZE,
Dsha2-ce-glue.c62 bool finalize = !sctx->sst.count && !(len % SHA256_BLOCK_SIZE); in sha256_ce_finup()
117 .cra_blocksize = SHA256_BLOCK_SIZE,
131 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/arm/crypto/
Dsha2-ce-glue.c38 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha2_ce_update()
81 .cra_blocksize = SHA256_BLOCK_SIZE,
95 .cra_blocksize = SHA256_BLOCK_SIZE,
Dsha256_neon_glue.c38 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha256_update()
82 .cra_blocksize = SHA256_BLOCK_SIZE,
Dsha256_glue.c74 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/x86/crypto/
Dsha256_ssse3_glue.c53 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha256_update()
112 .cra_blocksize = SHA256_BLOCK_SIZE,
178 .cra_blocksize = SHA256_BLOCK_SIZE,
260 .cra_blocksize = SHA256_BLOCK_SIZE,
340 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/x86/crypto/sha256-mb/
Dsha256_mb.c109 inline uint32_t sha256_pad(uint8_t padblock[SHA256_BLOCK_SIZE * 2], in sha256_pad()
112 uint32_t i = total_len & (SHA256_BLOCK_SIZE - 1); in sha256_pad()
114 memset(&padblock[i], 0, SHA256_BLOCK_SIZE); in sha256_pad()
117 i += ((SHA256_BLOCK_SIZE - 1) & in sha256_pad()
157 copy_len = len & (SHA256_BLOCK_SIZE-1); in sha256_ctx_mgr_resubmit()
170 assert((len % SHA256_BLOCK_SIZE) == 0); in sha256_ctx_mgr_resubmit()
282 if (ctx->partial_block_buffer_length || len < SHA256_BLOCK_SIZE) { in sha256_ctx_mgr_submit()
287 uint32_t copy_len = SHA256_BLOCK_SIZE - in sha256_ctx_mgr_submit()
305 assert(ctx->partial_block_buffer_length <= SHA256_BLOCK_SIZE); in sha256_ctx_mgr_submit()
311 if (ctx->partial_block_buffer_length >= SHA256_BLOCK_SIZE) { in sha256_ctx_mgr_submit()
[all …]
Dsha256_mb_ctx.h129 uint8_t partial_block_buffer[SHA256_BLOCK_SIZE * 2];
/Linux-v4.19/drivers/crypto/
Dpadlock-sha.c164 leftover = ((state.count - 1) & (SHA256_BLOCK_SIZE - 1)) + 1; in padlock_sha256_finup()
165 space = SHA256_BLOCK_SIZE - leftover; in padlock_sha256_finup()
274 .cra_blocksize = SHA256_BLOCK_SIZE,
396 if ((partial + len) >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano()
402 done + SHA256_BLOCK_SIZE); in padlock_sha256_update_nano()
407 done += SHA256_BLOCK_SIZE; in padlock_sha256_update_nano()
412 if (len - done >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano()
503 .cra_blocksize = SHA256_BLOCK_SIZE,
Dsahara.c37 #define SAHARA_MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE
1191 SHA_BUFFER_LEN + SHA256_BLOCK_SIZE); in sahara_sha_cra_init()
1284 .cra_blocksize = SHA256_BLOCK_SIZE,
Domap-sham.c435 d = SHA256_BLOCK_SIZE; in get_block_size()
966 bs = SHA256_BLOCK_SIZE; in omap_sham_init()
1586 .cra_blocksize = SHA256_BLOCK_SIZE,
1631 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/drivers/crypto/qce/
Dsha.h23 #define QCE_SHA_MAX_BLOCKSIZE SHA256_BLOCK_SIZE
Dsha.c453 .blocksize = SHA256_BLOCK_SIZE,
471 .blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/s390/crypto/
Dsha_common.c61 plen = (bsize > SHA256_BLOCK_SIZE) ? 16 : 8; in s390_sha_final()
Dsha256_s390.c74 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/arch/powerpc/crypto/
Dsha256-spe-glue.c234 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/drivers/crypto/stm32/
Dstm32-hash.c121 #define HASH_MAX_KEY_SIZE (SHA256_BLOCK_SIZE * 8)
1311 .cra_blocksize = SHA256_BLOCK_SIZE,
1337 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/drivers/crypto/mediatek/
Dmtk-sha.c396 ctx->bs = SHA256_BLOCK_SIZE; in mtk_sha_init()
984 .cra_blocksize = SHA256_BLOCK_SIZE,
1061 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/crypto/
Dsha256_generic.c239 src += SHA256_BLOCK_SIZE; in sha256_generic_block_fn()
275 .cra_blocksize = SHA256_BLOCK_SIZE,
/Linux-v4.19/drivers/crypto/rockchip/
Drk3288_crypto_ahash.c366 .cra_blocksize = SHA256_BLOCK_SIZE,

12