/Linux-v4.19/include/crypto/ |
D | sha256_base.h | 61 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_update() 65 if (unlikely((partial + len) >= SHA256_BLOCK_SIZE)) { in sha256_base_do_update() 69 int p = SHA256_BLOCK_SIZE - partial; in sha256_base_do_update() 78 blocks = len / SHA256_BLOCK_SIZE; in sha256_base_do_update() 79 len %= SHA256_BLOCK_SIZE; in sha256_base_do_update() 83 data += blocks * SHA256_BLOCK_SIZE; in sha256_base_do_update() 96 const int bit_offset = SHA256_BLOCK_SIZE - sizeof(__be64); in sha256_base_do_finalize() 99 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_base_do_finalize() 103 memset(sctx->buf + partial, 0x0, SHA256_BLOCK_SIZE - partial); in sha256_base_do_finalize()
|
D | sha.h | 18 #define SHA256_BLOCK_SIZE 64 macro 87 u8 buf[SHA256_BLOCK_SIZE];
|
/Linux-v4.19/arch/sparc/crypto/ |
D | sha256_glue.c | 68 done = SHA256_BLOCK_SIZE - partial; in __sha256_sparc64_update() 72 if (len - done >= SHA256_BLOCK_SIZE) { in __sha256_sparc64_update() 73 const unsigned int rounds = (len - done) / SHA256_BLOCK_SIZE; in __sha256_sparc64_update() 76 done += rounds * SHA256_BLOCK_SIZE; in __sha256_sparc64_update() 86 unsigned int partial = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_update() 89 if (partial + len < SHA256_BLOCK_SIZE) { in sha256_sparc64_update() 104 static const u8 padding[SHA256_BLOCK_SIZE] = { 0x80, }; in sha256_sparc64_final() 109 index = sctx->count % SHA256_BLOCK_SIZE; in sha256_sparc64_final() 110 padlen = (index < 56) ? (56 - index) : ((SHA256_BLOCK_SIZE+56) - index); in sha256_sparc64_final() 172 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha256.c | 111 partial = sctx->count % SHA256_BLOCK_SIZE; in __octeon_sha256_update() 116 if ((partial + len) >= SHA256_BLOCK_SIZE) { in __octeon_sha256_update() 120 done + SHA256_BLOCK_SIZE); in __octeon_sha256_update() 126 done += SHA256_BLOCK_SIZE; in __octeon_sha256_update() 128 } while (done + SHA256_BLOCK_SIZE <= len); in __octeon_sha256_update() 147 if ((sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in octeon_sha256_update() 242 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/nx/ |
D | nx-sha256.c | 80 u64 buf_len = (sctx->count % SHA256_BLOCK_SIZE); in nx_sha256_update() 88 total = (sctx->count % SHA256_BLOCK_SIZE) + len; in nx_sha256_update() 89 if (total < SHA256_BLOCK_SIZE) { in nx_sha256_update() 141 to_process = to_process & ~(SHA256_BLOCK_SIZE - 1); in nx_sha256_update() 176 } while (leftover >= SHA256_BLOCK_SIZE); in nx_sha256_update() 209 if (sctx->count >= SHA256_BLOCK_SIZE) { in nx_sha256_final() 222 len = sctx->count & (SHA256_BLOCK_SIZE - 1); in nx_sha256_final() 226 if (len != (sctx->count & (SHA256_BLOCK_SIZE - 1))) { in nx_sha256_final() 291 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/arm64/crypto/ |
D | sha256-glue.c | 71 .base.cra_blocksize = SHA256_BLOCK_SIZE, 105 chunk + sctx->count % SHA256_BLOCK_SIZE > SHA256_BLOCK_SIZE) in sha256_update_neon() 106 chunk = SHA256_BLOCK_SIZE - in sha256_update_neon() 107 sctx->count % SHA256_BLOCK_SIZE; in sha256_update_neon() 154 .base.cra_blocksize = SHA256_BLOCK_SIZE,
|
D | sha2-ce-glue.c | 62 bool finalize = !sctx->sst.count && !(len % SHA256_BLOCK_SIZE); in sha256_ce_finup() 117 .cra_blocksize = SHA256_BLOCK_SIZE, 131 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/arm/crypto/ |
D | sha2-ce-glue.c | 38 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha2_ce_update() 81 .cra_blocksize = SHA256_BLOCK_SIZE, 95 .cra_blocksize = SHA256_BLOCK_SIZE,
|
D | sha256_neon_glue.c | 38 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha256_update() 82 .cra_blocksize = SHA256_BLOCK_SIZE,
|
D | sha256_glue.c | 74 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/x86/crypto/ |
D | sha256_ssse3_glue.c | 53 (sctx->count % SHA256_BLOCK_SIZE) + len < SHA256_BLOCK_SIZE) in sha256_update() 112 .cra_blocksize = SHA256_BLOCK_SIZE, 178 .cra_blocksize = SHA256_BLOCK_SIZE, 260 .cra_blocksize = SHA256_BLOCK_SIZE, 340 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/x86/crypto/sha256-mb/ |
D | sha256_mb.c | 109 inline uint32_t sha256_pad(uint8_t padblock[SHA256_BLOCK_SIZE * 2], in sha256_pad() 112 uint32_t i = total_len & (SHA256_BLOCK_SIZE - 1); in sha256_pad() 114 memset(&padblock[i], 0, SHA256_BLOCK_SIZE); in sha256_pad() 117 i += ((SHA256_BLOCK_SIZE - 1) & in sha256_pad() 157 copy_len = len & (SHA256_BLOCK_SIZE-1); in sha256_ctx_mgr_resubmit() 170 assert((len % SHA256_BLOCK_SIZE) == 0); in sha256_ctx_mgr_resubmit() 282 if (ctx->partial_block_buffer_length || len < SHA256_BLOCK_SIZE) { in sha256_ctx_mgr_submit() 287 uint32_t copy_len = SHA256_BLOCK_SIZE - in sha256_ctx_mgr_submit() 305 assert(ctx->partial_block_buffer_length <= SHA256_BLOCK_SIZE); in sha256_ctx_mgr_submit() 311 if (ctx->partial_block_buffer_length >= SHA256_BLOCK_SIZE) { in sha256_ctx_mgr_submit() [all …]
|
D | sha256_mb_ctx.h | 129 uint8_t partial_block_buffer[SHA256_BLOCK_SIZE * 2];
|
/Linux-v4.19/drivers/crypto/ |
D | padlock-sha.c | 164 leftover = ((state.count - 1) & (SHA256_BLOCK_SIZE - 1)) + 1; in padlock_sha256_finup() 165 space = SHA256_BLOCK_SIZE - leftover; in padlock_sha256_finup() 274 .cra_blocksize = SHA256_BLOCK_SIZE, 396 if ((partial + len) >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano() 402 done + SHA256_BLOCK_SIZE); in padlock_sha256_update_nano() 407 done += SHA256_BLOCK_SIZE; in padlock_sha256_update_nano() 412 if (len - done >= SHA256_BLOCK_SIZE) { in padlock_sha256_update_nano() 503 .cra_blocksize = SHA256_BLOCK_SIZE,
|
D | sahara.c | 37 #define SAHARA_MAX_SHA_BLOCK_SIZE SHA256_BLOCK_SIZE 1191 SHA_BUFFER_LEN + SHA256_BLOCK_SIZE); in sahara_sha_cra_init() 1284 .cra_blocksize = SHA256_BLOCK_SIZE,
|
D | omap-sham.c | 435 d = SHA256_BLOCK_SIZE; in get_block_size() 966 bs = SHA256_BLOCK_SIZE; in omap_sham_init() 1586 .cra_blocksize = SHA256_BLOCK_SIZE, 1631 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/qce/ |
D | sha.h | 23 #define QCE_SHA_MAX_BLOCKSIZE SHA256_BLOCK_SIZE
|
D | sha.c | 453 .blocksize = SHA256_BLOCK_SIZE, 471 .blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/s390/crypto/ |
D | sha_common.c | 61 plen = (bsize > SHA256_BLOCK_SIZE) ? 16 : 8; in s390_sha_final()
|
D | sha256_s390.c | 74 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/arch/powerpc/crypto/ |
D | sha256-spe-glue.c | 234 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/stm32/ |
D | stm32-hash.c | 121 #define HASH_MAX_KEY_SIZE (SHA256_BLOCK_SIZE * 8) 1311 .cra_blocksize = SHA256_BLOCK_SIZE, 1337 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/mediatek/ |
D | mtk-sha.c | 396 ctx->bs = SHA256_BLOCK_SIZE; in mtk_sha_init() 984 .cra_blocksize = SHA256_BLOCK_SIZE, 1061 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/crypto/ |
D | sha256_generic.c | 239 src += SHA256_BLOCK_SIZE; in sha256_generic_block_fn() 275 .cra_blocksize = SHA256_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/rockchip/ |
D | rk3288_crypto_ahash.c | 366 .cra_blocksize = SHA256_BLOCK_SIZE,
|