/Linux-v4.19/include/crypto/ |
D | sha512_base.h | 61 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_update() 67 if (unlikely((partial + len) >= SHA512_BLOCK_SIZE)) { in sha512_base_do_update() 71 int p = SHA512_BLOCK_SIZE - partial; in sha512_base_do_update() 80 blocks = len / SHA512_BLOCK_SIZE; in sha512_base_do_update() 81 len %= SHA512_BLOCK_SIZE; in sha512_base_do_update() 85 data += blocks * SHA512_BLOCK_SIZE; in sha512_base_do_update() 98 const int bit_offset = SHA512_BLOCK_SIZE - sizeof(__be64[2]); in sha512_base_do_finalize() 101 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_base_do_finalize() 105 memset(sctx->buf + partial, 0x0, SHA512_BLOCK_SIZE - partial); in sha512_base_do_finalize()
|
D | sha.h | 24 #define SHA512_BLOCK_SIZE 128 macro 93 u8 buf[SHA512_BLOCK_SIZE];
|
/Linux-v4.19/arch/sparc/crypto/ |
D | sha512_glue.c | 68 done = SHA512_BLOCK_SIZE - partial; in __sha512_sparc64_update() 72 if (len - done >= SHA512_BLOCK_SIZE) { in __sha512_sparc64_update() 73 const unsigned int rounds = (len - done) / SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 76 done += rounds * SHA512_BLOCK_SIZE; in __sha512_sparc64_update() 86 unsigned int partial = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_update() 89 if (partial + len < SHA512_BLOCK_SIZE) { in sha512_sparc64_update() 105 static const u8 padding[SHA512_BLOCK_SIZE] = { 0x80, }; in sha512_sparc64_final() 112 index = sctx->count[0] % SHA512_BLOCK_SIZE; in sha512_sparc64_final() 113 padlen = (index < 112) ? (112 - index) : ((SHA512_BLOCK_SIZE+112) - index); in sha512_sparc64_final() 157 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/nx/ |
D | nx-sha512.c | 80 u64 buf_len = (sctx->count[0] % SHA512_BLOCK_SIZE); in nx_sha512_update() 88 total = (sctx->count[0] % SHA512_BLOCK_SIZE) + len; in nx_sha512_update() 89 if (total < SHA512_BLOCK_SIZE) { in nx_sha512_update() 140 to_process = to_process & ~(SHA512_BLOCK_SIZE - 1); in nx_sha512_update() 180 } while (leftover >= SHA512_BLOCK_SIZE); in nx_sha512_update() 213 if (sctx->count[0] >= SHA512_BLOCK_SIZE) { in nx_sha512_final() 231 len = sctx->count[0] & (SHA512_BLOCK_SIZE - 1); in nx_sha512_final() 235 if (len != (sctx->count[0] & (SHA512_BLOCK_SIZE - 1))) { in nx_sha512_final() 297 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/arch/mips/cavium-octeon/crypto/ |
D | octeon-sha512.c | 123 index = sctx->count[0] % SHA512_BLOCK_SIZE; in __octeon_sha512_update() 129 part_len = SHA512_BLOCK_SIZE - index; in __octeon_sha512_update() 136 for (i = part_len; i + SHA512_BLOCK_SIZE <= len; in __octeon_sha512_update() 137 i += SHA512_BLOCK_SIZE) in __octeon_sha512_update() 161 if ((sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in octeon_sha512_update() 238 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/arch/arm/crypto/ |
D | sha512-neon-glue.c | 34 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_neon_update() 93 .cra_blocksize = SHA512_BLOCK_SIZE,
|
D | sha512-glue.c | 66 .cra_blocksize = SHA512_BLOCK_SIZE, 80 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/arch/x86/crypto/ |
D | sha512_ssse3_glue.c | 53 (sctx->count[0] % SHA512_BLOCK_SIZE) + len < SHA512_BLOCK_SIZE) in sha512_update() 112 .cra_blocksize = SHA512_BLOCK_SIZE, 189 .cra_blocksize = SHA512_BLOCK_SIZE, 260 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/arch/x86/crypto/sha512-mb/ |
D | sha512_mb.c | 110 inline uint32_t sha512_pad(uint8_t padblock[SHA512_BLOCK_SIZE * 2], in sha512_pad() 113 uint32_t i = total_len & (SHA512_BLOCK_SIZE - 1); in sha512_pad() 115 memset(&padblock[i], 0, SHA512_BLOCK_SIZE); in sha512_pad() 118 i += ((SHA512_BLOCK_SIZE - 1) & in sha512_pad() 157 copy_len = len & (SHA512_BLOCK_SIZE-1); in sha512_ctx_mgr_resubmit() 170 assert((len % SHA512_BLOCK_SIZE) == 0); in sha512_ctx_mgr_resubmit() 302 if (ctx->partial_block_buffer_length || len < SHA512_BLOCK_SIZE) { in sha512_ctx_mgr_submit() 306 uint32_t copy_len = SHA512_BLOCK_SIZE - in sha512_ctx_mgr_submit() 326 assert(ctx->partial_block_buffer_length <= SHA512_BLOCK_SIZE); in sha512_ctx_mgr_submit() 331 if (ctx->partial_block_buffer_length >= SHA512_BLOCK_SIZE) { in sha512_ctx_mgr_submit() [all …]
|
D | sha512_mb_ctx.h | 123 uint8_t partial_block_buffer[SHA512_BLOCK_SIZE * 2];
|
/Linux-v4.19/arch/arm64/crypto/ |
D | sha512-ce-glue.c | 90 .base.cra_blocksize = SHA512_BLOCK_SIZE, 102 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
D | sha512-glue.c | 66 .base.cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/arch/s390/crypto/ |
D | sha.h | 18 #define SHA_MAX_BLOCK_SIZE SHA512_BLOCK_SIZE
|
D | sha512_s390.c | 79 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/ccree/ |
D | cc_hash.h | 19 #define CC_MAX_HASH_BLCK_SIZE SHA512_BLOCK_SIZE
|
/Linux-v4.19/crypto/ |
D | sha512_generic.c | 161 src += SHA512_BLOCK_SIZE; in sha512_generic_block_fn() 197 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/ccp/ |
D | ccp-crypto.h | 185 #define MAX_SHA_BLOCK_SIZE SHA512_BLOCK_SIZE
|
D | ccp-crypto-sha.c | 417 .block_size = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/mediatek/ |
D | mtk-sha.c | 103 u8 ipad[SHA512_BLOCK_SIZE] __aligned(sizeof(u32)); 104 u8 opad[SHA512_BLOCK_SIZE] __aligned(sizeof(u32)); 404 ctx->bs = SHA512_BLOCK_SIZE; in mtk_sha_init() 1111 .cra_blocksize = SHA512_BLOCK_SIZE, 1162 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/axis/ |
D | artpec6_crypto.c | 294 char partial_buffer[SHA512_BLOCK_SIZE]; 295 char partial_buffer_out[SHA512_BLOCK_SIZE]; 296 char key_buffer[SHA512_BLOCK_SIZE]; 297 char pad_buffer[SHA512_BLOCK_SIZE + 32]; 308 char partial_buffer[SHA512_BLOCK_SIZE]; 317 char hmac_key[SHA512_BLOCK_SIZE]; 2828 .cra_blocksize = SHA512_BLOCK_SIZE, 2852 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/ |
D | omap-sham.c | 139 #define BUFLEN SHA512_BLOCK_SIZE 166 u8 ipad[SHA512_BLOCK_SIZE] OMAP_ALIGNED; 167 u8 opad[SHA512_BLOCK_SIZE] OMAP_ALIGNED; 439 d = SHA512_BLOCK_SIZE; in get_block_size() 974 bs = SHA512_BLOCK_SIZE; in omap_sham_init() 1677 .cra_blocksize = SHA512_BLOCK_SIZE, 1722 .cra_blocksize = SHA512_BLOCK_SIZE,
|
D | atmel-sha.c | 109 u8 buffer[SHA_BUFFER_LEN + SHA512_BLOCK_SIZE] __aligned(sizeof(u32)); 461 ctx->block_size = SHA512_BLOCK_SIZE; in atmel_sha_init() 1371 .cra_blocksize = SHA512_BLOCK_SIZE, 1682 u8 buffer[SHA512_BLOCK_SIZE]; 1734 u32 ipad[SHA512_BLOCK_SIZE / sizeof(u32)]; 1735 u32 opad[SHA512_BLOCK_SIZE / sizeof(u32)]; 1787 ctx->block_size = SHA512_BLOCK_SIZE; in atmel_sha_hmac_setup() 2207 .cra_blocksize = SHA512_BLOCK_SIZE,
|
/Linux-v4.19/drivers/crypto/bcm/ |
D | cipher.h | 60 #define MAX_HASH_BLOCK_SIZE SHA512_BLOCK_SIZE
|
/Linux-v4.19/drivers/crypto/inside-secure/ |
D | safexcel_hash.c | 44 u8 cache[SHA512_BLOCK_SIZE] __aligned(sizeof(u32)); 48 u8 cache_next[SHA512_BLOCK_SIZE] __aligned(sizeof(u32)); 1350 .cra_blocksize = SHA512_BLOCK_SIZE, 1478 .cra_blocksize = SHA512_BLOCK_SIZE,
|
D | safexcel.h | 632 u8 cache[SHA512_BLOCK_SIZE];
|