/Linux-v5.4/include/crypto/ |
D | cbc.h | 19 unsigned int bsize = crypto_skcipher_blocksize(tfm); in crypto_cbc_encrypt_segment() local 26 crypto_xor(iv, src, bsize); in crypto_cbc_encrypt_segment() 28 memcpy(iv, dst, bsize); in crypto_cbc_encrypt_segment() 30 src += bsize; in crypto_cbc_encrypt_segment() 31 dst += bsize; in crypto_cbc_encrypt_segment() 32 } while ((nbytes -= bsize) >= bsize); in crypto_cbc_encrypt_segment() 41 unsigned int bsize = crypto_skcipher_blocksize(tfm); in crypto_cbc_encrypt_inplace() local 47 crypto_xor(src, iv, bsize); in crypto_cbc_encrypt_inplace() 51 src += bsize; in crypto_cbc_encrypt_inplace() 52 } while ((nbytes -= bsize) >= bsize); in crypto_cbc_encrypt_inplace() [all …]
|
D | ctr.h | 48 int bsize = min(nbytes, blocksize); in crypto_ctr_encrypt_walk() local 52 crypto_xor_cpy(dst, src, buf, bsize); in crypto_ctr_encrypt_walk() 55 dst += bsize; in crypto_ctr_encrypt_walk() 56 src += bsize; in crypto_ctr_encrypt_walk() 57 nbytes -= bsize; in crypto_ctr_encrypt_walk()
|
/Linux-v5.4/crypto/ |
D | pcbc.c | 23 int bsize = crypto_cipher_blocksize(tfm); in crypto_pcbc_encrypt_segment() local 30 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_segment() 32 crypto_xor_cpy(iv, dst, src, bsize); in crypto_pcbc_encrypt_segment() 34 src += bsize; in crypto_pcbc_encrypt_segment() 35 dst += bsize; in crypto_pcbc_encrypt_segment() 36 } while ((nbytes -= bsize) >= bsize); in crypto_pcbc_encrypt_segment() 45 int bsize = crypto_cipher_blocksize(tfm); in crypto_pcbc_encrypt_inplace() local 52 memcpy(tmpbuf, src, bsize); in crypto_pcbc_encrypt_inplace() 53 crypto_xor(iv, src, bsize); in crypto_pcbc_encrypt_inplace() 55 crypto_xor_cpy(iv, tmpbuf, src, bsize); in crypto_pcbc_encrypt_inplace() [all …]
|
D | cfb.c | 60 const unsigned int bsize = crypto_cfb_bsize(tfm); in crypto_cfb_encrypt_segment() local 68 crypto_xor(dst, src, bsize); in crypto_cfb_encrypt_segment() 71 src += bsize; in crypto_cfb_encrypt_segment() 72 dst += bsize; in crypto_cfb_encrypt_segment() 73 } while ((nbytes -= bsize) >= bsize); in crypto_cfb_encrypt_segment() 75 memcpy(walk->iv, iv, bsize); in crypto_cfb_encrypt_segment() 83 const unsigned int bsize = crypto_cfb_bsize(tfm); in crypto_cfb_encrypt_inplace() local 91 crypto_xor(src, tmp, bsize); in crypto_cfb_encrypt_inplace() 94 src += bsize; in crypto_cfb_encrypt_inplace() 95 } while ((nbytes -= bsize) >= bsize); in crypto_cfb_encrypt_inplace() [all …]
|
D | cts.c | 107 int bsize = crypto_skcipher_blocksize(tfm); in cts_cbc_encrypt() local 116 sg = scatterwalk_ffwd(rctx->sg, req->dst, offset - bsize); in cts_cbc_encrypt() 117 scatterwalk_map_and_copy(d + bsize, sg, 0, bsize, 0); in cts_cbc_encrypt() 119 memset(d, 0, bsize); in cts_cbc_encrypt() 122 scatterwalk_map_and_copy(d, sg, 0, bsize + lastn, 1); in cts_cbc_encrypt() 128 skcipher_request_set_crypt(subreq, sg, sg, bsize, req->iv); in cts_cbc_encrypt() 153 int bsize = crypto_skcipher_blocksize(tfm); in crypto_cts_encrypt() local 159 if (nbytes < bsize) in crypto_cts_encrypt() 162 if (nbytes == bsize) { in crypto_cts_encrypt() 171 offset = rounddown(nbytes - 1, bsize); in crypto_cts_encrypt() [all …]
|
D | ctr.c | 30 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_final() local 42 crypto_inc(ctrblk, bsize); in crypto_ctr_crypt_final() 50 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_segment() local 59 crypto_xor(dst, src, bsize); in crypto_ctr_crypt_segment() 62 crypto_inc(ctrblk, bsize); in crypto_ctr_crypt_segment() 64 src += bsize; in crypto_ctr_crypt_segment() 65 dst += bsize; in crypto_ctr_crypt_segment() 66 } while ((nbytes -= bsize) >= bsize); in crypto_ctr_crypt_segment() 76 unsigned int bsize = crypto_cipher_blocksize(tfm); in crypto_ctr_crypt_inplace() local 87 crypto_xor(src, keystream, bsize); in crypto_ctr_crypt_inplace() [all …]
|
D | ofb.c | 21 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ofb_crypt() local 27 while (walk.nbytes >= bsize) { in crypto_ofb_crypt() 35 crypto_xor_cpy(dst, src, iv, bsize); in crypto_ofb_crypt() 36 dst += bsize; in crypto_ofb_crypt() 37 src += bsize; in crypto_ofb_crypt() 38 } while ((nbytes -= bsize) >= bsize); in crypto_ofb_crypt()
|
D | ecb.c | 19 const unsigned int bsize = crypto_cipher_blocksize(cipher); in crypto_ecb_crypt() local 33 src += bsize; in crypto_ecb_crypt() 34 dst += bsize; in crypto_ecb_crypt() 35 } while ((nbytes -= bsize) >= bsize); in crypto_ecb_crypt()
|
D | ablkcipher.c | 139 unsigned int bsize, in ablkcipher_next_slow() argument 143 unsigned aligned_bsize = ALIGN(bsize, alignmask + 1); in ablkcipher_next_slow() 159 src = dst = ablkcipher_get_spot(dst, bsize); in ablkcipher_next_slow() 161 p->len = bsize; in ablkcipher_next_slow() 164 scatterwalk_copychunks(src, &walk->in, bsize, 0); in ablkcipher_next_slow() 168 walk->nbytes = bsize; in ablkcipher_next_slow() 217 unsigned int alignmask, bsize, n; in ablkcipher_walk_next() local 231 bsize = min(walk->blocksize, n); in ablkcipher_walk_next() 235 if (n < bsize || in ablkcipher_walk_next() 238 err = ablkcipher_next_slow(req, walk, bsize, alignmask, in ablkcipher_walk_next()
|
D | blkcipher.c | 69 unsigned int bsize) in blkcipher_done_slow() argument 74 addr = blkcipher_get_spot(addr, bsize); in blkcipher_done_slow() 75 scatterwalk_copychunks(addr, &walk->out, bsize, 1); in blkcipher_done_slow() 141 unsigned int bsize, in blkcipher_next_slow() argument 145 unsigned aligned_bsize = ALIGN(bsize, alignmask + 1); in blkcipher_next_slow() 163 walk->dst.virt.addr = blkcipher_get_spot(walk->dst.virt.addr, bsize); in blkcipher_next_slow() 165 aligned_bsize, bsize); in blkcipher_next_slow() 167 scatterwalk_copychunks(walk->src.virt.addr, &walk->in, bsize, 0); in blkcipher_next_slow() 169 walk->nbytes = bsize; in blkcipher_next_slow() 219 unsigned int bsize; in blkcipher_walk_next() local [all …]
|
/Linux-v5.4/arch/x86/crypto/ |
D | cast5_avx_glue.c | 53 const unsigned int bsize = CAST5_BLOCK_SIZE; in ecb_crypt() local 67 if (nbytes >= bsize * CAST5_PARALLEL_BLOCKS) { in ecb_crypt() 72 wsrc += bsize * CAST5_PARALLEL_BLOCKS; in ecb_crypt() 73 wdst += bsize * CAST5_PARALLEL_BLOCKS; in ecb_crypt() 74 nbytes -= bsize * CAST5_PARALLEL_BLOCKS; in ecb_crypt() 75 } while (nbytes >= bsize * CAST5_PARALLEL_BLOCKS); in ecb_crypt() 77 if (nbytes < bsize) in ecb_crypt() 87 wsrc += bsize; in ecb_crypt() 88 wdst += bsize; in ecb_crypt() 89 nbytes -= bsize; in ecb_crypt() [all …]
|
D | blowfish_glue.c | 75 unsigned int bsize = BF_BLOCK_SIZE; in ecb_crypt() local 89 if (nbytes >= bsize * 4) { in ecb_crypt() 93 wsrc += bsize * 4; in ecb_crypt() 94 wdst += bsize * 4; in ecb_crypt() 95 nbytes -= bsize * 4; in ecb_crypt() 96 } while (nbytes >= bsize * 4); in ecb_crypt() 98 if (nbytes < bsize) in ecb_crypt() 106 wsrc += bsize; in ecb_crypt() 107 wdst += bsize; in ecb_crypt() 108 nbytes -= bsize; in ecb_crypt() [all …]
|
D | des3_ede_glue.c | 78 const unsigned int bsize = DES3_EDE_BLOCK_SIZE; in ecb_crypt() local 90 if (nbytes >= bsize * 3) { in ecb_crypt() 95 wsrc += bsize * 3; in ecb_crypt() 96 wdst += bsize * 3; in ecb_crypt() 97 nbytes -= bsize * 3; in ecb_crypt() 98 } while (nbytes >= bsize * 3); in ecb_crypt() 100 if (nbytes < bsize) in ecb_crypt() 108 wsrc += bsize; in ecb_crypt() 109 wdst += bsize; in ecb_crypt() 110 nbytes -= bsize; in ecb_crypt() [all …]
|
D | glue_helper.c | 25 const unsigned int bsize = 128 / 8; in glue_ecb_req_128bit() local 39 fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, in glue_ecb_req_128bit() 42 func_bytes = bsize * gctx->funcs[i].num_blocks; in glue_ecb_req_128bit() 55 if (nbytes < bsize) in glue_ecb_req_128bit() 70 const unsigned int bsize = 128 / 8; in glue_cbc_encrypt_req_128bit() local 88 nbytes -= bsize; in glue_cbc_encrypt_req_128bit() 89 } while (nbytes >= bsize); in glue_cbc_encrypt_req_128bit() 102 const unsigned int bsize = 128 / 8; in glue_cbc_decrypt_req_128bit() local 117 fpu_enabled = glue_fpu_begin(bsize, gctx->fpu_blocks_limit, in glue_cbc_decrypt_req_128bit() 120 src += nbytes / bsize - 1; in glue_cbc_decrypt_req_128bit() [all …]
|
/Linux-v5.4/arch/s390/crypto/ |
D | sha_common.c | 19 unsigned int bsize = crypto_shash_blocksize(desc->tfm); in s390_sha_update() local 23 index = ctx->count % bsize; in s390_sha_update() 26 if ((index + len) < bsize) in s390_sha_update() 31 memcpy(ctx->buf + index, data, bsize - index); in s390_sha_update() 32 cpacf_kimd(ctx->func, ctx->state, ctx->buf, bsize); in s390_sha_update() 33 data += bsize - index; in s390_sha_update() 34 len -= bsize - index; in s390_sha_update() 39 if (len >= bsize) { in s390_sha_update() 40 n = (len / bsize) * bsize; in s390_sha_update() 75 unsigned int bsize = crypto_shash_blocksize(desc->tfm); in s390_sha_final() local [all …]
|
/Linux-v5.4/arch/m68k/emu/ |
D | nfblock.c | 56 u32 blocks, bsize; member 99 static int __init nfhd_init_one(int id, u32 blocks, u32 bsize) in nfhd_init_one() argument 105 blocks, bsize); in nfhd_init_one() 107 if (bsize < 512 || (bsize & (bsize - 1))) { in nfhd_init_one() 118 dev->bsize = bsize; in nfhd_init_one() 119 dev->bshift = ffs(bsize) - 10; in nfhd_init_one() 127 blk_queue_logical_block_size(dev->queue, bsize); in nfhd_init_one() 138 set_capacity(dev->disk, (sector_t)blocks * (bsize / 512)); in nfhd_init_one() 157 u32 blocks, bsize; in nfhd_init() local 175 if (nfhd_get_capacity(i, 0, &blocks, &bsize)) in nfhd_init() [all …]
|
/Linux-v5.4/lib/mpi/ |
D | mpi-pow.c | 31 mpi_size_t esize, msize, bsize, rsize; in mpi_powm() local 79 bsize = base->nlimbs; in mpi_powm() 81 if (bsize > msize) { /* The base is larger than the module. Reduce it. */ in mpi_powm() 84 bp = bp_marker = mpi_alloc_limb_space(bsize + 1); in mpi_powm() 87 MPN_COPY(bp, base->d, bsize); in mpi_powm() 90 mpihelp_divrem(bp + msize, 0, bp, bsize, mp, msize); in mpi_powm() 91 bsize = msize; in mpi_powm() 94 MPN_NORMALIZE(bp, bsize); in mpi_powm() 98 if (!bsize) { in mpi_powm() 122 bp = bp_marker = mpi_alloc_limb_space(bsize); in mpi_powm() [all …]
|
/Linux-v5.4/fs/squashfs/ |
D | file_direct.c | 21 static int squashfs_read_cache(struct page *target_page, u64 block, int bsize, 25 int squashfs_readpage_block(struct page *target_page, u64 block, int bsize, in squashfs_readpage_block() argument 84 res = squashfs_read_cache(target_page, block, bsize, pages, in squashfs_readpage_block() 93 res = squashfs_read_data(inode->i_sb, block, bsize, NULL, actor); in squashfs_readpage_block() 144 static int squashfs_read_cache(struct page *target_page, u64 block, int bsize, in squashfs_read_cache() argument 149 block, bsize); in squashfs_read_cache() 154 bsize); in squashfs_read_cache()
|
D | file_cache.c | 21 int squashfs_readpage_block(struct page *page, u64 block, int bsize, int expected) in squashfs_readpage_block() argument 25 block, bsize); in squashfs_readpage_block() 30 bsize); in squashfs_readpage_block()
|
/Linux-v5.4/fs/freevxfs/ |
D | vxfs_lookup.c | 83 u_long bsize = ip->i_sb->s_blocksize; in vxfs_find_entry() local 104 if ((pos & (bsize - 1)) < 4) { in vxfs_find_entry() 116 pos += bsize - 1; in vxfs_find_entry() 117 pos &= ~(bsize - 1); in vxfs_find_entry() 219 u_long bsize = sbp->s_blocksize; in vxfs_readdir() local 255 if ((pos & (bsize - 1)) < 4) { in vxfs_readdir() 267 pos += bsize - 1; in vxfs_readdir() 268 pos &= ~(bsize - 1); in vxfs_readdir()
|
D | vxfs_olt.c | 57 vxfs_oblock(struct super_block *sbp, daddr_t block, u_long bsize) in vxfs_oblock() argument 59 BUG_ON(sbp->s_blocksize % bsize); in vxfs_oblock() 60 return (block * (sbp->s_blocksize / bsize)); in vxfs_oblock() 77 vxfs_read_olt(struct super_block *sbp, u_long bsize) in vxfs_read_olt() argument 84 bp = sb_bread(sbp, vxfs_oblock(sbp, infp->vsi_oltext, bsize)); in vxfs_read_olt()
|
/Linux-v5.4/fs/affs/ |
D | file.c | 510 u32 bidx, boff, bsize; in affs_do_readpage_ofs() local 516 bsize = AFFS_SB(sb)->s_data_blksize; in affs_do_readpage_ofs() 518 bidx = tmp / bsize; in affs_do_readpage_ofs() 519 boff = tmp % bsize; in affs_do_readpage_ofs() 525 tmp = min(bsize - boff, to - pos); in affs_do_readpage_ofs() 526 BUG_ON(pos + tmp > to || tmp > bsize); in affs_do_readpage_ofs() 545 u32 size, bsize; in affs_extent_file_ofs() local 549 bsize = AFFS_SB(sb)->s_data_blksize; in affs_extent_file_ofs() 552 bidx = size / bsize; in affs_extent_file_ofs() 553 boff = size % bsize; in affs_extent_file_ofs() [all …]
|
/Linux-v5.4/tools/perf/util/ |
D | mem2node.c | 38 phys_entry__init(struct phys_entry *entry, u64 start, u64 bsize, u64 node) in phys_entry__init() argument 41 entry->end = start + bsize; in phys_entry__init() 50 u64 bsize = env->memory_bsize; in mem2node__init() local 76 start = bit * bsize; in mem2node__init() 87 prev->end += bsize; in mem2node__init() 92 phys_entry__init(&entries[j++], start, bsize, n->node); in mem2node__init()
|
/Linux-v5.4/fs/jfs/ |
D | jfs_mount.c | 298 s32 bsize; in chkSuper() local 314 bsize = le32_to_cpu(j_sb->s_bsize); in chkSuper() 316 if (bsize != PSIZE) { in chkSuper() 331 AIM_bytesize = lengthPXD(&(j_sb->s_aim2)) * bsize; in chkSuper() 333 AIT_bytesize = lengthPXD(&(j_sb->s_ait2)) * bsize; in chkSuper() 334 AIM_byte_addr = addressPXD(&(j_sb->s_aim2)) * bsize; in chkSuper() 335 AIT_byte_addr = addressPXD(&(j_sb->s_ait2)) * bsize; in chkSuper() 337 fsckwsp_addr = addressPXD(&(j_sb->s_fsckpxd)) * bsize; in chkSuper() 365 sbi->bsize = bsize; in chkSuper()
|
/Linux-v5.4/fs/iomap/ |
D | seek.c | 22 unsigned int bsize = i_blocksize(inode), off; in page_seek_hole_data() local 42 if (bsize == PAGE_SIZE || !ops->is_partially_uptodate) in page_seek_hole_data() 49 for (off = 0; off < PAGE_SIZE; off += bsize) { in page_seek_hole_data() 50 if (offset_in_page(*lastoff) >= off + bsize) in page_seek_hole_data() 52 if (ops->is_partially_uptodate(page, off, bsize) == seek_data) { in page_seek_hole_data() 56 *lastoff = poff + off + bsize; in page_seek_hole_data()
|