Lines Matching full:stride

41 	unsigned int stride;  in show_root_entry_hash()  local
44 stride = regmap_get_reg_stride(sec->m10bmc->regmap); in show_root_entry_hash()
53 if ((sha_num_bytes % stride) || in show_root_entry_hash()
62 hash, sha_num_bytes / stride); in show_root_entry_hash()
65 reh_addr, sha_num_bytes / stride, ret); in show_root_entry_hash()
93 unsigned int i, stride, size = CSK_32ARRAY_SIZE * sizeof(u32); in show_canceled_csk() local
100 stride = regmap_get_reg_stride(sec->m10bmc->regmap); in show_canceled_csk()
101 if (size % stride) { in show_canceled_csk()
103 "CSK vector size (0x%x) not aligned to stride (0x%x)\n", in show_canceled_csk()
104 size, stride); in show_canceled_csk()
110 size / stride); in show_canceled_csk()
113 addr, size / stride, ret); in show_canceled_csk()
144 unsigned int stride, num_bits; in flash_count_show() local
148 stride = regmap_get_reg_stride(sec->m10bmc->regmap); in flash_count_show()
151 if (FLASH_COUNT_SIZE % stride) { in flash_count_show()
153 "FLASH_COUNT_SIZE (0x%x) not aligned to stride (0x%x)\n", in flash_count_show()
154 FLASH_COUNT_SIZE, stride); in flash_count_show()
164 flash_buf, FLASH_COUNT_SIZE / stride); in flash_count_show()
168 STAGING_FLASH_COUNT, FLASH_COUNT_SIZE / stride, ret); in flash_count_show()
449 unsigned int stride, extra = 0; in m10bmc_sec_write() local
452 stride = regmap_get_reg_stride(sec->m10bmc->regmap); in m10bmc_sec_write()
464 WARN_ON_ONCE(WRITE_BLOCK_SIZE % stride); in m10bmc_sec_write()
469 blk_size / stride); in m10bmc_sec_write()
474 * If blk_size is not aligned to stride, then handle the extra in m10bmc_sec_write()
477 if (blk_size % stride) { in m10bmc_sec_write()
478 extra_offset = offset + ALIGN_DOWN(blk_size, stride); in m10bmc_sec_write()
479 memcpy(&extra, (u8 *)(data + extra_offset), blk_size % stride); in m10bmc_sec_write()