Lines Matching full:block
23 static size_t chunk_size(const struct bt_mesh_blob_block *block, in chunk_size() argument
26 if ((chunk_idx == block->chunk_count - 1) && in chunk_size()
27 (block->size % CHUNK_SIZE)) { in chunk_size()
28 return block->size % CHUNK_SIZE; in chunk_size()
52 struct bt_mesh_blob_block block = { 0 }; in ZTEST() local
98 block.chunk_count = in ZTEST()
101 block.size = remaining > CONFIG_BT_MESH_BLOB_BLOCK_SIZE_MAX in ZTEST()
106 err = blob_flash_stream.io.block_start(&blob_flash_stream.io, &xfer, &block); in ZTEST()
109 /* `block_start` in write mode will erase flash pages that can fit block. in ZTEST()
110 * Assert that at least block size of data was not erased in read mode in ZTEST()
112 flash_area_read(blob_flash_stream.area, block.offset, ctrl_data, block.size); in ZTEST()
113 zassert_mem_equal(ctrl_data, &test_data[block.offset], block.size, in ZTEST()
118 block.offset = block_idx * (1 << block_size_to_log(block.size)); in ZTEST()
120 for (i = 0; i < block.chunk_count; i++) { in ZTEST()
121 chunk.size = chunk_size(&block, chunk_idx); in ZTEST()
124 err = blob_flash_stream.io.rd(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
126 err, block.offset + chunk.offset, chunk.size); in ZTEST()
143 err = blob_flash_stream.io.rd(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
147 block.offset++; in ZTEST()
148 err = blob_flash_stream.io.rd(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
151 block.offset--; in ZTEST()
153 err = blob_flash_stream.io.rd(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
162 struct bt_mesh_blob_block block = { 0 }; in ZTEST() local
195 block.chunk_count = in ZTEST()
198 block.size = remaining > CONFIG_BT_MESH_BLOB_BLOCK_SIZE_MAX in ZTEST()
201 block.offset = block_idx * (1 << block_size_to_log(block.size)); in ZTEST()
203 err = blob_flash_stream.io.block_start(&blob_flash_stream.io, &xfer, &block); in ZTEST()
206 flash_area_read(blob_flash_stream.area, block.offset, in ZTEST()
207 ctrl_data, block.size); in ZTEST()
209 zassert_mem_equal(ctrl_data, erased_block_data, block.size, in ZTEST()
214 for (i = 0; i < block.chunk_count; i++) { in ZTEST()
215 chunk.size = chunk_size(&block, chunk_idx); in ZTEST()
219 &test_data[chunk.offset + block.offset], in ZTEST()
222 err = blob_flash_stream.io.wr(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
229 ROUND_UP((block.offset + chunk.offset) % in ZTEST()
232 (block.offset + chunk.offset) % in ZTEST()
235 flash_area_read(blob_flash_stream.area, block.offset + chunk.offset, in ZTEST()
264 err = blob_flash_stream.io.wr(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
268 block.offset++; in ZTEST()
269 err = blob_flash_stream.io.wr(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()
272 block.offset--; in ZTEST()
274 err = blob_flash_stream.io.wr(&blob_flash_stream.io, &xfer, &block, &chunk); in ZTEST()