/Zephyr-latest/tests/unit/winstream/ |
D | main.c | 29 uint32_t bytes = sys_winstream_read(ws, &seq, &c, 1); in ZTEST() local 31 zassert_true(bytes == 1, ""); in ZTEST() 36 bytes = sys_winstream_read(ws, &seq, &c, 1); in ZTEST() 37 zassert_true(bytes == 0, ""); in ZTEST() 52 bytes = sys_winstream_read(ws, &seq, readback, sizeof(readback)); in ZTEST() 54 zassert_true(bytes == 0, ""); in ZTEST() 59 bytes = sys_winstream_read(ws, &seq, readback, sizeof(readback)); in ZTEST() 61 zassert_true(bytes == 0, ""); in ZTEST() 68 bytes = sys_winstream_read(ws, &seq, readback, sizeof(readback)); in ZTEST() 69 zassert_true(bytes == ws->len / 2, ""); in ZTEST() [all …]
|
/Zephyr-latest/soc/intel/intel_adsp/common/include/ |
D | soc_util.h | 12 static ALWAYS_INLINE void bmemcpy(void *dest, void *src, size_t bytes) in bmemcpy() argument 17 sys_cache_data_invd_range(src, bytes); in bmemcpy() 18 for (size_t i = 0; i < (bytes >> 2); i++) { in bmemcpy() 22 sys_cache_data_flush_range(dest, bytes); in bmemcpy() 26 static ALWAYS_INLINE void bbzero(void *dest, size_t bytes) in bbzero() argument 30 for (size_t i = 0; i < (bytes >> 2); i++) { in bbzero() 34 sys_cache_data_flush_range(dest, bytes); in bbzero()
|
/Zephyr-latest/include/zephyr/sys/ |
D | sys_heap.h | 109 void sys_heap_init(struct sys_heap *heap, void *mem, size_t bytes); 128 void *sys_heap_alloc(struct sys_heap *heap, size_t bytes); 143 void *sys_heap_aligned_alloc(struct sys_heap *heap, size_t align, size_t bytes); 179 size_t align, size_t bytes); 181 #define sys_heap_realloc(heap, ptr, bytes) \ argument 182 sys_heap_aligned_realloc(heap, ptr, 0, bytes) 252 void sys_heap_stress(void *(*alloc_fn)(void *arg, size_t bytes),
|
D | heap_listener.h | 72 void *mem, size_t bytes); 91 void *mem, size_t bytes); 148 void heap_listener_notify_alloc(uintptr_t heap_id, void *mem, size_t bytes); 160 void heap_listener_notify_free(uintptr_t heap_id, void *mem, size_t bytes); 275 static inline void heap_listener_notify_alloc(uintptr_t heap_id, void *mem, size_t bytes) 279 ARG_UNUSED(bytes); 282 static inline void heap_listener_notify_free(uintptr_t heap_id, void *mem, size_t bytes) 286 ARG_UNUSED(bytes);
|
D | multi_heap.h | 123 void *sys_multi_heap_alloc(struct sys_multi_heap *mheap, void *cfg, size_t bytes); 139 void *cfg, size_t align, size_t bytes); 192 void *ptr, size_t align, size_t bytes); 194 #define sys_multi_heap_realloc(mheap, cfg, ptr, bytes) \ argument 195 sys_multi_heap_aligned_realloc(mheap, cfg, ptr, 0, bytes)
|
/Zephyr-latest/lib/heap/ |
D | multi_heap.c | 44 void *sys_multi_heap_alloc(struct sys_multi_heap *mheap, void *cfg, size_t bytes) in sys_multi_heap_alloc() argument 46 return mheap->choice(mheap, cfg, 0, bytes); in sys_multi_heap_alloc() 50 void *cfg, size_t align, size_t bytes) in sys_multi_heap_aligned_alloc() argument 52 return mheap->choice(mheap, cfg, align, bytes); in sys_multi_heap_aligned_alloc() 96 void *ptr, size_t align, size_t bytes) in sys_multi_heap_aligned_realloc() argument 100 return sys_multi_heap_aligned_alloc(mheap, cfg, align, bytes); in sys_multi_heap_aligned_realloc() 102 if (bytes == 0) { in sys_multi_heap_aligned_realloc() 112 void *new_ptr = sys_heap_aligned_realloc(rec->heap, ptr, align, bytes); in sys_multi_heap_aligned_realloc() 121 new_ptr = sys_multi_heap_aligned_alloc(mheap, cfg, align, bytes); in sys_multi_heap_aligned_realloc() 123 memcpy(new_ptr, ptr, MIN(old_size, bytes)); in sys_multi_heap_aligned_realloc()
|
D | heap.c | 263 void *sys_heap_alloc(struct sys_heap *heap, size_t bytes) in sys_heap_alloc() argument 268 if ((bytes == 0U) || size_too_big(h, bytes)) { in sys_heap_alloc() 272 chunksz_t chunk_sz = bytes_to_chunksz(h, bytes); in sys_heap_alloc() 297 IF_ENABLED(CONFIG_MSAN, (__msan_allocated_memory(mem, bytes))); in sys_heap_alloc() 301 void *sys_heap_aligned_alloc(struct sys_heap *heap, size_t align, size_t bytes) in sys_heap_aligned_alloc() argument 319 return sys_heap_alloc(heap, bytes); in sys_heap_aligned_alloc() 326 if ((bytes == 0) || size_too_big(h, bytes)) { in sys_heap_aligned_alloc() 335 chunksz_t padded_sz = bytes_to_chunksz(h, bytes + align - gap); in sys_heap_aligned_alloc() 345 chunk_unit_t *end = (chunk_unit_t *) ROUND_UP(mem + bytes, CHUNK_UNIT); in sys_heap_aligned_alloc() 375 IF_ENABLED(CONFIG_MSAN, (__msan_allocated_memory(mem, bytes))); in sys_heap_aligned_alloc() [all …]
|
D | heap.h | 58 typedef struct { char bytes[CHUNK_UNIT]; } chunk_unit_t; member 91 static inline bool big_heap_bytes(size_t bytes) in big_heap_bytes() argument 93 return big_heap_chunks(bytes / CHUNK_UNIT); in big_heap_bytes() 230 static inline chunksz_t chunksz(size_t bytes) in chunksz() argument 232 return (bytes + CHUNK_UNIT - 1U) / CHUNK_UNIT; in chunksz() 235 static inline chunksz_t bytes_to_chunksz(struct z_heap *h, size_t bytes) in bytes_to_chunksz() argument 237 return chunksz(chunk_header_bytes(h) + bytes); in bytes_to_chunksz() 256 static inline bool size_too_big(struct z_heap *h, size_t bytes) in size_too_big() argument 262 return (bytes / CHUNK_UNIT) >= h->end_chunk; in size_too_big()
|
D | heap_listener.c | 31 void heap_listener_notify_alloc(uintptr_t heap_id, void *mem, size_t bytes) in heap_listener_notify_alloc() argument 40 listener->alloc_cb(heap_id, mem, bytes); in heap_listener_notify_alloc() 47 void heap_listener_notify_free(uintptr_t heap_id, void *mem, size_t bytes) in heap_listener_notify_free() argument 56 listener->free_cb(heap_id, mem, bytes); in heap_listener_notify_free()
|
/Zephyr-latest/subsys/net/l2/ppp/ |
D | ppp_stats.h | 17 uint32_t bytes) in ppp_stats_update_bytes_rx() argument 32 stats->bytes.received += bytes; in ppp_stats_update_bytes_rx() 36 uint32_t bytes) in ppp_stats_update_bytes_tx() argument 51 stats->bytes.sent += bytes; in ppp_stats_update_bytes_tx() 128 #define ppp_stats_update_bytes_rx(iface, bytes) argument 129 #define ppp_stats_update_bytes_tx(iface, bytes) argument
|
/Zephyr-latest/kernel/ |
D | kheap.c | 15 void k_heap_init(struct k_heap *heap, void *mem, size_t bytes) in k_heap_init() argument 18 sys_heap_init(&heap->heap, mem, bytes); in k_heap_init() 65 void *k_heap_aligned_alloc(struct k_heap *heap, size_t align, size_t bytes, in k_heap_aligned_alloc() argument 80 ret = sys_heap_aligned_alloc(&heap->heap, align, bytes); in k_heap_aligned_alloc() 108 void *k_heap_alloc(struct k_heap *heap, size_t bytes, k_timeout_t timeout) in k_heap_alloc() argument 112 void *ret = k_heap_aligned_alloc(heap, sizeof(void *), bytes, timeout); in k_heap_alloc() 138 void *k_heap_realloc(struct k_heap *heap, void *ptr, size_t bytes, k_timeout_t timeout) in k_heap_realloc() argument 145 SYS_PORT_TRACING_OBJ_FUNC_ENTER(k_heap, realloc, heap, ptr, bytes, timeout); in k_heap_realloc() 150 ret = sys_heap_aligned_realloc(&heap->heap, ptr, sizeof(void *), bytes); in k_heap_realloc() 162 SYS_PORT_TRACING_OBJ_FUNC_EXIT(k_heap, realloc, heap, ptr, bytes, timeout, ret); in k_heap_realloc()
|
/Zephyr-latest/subsys/usb/device/ |
D | usb_transfer.c | 75 uint32_t bytes; in usb_transfer_work() local 102 ret = usb_write(ep, trans->buffer, trans->bsize, &bytes); in usb_transfer_work() 110 trans->buffer += bytes; in usb_transfer_work() 111 trans->bsize -= bytes; in usb_transfer_work() 112 trans->tsize += bytes; in usb_transfer_work() 115 &bytes); in usb_transfer_work() 123 trans->buffer += bytes; in usb_transfer_work() 124 trans->bsize -= bytes; in usb_transfer_work() 125 trans->tsize += bytes; in usb_transfer_work() 128 if (!bytes || (bytes % usb_dc_ep_mps(ep)) || !trans->bsize) { in usb_transfer_work() [all …]
|
/Zephyr-latest/tests/drivers/can/api/src/ |
D | utilities.c | 47 uint8_t bytes; in ZTEST() local 50 for (bytes = 0; bytes <= 8; bytes++) { in ZTEST() 51 zassert_equal(can_bytes_to_dlc(bytes), bytes, "wrong DLC for %u byte(s)", bytes); in ZTEST()
|
/Zephyr-latest/include/zephyr/arch/xtensa/ |
D | cache.h | 28 static ALWAYS_INLINE int arch_dcache_flush_range(void *addr, size_t bytes) in arch_dcache_flush_range() argument 33 size_t last = ROUND_UP(((long)addr) + bytes, step); in arch_dcache_flush_range() 36 for (line = first; bytes && line < last; line += step) { in arch_dcache_flush_range() 44 static ALWAYS_INLINE int arch_dcache_flush_and_invd_range(void *addr, size_t bytes) in arch_dcache_flush_and_invd_range() argument 49 size_t last = ROUND_UP(((long)addr) + bytes, step); in arch_dcache_flush_and_invd_range() 52 for (line = first; bytes && line < last; line += step) { in arch_dcache_flush_and_invd_range() 60 static ALWAYS_INLINE int arch_dcache_invd_range(void *addr, size_t bytes) in arch_dcache_invd_range() argument 65 size_t last = ROUND_UP(((long)addr) + bytes, step); in arch_dcache_invd_range() 68 for (line = first; bytes && line < last; line += step) { in arch_dcache_invd_range()
|
/Zephyr-latest/drivers/entropy/ |
D | Kconfig.smartbond | 22 Buffer length in bytes used to store entropy bytes generated by the 30 Low water-mark threshold in bytes to trigger entropy generation for 31 thread mode consumers. As soon as the number of available bytes in the 40 Buffer length in bytes used to store entropy bytes generated by the 48 Low water-mark threshold in bytes to trigger entropy generation for 49 ISR consumers. As soon as the number of available bytes in the
|
D | entropy_smartbond.c | 210 static const uint8_t *rng_pool_put_bytes(struct rng_pool *rngp, const uint8_t *bytes, in rng_pool_put_bytes() argument 216 for (; bytes < limit; ++bytes) { in rng_pool_put_bytes() 217 if (rng_pool_put(rngp, *bytes) < 0) { in rng_pool_put_bytes() 223 return bytes; in rng_pool_put_bytes() 278 uint16_t bytes; in entropy_smartbond_get_entropy() local 281 bytes = rng_pool_get((struct rng_pool *)(entropy_smartbond_data.thr), buf, len); in entropy_smartbond_get_entropy() 284 if (bytes == 0U) { in entropy_smartbond_get_entropy() 290 len -= bytes; in entropy_smartbond_get_entropy() 291 buf += bytes; in entropy_smartbond_get_entropy() 329 uint8_t bytes[4]; in entropy_smartbond_get_entropy_isr() local [all …]
|
D | Kconfig.nrf5 | 39 Buffer length in bytes used to store entropy bytes generated by the 47 Low water-mark threshold in bytes to trigger entropy generation for 48 thread mode consumers. As soon as the number of available bytes in the 57 Buffer length in bytes used to store entropy bytes generated by the 65 Low water-mark threshold in bytes to trigger entropy generation for 66 ISR consumers. As soon as the number of available bytes in the
|
D | Kconfig.stm32 | 25 Buffer length in bytes used to store entropy bytes generated by the 33 Low water-mark threshold in bytes to trigger entropy generation for 34 thread mode consumers. As soon as the number of available bytes in the 43 Buffer length in bytes used to store entropy bytes generated by the 51 Low water-mark threshold in bytes to trigger entropy generation for 52 ISR consumers. As soon as the number of available bytes in the
|
/Zephyr-latest/samples/modules/tflite-micro/tflm_ethosu/src/ |
D | inference_process.cpp | 30 if (src.bytes > dst.size) { in copyOutput() 31 printk("Tensor size mismatch (bytes): actual=%d, expected%d.\n", src.bytes, in copyOutput() 36 copy(src.data.uint8, src.data.uint8 + src.bytes, static_cast<uint8_t *>(dst.data)); in copyOutput() 37 dst.size = src.bytes; in copyOutput() 144 if (input.size != tensor->bytes) { in runJob() 146 input.size, tensor->bytes); in runJob() 187 if (expected.size != output->bytes) { in runJob() 189 i, expected.size, output->bytes); in runJob() 193 for (unsigned int j = 0; j < output->bytes; ++j) { in runJob()
|
/Zephyr-latest/subsys/llext/ |
D | llext_priv.h | 28 static inline void *llext_alloc(size_t bytes) in llext_alloc() argument 32 return k_heap_alloc(&llext_heap, bytes, K_NO_WAIT); in llext_alloc() 35 static inline void *llext_aligned_alloc(size_t align, size_t bytes) in llext_aligned_alloc() argument 39 return k_heap_aligned_alloc(&llext_heap, align, bytes, K_NO_WAIT); in llext_aligned_alloc()
|
/Zephyr-latest/drivers/ipm/ |
D | Kconfig.imx | 28 bool "4 bytes" 31 and a maximum size of 4 bytes each. 34 bool "8 bytes" 37 and a maximum size of 8 bytes each. 40 bool "16 bytes" 43 and a maximum size of 16 bytes.
|
/Zephyr-latest/subsys/net/l2/ethernet/ |
D | eth_stats.h | 18 uint32_t bytes) in eth_stats_update_bytes_rx() argument 33 stats->bytes.received += bytes; in eth_stats_update_bytes_rx() 37 uint32_t bytes) in eth_stats_update_bytes_tx() argument 52 stats->bytes.sent += bytes; in eth_stats_update_bytes_tx() 226 #define eth_stats_update_bytes_rx(iface, bytes) argument 227 #define eth_stats_update_bytes_tx(iface, bytes) argument
|
/Zephyr-latest/drivers/crypto/ |
D | Kconfig.it8xxx2 | 12 It requires 256 + 256 bytes in the RAM's first 4k-bytes to calculate 23 It requires 1024 + 256 bytes in the RAM's first 4k-bytes to calculate
|
/Zephyr-latest/subsys/net/ip/ |
D | net_stats.h | 52 uint32_t bytes) in net_stats_update_bytes_recv() argument 54 UPDATE_STAT(iface, stats.bytes.received += bytes); in net_stats_update_bytes_recv() 58 uint32_t bytes) in net_stats_update_bytes_sent() argument 60 UPDATE_STAT(iface, stats.bytes.sent += bytes); in net_stats_update_bytes_sent() 66 #define net_stats_update_bytes_recv(iface, bytes) argument 67 #define net_stats_update_bytes_sent(iface, bytes) argument 237 static inline void net_stats_update_tcp_sent(struct net_if *iface, uint32_t bytes) in net_stats_update_tcp_sent() argument 239 UPDATE_STAT(iface, stats.tcp.bytes.sent += bytes); in net_stats_update_tcp_sent() 242 static inline void net_stats_update_tcp_recv(struct net_if *iface, uint32_t bytes) in net_stats_update_tcp_recv() argument 244 UPDATE_STAT(iface, stats.tcp.bytes.received += bytes); in net_stats_update_tcp_recv() [all …]
|
/Zephyr-latest/include/zephyr/mem_mgmt/ |
D | mem_attr_heap.h | 48 void *mem_attr_heap_alloc(uint32_t attr, size_t bytes); 64 void *mem_attr_heap_aligned_alloc(uint32_t attr, size_t align, size_t bytes);
|