/Linux-v6.6/drivers/infiniband/hw/cxgb4/ |
D | id_table.c | 44 u32 c4iw_id_alloc(struct c4iw_id_table *alloc) in c4iw_id_alloc() argument 49 spin_lock_irqsave(&alloc->lock, flags); in c4iw_id_alloc() 51 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in c4iw_id_alloc() 52 if (obj >= alloc->max) in c4iw_id_alloc() 53 obj = find_first_zero_bit(alloc->table, alloc->max); in c4iw_id_alloc() 55 if (obj < alloc->max) { in c4iw_id_alloc() 56 if (alloc->flags & C4IW_ID_TABLE_F_RANDOM) in c4iw_id_alloc() 57 alloc->last += get_random_u32_below(RANDOM_SKIP); in c4iw_id_alloc() 59 alloc->last = obj + 1; in c4iw_id_alloc() 60 if (alloc->last >= alloc->max) in c4iw_id_alloc() [all …]
|
/Linux-v6.6/drivers/android/ |
D | binder_alloc.c | 60 static size_t binder_alloc_buffer_size(struct binder_alloc *alloc, in binder_alloc_buffer_size() argument 63 if (list_is_last(&buffer->entry, &alloc->buffers)) in binder_alloc_buffer_size() 64 return alloc->buffer + alloc->buffer_size - buffer->user_data; in binder_alloc_buffer_size() 68 static void binder_insert_free_buffer(struct binder_alloc *alloc, in binder_insert_free_buffer() argument 71 struct rb_node **p = &alloc->free_buffers.rb_node; in binder_insert_free_buffer() 79 new_buffer_size = binder_alloc_buffer_size(alloc, new_buffer); in binder_insert_free_buffer() 83 alloc->pid, new_buffer_size, new_buffer); in binder_insert_free_buffer() 90 buffer_size = binder_alloc_buffer_size(alloc, buffer); in binder_insert_free_buffer() 98 rb_insert_color(&new_buffer->rb_node, &alloc->free_buffers); in binder_insert_free_buffer() 102 struct binder_alloc *alloc, struct binder_buffer *new_buffer) in binder_insert_allocated_buffer_locked() argument [all …]
|
D | binder_alloc_selftest.c | 92 static bool check_buffer_pages_allocated(struct binder_alloc *alloc, in check_buffer_pages_allocated() argument 103 page_index = (page_addr - alloc->buffer) / PAGE_SIZE; in check_buffer_pages_allocated() 104 if (!alloc->pages[page_index].page_ptr || in check_buffer_pages_allocated() 105 !list_empty(&alloc->pages[page_index].lru)) { in check_buffer_pages_allocated() 107 alloc->pages[page_index].page_ptr ? in check_buffer_pages_allocated() 115 static void binder_selftest_alloc_buf(struct binder_alloc *alloc, in binder_selftest_alloc_buf() argument 122 buffers[i] = binder_alloc_new_buf(alloc, sizes[i], 0, 0, 0, 0); in binder_selftest_alloc_buf() 124 !check_buffer_pages_allocated(alloc, buffers[i], in binder_selftest_alloc_buf() 132 static void binder_selftest_free_buf(struct binder_alloc *alloc, in binder_selftest_free_buf() argument 139 binder_alloc_free_buf(alloc, buffers[seq[i]]); in binder_selftest_free_buf() [all …]
|
D | binder_alloc.h | 72 struct binder_alloc *alloc; member 117 void binder_selftest_alloc(struct binder_alloc *alloc); 119 static inline void binder_selftest_alloc(struct binder_alloc *alloc) {} in binder_selftest_alloc() argument 124 extern struct binder_buffer *binder_alloc_new_buf(struct binder_alloc *alloc, 130 extern void binder_alloc_init(struct binder_alloc *alloc); 133 extern void binder_alloc_vma_close(struct binder_alloc *alloc); 135 binder_alloc_prepare_to_free(struct binder_alloc *alloc, 137 extern void binder_alloc_free_buf(struct binder_alloc *alloc, 139 extern int binder_alloc_mmap_handler(struct binder_alloc *alloc, 141 extern void binder_alloc_deferred_release(struct binder_alloc *alloc); [all …]
|
D | binder_trace.h | 319 TP_PROTO(struct binder_alloc *alloc, bool allocate, 321 TP_ARGS(alloc, allocate, start, end), 329 __entry->proc = alloc->pid; 331 __entry->offset = start - alloc->buffer; 340 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 341 TP_ARGS(alloc, page_index), 347 __entry->proc = alloc->pid; 355 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), 356 TP_ARGS(alloc, page_index)); 359 TP_PROTO(const struct binder_alloc *alloc, size_t page_index), [all …]
|
/Linux-v6.6/sound/isa/gus/ |
D | gus_mem.c | 18 void snd_gf1_mem_lock(struct snd_gf1_mem * alloc, int xup) in snd_gf1_mem_lock() argument 21 mutex_lock(&alloc->memory_mutex); in snd_gf1_mem_lock() 23 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_lock() 28 snd_gf1_mem_xalloc(struct snd_gf1_mem *alloc, struct snd_gf1_mem_block *block, in snd_gf1_mem_xalloc() argument 43 pblock = alloc->first; in snd_gf1_mem_xalloc() 49 if (pblock == alloc->first) in snd_gf1_mem_xalloc() 50 alloc->first = nblock; in snd_gf1_mem_xalloc() 53 mutex_unlock(&alloc->memory_mutex); in snd_gf1_mem_xalloc() 59 if (alloc->last == NULL) { in snd_gf1_mem_xalloc() 61 alloc->first = alloc->last = nblock; in snd_gf1_mem_xalloc() [all …]
|
/Linux-v6.6/drivers/infiniband/hw/mthca/ |
D | mthca_allocator.c | 40 u32 mthca_alloc(struct mthca_alloc *alloc) in mthca_alloc() argument 45 spin_lock_irqsave(&alloc->lock, flags); in mthca_alloc() 47 obj = find_next_zero_bit(alloc->table, alloc->max, alloc->last); in mthca_alloc() 48 if (obj >= alloc->max) { in mthca_alloc() 49 alloc->top = (alloc->top + alloc->max) & alloc->mask; in mthca_alloc() 50 obj = find_first_zero_bit(alloc->table, alloc->max); in mthca_alloc() 53 if (obj < alloc->max) { in mthca_alloc() 54 __set_bit(obj, alloc->table); in mthca_alloc() 55 obj |= alloc->top; in mthca_alloc() 59 spin_unlock_irqrestore(&alloc->lock, flags); in mthca_alloc() [all …]
|
D | mthca_uar.c | 40 uar->index = mthca_alloc(&dev->uar_table.alloc); in mthca_uar_alloc() 51 mthca_free(&dev->uar_table.alloc, uar->index); in mthca_uar_free() 58 ret = mthca_alloc_init(&dev->uar_table.alloc, in mthca_init_uar_table() 67 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_init_uar_table() 77 mthca_alloc_cleanup(&dev->uar_table.alloc); in mthca_cleanup_uar_table()
|
D | mthca_pd.c | 46 pd->pd_num = mthca_alloc(&dev->pd_table.alloc); in mthca_pd_alloc() 56 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_alloc() 66 mthca_free(&dev->pd_table.alloc, pd->pd_num); in mthca_pd_free() 71 return mthca_alloc_init(&dev->pd_table.alloc, in mthca_init_pd_table() 80 mthca_alloc_cleanup(&dev->pd_table.alloc); in mthca_cleanup_pd_table()
|
/Linux-v6.6/fs/ocfs2/ |
D | localalloc.c | 35 static u32 ocfs2_local_alloc_count_bits(struct ocfs2_dinode *alloc); 38 struct ocfs2_dinode *alloc, 42 static void ocfs2_clear_local_alloc(struct ocfs2_dinode *alloc); 46 struct ocfs2_dinode *alloc, 271 struct ocfs2_dinode *alloc = NULL; in ocfs2_load_local_alloc() local 305 alloc = (struct ocfs2_dinode *) alloc_bh->b_data; in ocfs2_load_local_alloc() 306 la = OCFS2_LOCAL_ALLOC(alloc); in ocfs2_load_local_alloc() 308 if (!(le32_to_cpu(alloc->i_flags) & in ocfs2_load_local_alloc() 325 num_used = ocfs2_local_alloc_count_bits(alloc); in ocfs2_load_local_alloc() 330 || alloc->id1.bitmap1.i_used in ocfs2_load_local_alloc() [all …]
|
/Linux-v6.6/fs/xfs/libxfs/ |
D | xfs_alloc_btree.c | 129 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 132 if (be32_to_cpu(rec->alloc.ar_blockcount) <= in xfs_allocbt_update_lastrec() 135 len = rec->alloc.ar_blockcount; in xfs_allocbt_update_lastrec() 185 key->alloc.ar_startblock = rec->alloc.ar_startblock; in xfs_allocbt_init_key_from_rec() 186 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_allocbt_init_key_from_rec() 196 x = be32_to_cpu(rec->alloc.ar_startblock); in xfs_bnobt_init_high_key_from_rec() 197 x += be32_to_cpu(rec->alloc.ar_blockcount) - 1; in xfs_bnobt_init_high_key_from_rec() 198 key->alloc.ar_startblock = cpu_to_be32(x); in xfs_bnobt_init_high_key_from_rec() 199 key->alloc.ar_blockcount = 0; in xfs_bnobt_init_high_key_from_rec() 207 key->alloc.ar_blockcount = rec->alloc.ar_blockcount; in xfs_cntbt_init_high_key_from_rec() [all …]
|
/Linux-v6.6/rust/alloc/ |
D | raw_vec.rs | 5 use core::alloc::LayoutError; 13 use crate::alloc::handle_alloc_error; 14 use crate::alloc::{Allocator, Global, Layout}; 56 alloc: A, field 122 pub const fn new_in(alloc: A) -> Self { in new_in() 124 Self { ptr: Unique::dangling(), cap: 0, alloc } in new_in() 131 pub fn with_capacity_in(capacity: usize, alloc: A) -> Self { in with_capacity_in() 132 Self::allocate_in(capacity, AllocInit::Uninitialized, alloc) in with_capacity_in() 138 pub fn try_with_capacity_in(capacity: usize, alloc: A) -> Result<Self, TryReserveError> { in try_with_capacity_in() 139 Self::try_allocate_in(capacity, AllocInit::Uninitialized, alloc) in try_with_capacity_in() [all …]
|
D | slice.rs | 27 use crate::alloc::Allocator; 29 use crate::alloc::{self, Global}; 94 use core::alloc::Allocator; 105 let (b, alloc) = Box::into_raw_with_allocator(b); in into_vec() 106 Vec::from_raw_parts_in(b as *mut T, len, len, alloc) in into_vec() 112 pub fn to_vec<T: ConvertVec, A: Allocator>(s: &[T], alloc: A) -> Vec<T, A> { in to_vec() 113 T::to_vec(s, alloc) in to_vec() 118 fn to_vec<A: Allocator>(s: &[Self], alloc: A) -> Vec<Self, A> in to_vec() 126 default fn to_vec<A: Allocator>(s: &[Self], alloc: A) -> Vec<Self, A> { in to_vec() 141 let mut vec = Vec::with_capacity_in(s.len(), alloc); in to_vec() [all …]
|
D | boxed.rs | 171 use crate::alloc::{handle_alloc_error, WriteCloneIntoRaw}; 172 use crate::alloc::{AllocError, Allocator, Global, Layout}; 382 pub fn new_in(x: T, alloc: A) -> Self in new_in() 386 let mut boxed = Self::new_uninit_in(alloc); in new_in() 410 pub fn try_new_in(x: T, alloc: A) -> Result<Self, AllocError> in try_new_in() 414 let mut boxed = Self::try_new_uninit_in(alloc)?; in try_new_in() 445 pub fn new_uninit_in(alloc: A) -> Box<mem::MaybeUninit<T>, A> in new_uninit_in() 452 match Box::try_new_uninit_in(alloc) { in new_uninit_in() 482 pub fn try_new_uninit_in(alloc: A) -> Result<Box<mem::MaybeUninit<T>, A>, AllocError> in try_new_uninit_in() 487 let ptr = alloc.allocate(layout)?.cast(); in try_new_uninit_in() [all …]
|
/Linux-v6.6/lib/zstd/compress/ |
D | zstd_cwksp.h | 247 void* const alloc = (BYTE*)ws->allocStart - bytes; in ZSTD_cwksp_reserve_internal_buffer_space() local 250 alloc, bytes, ZSTD_cwksp_available_space(ws) - bytes); in ZSTD_cwksp_reserve_internal_buffer_space() 252 assert(alloc >= bottom); in ZSTD_cwksp_reserve_internal_buffer_space() 253 if (alloc < bottom) { in ZSTD_cwksp_reserve_internal_buffer_space() 260 if (alloc < ws->tableValidEnd) { in ZSTD_cwksp_reserve_internal_buffer_space() 261 ws->tableValidEnd = alloc; in ZSTD_cwksp_reserve_internal_buffer_space() 263 ws->allocStart = alloc; in ZSTD_cwksp_reserve_internal_buffer_space() 264 return alloc; in ZSTD_cwksp_reserve_internal_buffer_space() 295 void* const alloc = ws->objectEnd; in ZSTD_cwksp_internal_advance_phase() local 296 … size_t const bytesToAlign = ZSTD_cwksp_bytes_to_align_ptr(alloc, ZSTD_CWKSP_ALIGNMENT_BYTES); in ZSTD_cwksp_internal_advance_phase() [all …]
|
/Linux-v6.6/tools/perf/util/ |
D | strbuf.c | 22 sb->alloc = sb->len = 0; in strbuf_init() 31 if (sb->alloc) { in strbuf_release() 39 char *res = sb->alloc ? sb->buf : NULL; in strbuf_detach() 51 if (nr < sb->alloc) in strbuf_grow() 57 if (alloc_nr(sb->alloc) > nr) in strbuf_grow() 58 nr = alloc_nr(sb->alloc); in strbuf_grow() 64 buf = realloc(sb->alloc ? sb->buf : NULL, nr * sizeof(*buf)); in strbuf_grow() 69 sb->alloc = nr; in strbuf_grow() 106 len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap); in strbuf_addv() 117 len = vsnprintf(sb->buf + sb->len, sb->alloc - sb->len, fmt, ap_saved); in strbuf_addv() [all …]
|
D | strbuf.h | 51 size_t alloc; member 65 return sb->alloc ? sb->alloc - sb->len - 1 : 0; in strbuf_avail() 71 if (!sb->alloc) { in strbuf_setlen() 76 assert(len < sb->alloc); in strbuf_setlen()
|
/Linux-v6.6/tools/lib/subcmd/ |
D | subcmd-util.h | 38 #define ALLOC_GROW(x, nr, alloc) \ argument 40 if ((nr) > alloc) { \ 41 if (alloc_nr(alloc) < (nr)) \ 42 alloc = (nr); \ 44 alloc = alloc_nr(alloc); \ 45 x = xrealloc((x), alloc * sizeof(*(x))); \
|
/Linux-v6.6/net/core/ |
D | page_pool.c | 294 pool->alloc.cache[pool->alloc.count++] = page; in page_pool_refill_alloc_cache() 306 } while (pool->alloc.count < PP_ALLOC_CACHE_REFILL); in page_pool_refill_alloc_cache() 309 if (likely(pool->alloc.count > 0)) { in page_pool_refill_alloc_cache() 310 page = pool->alloc.cache[--pool->alloc.count]; in page_pool_refill_alloc_cache() 323 if (likely(pool->alloc.count)) { in __page_pool_get_cached() 325 page = pool->alloc.cache[--pool->alloc.count]; in __page_pool_get_cached() 426 if (unlikely(pool->alloc.count > 0)) in __page_pool_alloc_pages_slow() 427 return pool->alloc.cache[--pool->alloc.count]; in __page_pool_alloc_pages_slow() 430 memset(&pool->alloc.cache, 0, sizeof(void *) * bulk); in __page_pool_alloc_pages_slow() 433 pool->alloc.cache); in __page_pool_alloc_pages_slow() [all …]
|
/Linux-v6.6/arch/xtensa/variants/dc233c/include/variant/ |
D | tie-asm.h | 77 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 85 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 97 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 115 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 141 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 149 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 161 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 179 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
|
/Linux-v6.6/arch/xtensa/variants/csp/include/variant/ |
D | tie-asm.h | 76 .macro xchal_ncp_store ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 84 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 96 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 116 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 142 .macro xchal_ncp_load ptr at1 at2 at3 at4 continue=0 ofs=-1 select=XTHAL_SAS_ALL alloc=0 150 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_GLOB) & ~(\alloc)) == 0 162 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_CC | XTHAL_SAS_CALR) & ~(\alloc)) == 0 182 .elseif ((XTHAL_SAS_OPT | XTHAL_SAS_NOCC | XTHAL_SAS_CALR) & ~(\alloc)) == 0
|
/Linux-v6.6/arch/s390/mm/ |
D | pgalloc.c | 531 unsigned long end, int alloc) in base_page_walk() argument 535 if (!alloc) in base_page_walk() 547 unsigned long end, int alloc) in base_segment_walk() argument 557 if (!alloc) in base_segment_walk() 565 rc = base_page_walk(table, addr, next, alloc); in base_segment_walk() 568 if (!alloc) in base_segment_walk() 576 unsigned long end, int alloc) in base_region3_walk() argument 586 if (!alloc) in base_region3_walk() 594 rc = base_segment_walk(table, addr, next, alloc); in base_region3_walk() 597 if (!alloc) in base_region3_walk() [all …]
|
/Linux-v6.6/fs/nfs/ |
D | nfs3acl.c | 260 struct posix_acl *orig = acl, *dfacl = NULL, *alloc; in nfs3_set_acl() local 267 alloc = get_inode_acl(inode, ACL_TYPE_DEFAULT); in nfs3_set_acl() 268 if (IS_ERR(alloc)) in nfs3_set_acl() 270 dfacl = alloc; in nfs3_set_acl() 274 alloc = get_inode_acl(inode, ACL_TYPE_ACCESS); in nfs3_set_acl() 275 if (IS_ERR(alloc)) in nfs3_set_acl() 278 acl = alloc; in nfs3_set_acl() 284 alloc = posix_acl_from_mode(inode->i_mode, GFP_KERNEL); in nfs3_set_acl() 285 if (IS_ERR(alloc)) in nfs3_set_acl() 287 acl = alloc; in nfs3_set_acl() [all …]
|
/Linux-v6.6/rust/ |
D | Makefile | 18 obj-$(CONFIG_RUST) += alloc.o bindings.o kernel.o 63 alloc-cfgs = \ 95 rustdoc-alloc rustdoc-kernel 122 rustdoc-alloc: private rustc_target_flags = $(alloc-cfgs) \ 124 rustdoc-alloc: $(src)/alloc/lib.rs rustdoc-core rustdoc-compiler_builtins FORCE 127 rustdoc-kernel: private rustc_target_flags = --extern alloc \ 131 rustdoc-compiler_builtins rustdoc-alloc $(obj)/libmacros.so \ 177 -L$(objtree)/$(obj) --extern alloc --extern kernel \ 239 cp -r $(srctree)/$(src)/alloc/* \ 240 $(objtree)/$(obj)/test/sysroot/lib/rustlib/src/rust/library/alloc/src; \ [all …]
|
/Linux-v6.6/drivers/char/agp/ |
D | compat_ioctl.c | 155 struct agp_allocate32 alloc; in compat_agpioc_allocate_wrap() local 158 if (copy_from_user(&alloc, arg, sizeof(alloc))) in compat_agpioc_allocate_wrap() 161 memory = agp_allocate_memory_wrap(alloc.pg_count, alloc.type); in compat_agpioc_allocate_wrap() 166 alloc.key = memory->key; in compat_agpioc_allocate_wrap() 167 alloc.physical = memory->physical; in compat_agpioc_allocate_wrap() 169 if (copy_to_user(arg, &alloc, sizeof(alloc))) { in compat_agpioc_allocate_wrap()
|