Home
last modified time | relevance | path

Searched full:chunks (Results 1 – 25 of 730) sorted by relevance

12345678910>>...30

/Linux-v6.6/mm/
Dzbud.c31 * zbud pages are divided into "chunks". The size of the chunks is fixed at
33 * into chunks allows organizing unbuddied zbud pages into a manageable number
34 * of unbuddied lists according to the number of free chunks available in the
63 * allocation granularity will be in chunks of size PAGE_SIZE/64. As one chunk
65 * 63 which shows the max number of free chunks in zbud page, also there will be
108 * @first_chunks: the size of the first buddy in chunks, 0 if free
109 * @last_chunks: the size of the last buddy in chunks, 0 if free
126 /* Converts an allocation size in bytes to size in zbud chunks */
180 /* Returns the number of free chunks in a zbud page */
249 int chunks, i, freechunks; in zbud_alloc() local
[all …]
Dz3fold.c18 * As in zbud, pages are divided into "chunks". The size of the chunks is
48 * allocation granularity will be in chunks of size PAGE_SIZE/64. Some chunks
51 * which shows the max number of free chunks in z3fold page, also there will
92 * struct z3fold_header - z3fold page metadata occupying first chunks of each
102 * @first_chunks: the size of the first buddy in chunks, 0 if free
103 * @middle_chunks: the size of the middle buddy in chunks, 0 if free
104 * @last_chunks: the size of the last buddy in chunks, 0 if free
185 /* Converts an allocation size in bytes to size in z3fold chunks */
514 * Returns the number of free chunks in a z3fold page.
523 * of chunks occupied by the first and the last objects. in num_free_chunks()
[all …]
/Linux-v6.6/net/sctp/
Dchunk.c43 INIT_LIST_HEAD(&msg->chunks); in sctp_datamsg_init()
65 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free()
81 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_destroy()
140 * down any such message into smaller chunks. Opportunistically, fragment
141 * the chunks down to the current MTU constraints. We may get refragmented
182 /* If the peer requested that we authenticate DATA chunks in sctp_datamsg_from_user()
183 * we need to account for bundling of the AUTH chunks along with in sctp_datamsg_from_user()
235 /* Create chunks for all DATA chunks. */ in sctp_datamsg_from_user()
280 list_add_tail(&chunk->frag_list, &msg->chunks); in sctp_datamsg_from_user()
289 list_for_each_safe(pos, temp, &msg->chunks) { in sctp_datamsg_from_user()
Doutqueue.c11 * bundling and queueing of outgoing SCTP chunks.
204 /* Free the outqueue structure and any related pending chunks.
212 /* Throw away unacknowledged chunks. */ in __sctp_outq_teardown()
224 /* Throw away chunks that have been gap ACKed. */ in __sctp_outq_teardown()
233 /* Throw away any chunks in the retransmit queue. */ in __sctp_outq_teardown()
242 /* Throw away any chunks that are in the abandoned queue. */ in __sctp_outq_teardown()
251 /* Throw away any leftover data chunks. */ in __sctp_outq_teardown()
260 /* Throw away any leftover control chunks. */ in __sctp_outq_teardown()
273 /* Free the outqueue structure and any related pending chunks. */
276 /* Throw away leftover chunks. */ in sctp_outq_free()
[all …]
Dauth.c175 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO
186 struct sctp_chunks_param *chunks, in sctp_auth_make_key_vector() argument
197 if (chunks) in sctp_auth_make_key_vector()
198 chunks_len = ntohs(chunks->param_hdr.length); in sctp_auth_make_key_vector()
209 if (chunks) { in sctp_auth_make_key_vector()
210 memcpy(new->data + offset, chunks, chunks_len); in sctp_auth_make_key_vector()
301 * The RANDOM parameter, the CHUNKS parameter and the HMAC-ALGO in sctp_auth_asoc_create_secret()
651 * chunks MUST NOT be listed in the CHUNKS parameter. However, if in __sctp_auth_cid()
652 * a CHUNKS parameter is received then the types for INIT, INIT-ACK, in __sctp_auth_cid()
653 * SHUTDOWN-COMPLETE and AUTH chunks MUST be ignored. in __sctp_auth_cid()
[all …]
Dulpqueue.c102 /* Create a temporary list to collect chunks on. */ in sctp_ulpq_tail_data()
266 /* Helper function to store chunks that need to be reassembled. */
412 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_reassembled()
414 * fragmented chunks that complete a datagram. in sctp_ulpq_retrieve_reassembled()
509 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_partial()
511 * sequence of fragmented chunks. in sctp_ulpq_retrieve_partial()
568 /* Helper function to reassemble chunks. Hold chunks on the reasm queue that
608 /* The chunks are held in the reasm queue sorted by TSN. in sctp_ulpq_retrieve_first()
610 * fragmented chunks that start a datagram. in sctp_ulpq_retrieve_first()
753 /* We are holding the chunks by stream, by SSN. */ in sctp_ulpq_retrieve_ordered()
[all …]
/Linux-v6.6/drivers/gpu/drm/radeon/
Dradeon_cs.c284 /* get chunks */ in radeon_cs_parser_init()
296 chunk_array_ptr = (uint64_t *)(unsigned long)(cs->chunks); in radeon_cs_parser_init()
303 p->chunks = kvcalloc(p->nchunks, sizeof(struct radeon_cs_chunk), GFP_KERNEL); in radeon_cs_parser_init()
304 if (p->chunks == NULL) { in radeon_cs_parser_init()
317 p->chunks[i].length_dw = user_chunk.length_dw; in radeon_cs_parser_init()
319 p->chunk_relocs = &p->chunks[i]; in radeon_cs_parser_init()
322 p->chunk_ib = &p->chunks[i]; in radeon_cs_parser_init()
324 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
328 p->chunk_const_ib = &p->chunks[i]; in radeon_cs_parser_init()
330 if (p->chunks[i].length_dw == 0) in radeon_cs_parser_init()
[all …]
/Linux-v6.6/arch/mips/ar7/
Dprom.c102 Name=Value pair in 2 chunks (len is the number of chunks)
149 struct psp_env_chunk *chunks = (struct psp_env_chunk *)psp_env_data; in parse_psp_env() local
151 memcpy_fromio(chunks, psp_env_base, PSP_ENV_SIZE); in parse_psp_env()
156 if ((chunks[i].num == 0xff) || ((i + chunks[i].len) > n)) in parse_psp_env()
158 value = chunks[i].data; in parse_psp_env()
159 if (chunks[i].num) { in parse_psp_env()
160 name = lookup_psp_var_map(chunks[i].num); in parse_psp_env()
167 i += chunks[i].len; in parse_psp_env()
/Linux-v6.6/drivers/net/ethernet/netronome/nfp/nfpcore/
Dnfp_nsp.c505 } *chunks; in nfp_nsp_command_buf_dma_sg() local
517 chunks = kcalloc(nseg, sizeof(*chunks), GFP_KERNEL); in nfp_nsp_command_buf_dma_sg()
518 if (!chunks) in nfp_nsp_command_buf_dma_sg()
526 chunks[i].chunk = kmalloc(chunk_size, in nfp_nsp_command_buf_dma_sg()
528 if (!chunks[i].chunk) in nfp_nsp_command_buf_dma_sg()
531 chunks[i].len = min_t(u64, chunk_size, max_size - off); in nfp_nsp_command_buf_dma_sg()
536 memcpy(chunks[i].chunk, arg->in_buf + off, coff); in nfp_nsp_command_buf_dma_sg()
538 memset(chunks[i].chunk + coff, 0, chunk_size - coff); in nfp_nsp_command_buf_dma_sg()
540 off += chunks[i].len; in nfp_nsp_command_buf_dma_sg()
548 addr = dma_map_single(dev, chunks[i].chunk, chunks[i].len, in nfp_nsp_command_buf_dma_sg()
[all …]
/Linux-v6.6/drivers/infiniband/ulp/rtrs/
DREADME28 session. A session is associated with a set of memory chunks reserved on the
36 chunks reserved for him on the server side. Their number, size and addresses
45 which of the memory chunks has been accessed and at which offset the message
80 the server (number of memory chunks which are going to be allocated for that
122 1. When processing a write request client selects one of the memory chunks
139 1. When processing a write request client selects one of the memory chunks
144 using the IMM field, Server invalidate rkey associated to the memory chunks
162 1. When processing a read request client selects one of the memory chunks
181 1. When processing a read request client selects one of the memory chunks
186 Server invalidate rkey associated to the memory chunks first, when it finishes,
/Linux-v6.6/include/net/sctp/
Dstructs.h372 struct sctp_chunks_param *chunks; member
521 /* Chunks waiting to be submitted to lower layer. */
522 struct list_head chunks; member
565 /* This field is used by chunks that hold fragmented data.
642 * spec violates the principle premis that all chunks are processed
703 /* This structure holds lists of chunks as we are assembling for
712 /* This contains the payload chunks. */
717 /* This is the total size of all chunks INCLUDING padding. */
782 * chunks sent to this address is currently being
955 /* This is the list of transports that have chunks to send. */
[all …]
/Linux-v6.6/tools/testing/selftests/drivers/net/mlxsw/spectrum/
Ddevlink_lib_spectrum.sh13 KVDL_CHILDREN="singles chunks large_chunks"
90 devlink_resource_size_set 32000 kvd linear chunks
99 devlink_resource_size_set 32000 kvd linear chunks
108 devlink_resource_size_set 49152 kvd linear chunks
/Linux-v6.6/fs/ocfs2/
Dquota.h40 struct list_head rc_list; /* List of chunks */
46 struct list_head r_list[OCFS2_MAXQUOTAS]; /* List of chunks to recover */
53 unsigned int dqi_chunks; /* Number of chunks in local quota file */
56 struct list_head dqi_chunk; /* List of chunks */
78 struct list_head qc_chunk; /* List of quotafile chunks */
/Linux-v6.6/net/xdp/
Dxdp_umem.c157 u64 chunks, npgs; in xdp_umem_reg() local
192 chunks = div_u64_rem(size, chunk_size, &chunks_rem); in xdp_umem_reg()
193 if (!chunks || chunks > U32_MAX) in xdp_umem_reg()
205 umem->chunks = chunks; in xdp_umem_reg()
/Linux-v6.6/drivers/infiniband/hw/usnic/
Dusnic_vnic.c44 struct usnic_vnic_res_chunk chunks[USNIC_VNIC_RES_TYPE_MAX]; member
117 for (i = 0; i < ARRAY_SIZE(vnic->chunks); i++) { in usnic_vnic_dump()
118 chunk = &vnic->chunks[i]; in usnic_vnic_dump()
222 return vnic->chunks[type].cnt; in usnic_vnic_res_cnt()
228 return vnic->chunks[type].free_cnt; in usnic_vnic_res_free_cnt()
254 src = &vnic->chunks[type]; in usnic_vnic_get_resources()
286 vnic->chunks[res->type].free_cnt++; in usnic_vnic_put_resources()
382 &vnic->chunks[res_type]); in usnic_vnic_discover_resources()
391 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_discover_resources()
427 usnic_vnic_free_res_chunk(&vnic->chunks[res_type]); in usnic_vnic_release_resources()
/Linux-v6.6/kernel/
Daudit_tree.c17 struct list_head chunks; member
68 * tree.chunks anchors chunk.owners[].list hash_lock
101 INIT_LIST_HEAD(&tree->chunks); in alloc_tree()
435 list_add(&chunk->owners[0].list, &tree->chunks); in create_chunk()
507 list_add(&p->list, &tree->chunks); in tag_chunk()
565 * Remove tree from chunks. If 'tagged' is set, remove tree only from tagged
566 * chunks. The function expects tagged chunks are all at the beginning of the
567 * chunks list.
572 while (!list_empty(&victim->chunks)) { in prune_tree_chunks()
577 p = list_first_entry(&victim->chunks, struct audit_node, list); in prune_tree_chunks()
[all …]
/Linux-v6.6/scripts/gdb/linux/
Dtimerlist.py163 chunks = []
169 chunks.append(buf[start:end])
171 chunks.append(',')
175 chunks[0] = chunks[0][0] # Cut off the first 0
177 return "".join(str(chunks))
/Linux-v6.6/drivers/mtd/nand/spi/
Dalliancememory.c78 * The free area in this kind of flash is divided into chunks where the in am_ooblayout_free()
79 * first 4 bytes of each chunk are unprotected. The number of chunks in am_ooblayout_free()
81 * have 8 chunks, the others 4 chunks. in am_ooblayout_free()
/Linux-v6.6/drivers/infiniband/hw/efa/
Defa_verbs.c110 struct pbl_chunk *chunks; member
1281 /* allocate a chunk list that consists of 4KB chunks */ in pbl_chunk_list_create()
1285 chunk_list->chunks = kcalloc(chunk_list_size, in pbl_chunk_list_create()
1286 sizeof(*chunk_list->chunks), in pbl_chunk_list_create()
1288 if (!chunk_list->chunks) in pbl_chunk_list_create()
1297 chunk_list->chunks[i].buf = kzalloc(EFA_CHUNK_SIZE, GFP_KERNEL); in pbl_chunk_list_create()
1298 if (!chunk_list->chunks[i].buf) in pbl_chunk_list_create()
1301 chunk_list->chunks[i].length = EFA_CHUNK_USED_SIZE; in pbl_chunk_list_create()
1303 chunk_list->chunks[chunk_list_size - 1].length = in pbl_chunk_list_create()
1307 /* fill the dma addresses of sg list pages to chunks: */ in pbl_chunk_list_create()
[all …]
/Linux-v6.6/drivers/comedi/drivers/ni_routing/tools/
Dconvert_csv_to_c.py228 chunks = [ self.output_file_top,
244 chunks.append('\t&{},'.format(dev_table_name))
273 chunks.append('\tNULL,') # terminate list
274 chunks.append('};')
275 return '\n'.join(chunks)
416 chunks = [ self.output_file_top,
432 chunks.append('\t&{},'.format(fam_table_name))
462 chunks.append('\tNULL,') # terminate list
463 chunks.append('};')
464 return '\n'.join(chunks)
/Linux-v6.6/arch/xtensa/lib/
Dchecksum.S51 srli a5, a3, 5 /* 32-byte chunks */
81 extui a5, a3, 2, 3 /* remaining 4-byte chunks */
131 srli a5, a3, 2 /* 4-byte chunks */
215 srli a10, a4, 5 /* 32-byte chunks */
254 extui a10, a4, 2, 3 /* remaining 4-byte chunks */
255 extui a4, a4, 0, 2 /* reset len for general-case, 2-byte chunks */
286 srli a10, a4, 1 /* 2-byte chunks */
/Linux-v6.6/drivers/md/
Ddm-snap-persistent.c35 * requires that we copy exception chunks to chunk aligned areas
47 * followed by as many exception chunks as can fit in the
114 * whole chunks worth of metadata in memory at once.
139 * When creating exceptions, all the chunks here and above are
142 * the exception store because chunks can be committed out of
146 * chunks here and above are free. It holds the value it would
147 * have held if all chunks had been committed in order of
472 * Keep track of the start of the free chunks. in insert_exceptions()
510 * Keeping reading chunks and inserting exceptions until in read_exceptions()
585 * Then there are (ps->current_area + 1) metadata chunks, each one in persistent_usage()
[all …]
/Linux-v6.6/drivers/parisc/
Diommu-helpers.h5 * iommu_fill_pdir - Insert coalesced scatter/gather chunks into the I/O Pdir.
7 * @startsg: The scatter/gather list of coalesced chunks.
11 * This function inserts the coalesced scatter/gather list chunks into the
93 ** Returns the number of DMA chunks.
/Linux-v6.6/arch/x86/kernel/cpu/resctrl/
Dmonitor.c226 u64 shift = 64 - width, chunks; in mbm_overflow_count() local
228 chunks = (cur_msr << shift) - (prev_msr << shift); in mbm_overflow_count()
229 return chunks >> shift; in mbm_overflow_count()
238 u64 msr_val, chunks; in resctrl_arch_rmid_read() local
250 am->chunks += mbm_overflow_count(am->prev_msr, msr_val, in resctrl_arch_rmid_read()
252 chunks = get_corrected_mbm_count(rmid, am->chunks); in resctrl_arch_rmid_read()
255 chunks = msr_val; in resctrl_arch_rmid_read()
258 *val = chunks * hw_res->mon_scale; in resctrl_arch_rmid_read()
428 * and delta bandwidth in MBps. The chunks value previously read by
429 * __mon_event_count() is compared with the chunks value from the previous
/Linux-v6.6/drivers/dma/sh/
Dshdma-base.c85 /* Mark all chunks of this descriptor as submitted, move to the queue */ in shdma_tx_submit()
88 * All chunks are on the global ld_free, so, we have to find in shdma_tx_submit()
97 if (chunk->chunks == 1) { in shdma_tx_submit()
134 * Make it int then, on error remove chunks from the in shdma_tx_submit()
347 * any chunks in a (partially) completed chain in __ld_cleanup()
356 if (desc->mark == DESC_COMPLETED && desc->chunks == 1) { in __ld_cleanup()
372 BUG_ON(desc->chunks != 1); in __ld_cleanup()
567 int chunks = 0; in shdma_prep_sg() local
572 chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len); in shdma_prep_sg()
582 * if more than one chunk is needed further chunks have cookie = -EINVAL in shdma_prep_sg()
[all …]

12345678910>>...30