/Linux-v4.19/net/sctp/ |
D | inqueue.c | 59 struct sctp_chunk *chunk, *tmp; in sctp_inq_free() local 62 list_for_each_entry_safe(chunk, tmp, &queue->in_chunk_list, list) { in sctp_inq_free() 63 list_del_init(&chunk->list); in sctp_inq_free() 64 sctp_chunk_free(chunk); in sctp_inq_free() 79 void sctp_inq_push(struct sctp_inq *q, struct sctp_chunk *chunk) in sctp_inq_push() argument 82 if (chunk->rcvr->dead) { in sctp_inq_push() 83 sctp_chunk_free(chunk); in sctp_inq_push() 92 list_add_tail(&chunk->list, &q->in_chunk_list); in sctp_inq_push() 93 if (chunk->asoc) in sctp_inq_push() 94 chunk->asoc->stats.ipackets++; in sctp_inq_push() [all …]
|
D | chunk.c | 75 struct sctp_chunk *chunk; in sctp_datamsg_free() local 80 list_for_each_entry(chunk, &msg->chunks, frag_list) in sctp_datamsg_free() 81 sctp_chunk_free(chunk); in sctp_datamsg_free() 90 struct sctp_chunk *chunk; in sctp_datamsg_destroy() local 102 chunk = list_entry(pos, struct sctp_chunk, frag_list); in sctp_datamsg_destroy() 105 asoc = chunk->asoc; in sctp_datamsg_destroy() 119 if (chunk->has_tsn) in sctp_datamsg_destroy() 124 ev = sctp_ulpevent_make_send_failed(asoc, chunk, sent, in sctp_datamsg_destroy() 130 sctp_chunk_put(chunk); in sctp_datamsg_destroy() 151 static void sctp_datamsg_assign(struct sctp_datamsg *msg, struct sctp_chunk *chunk) in sctp_datamsg_assign() argument [all …]
|
D | output.c | 61 struct sctp_chunk *chunk); 63 struct sctp_chunk *chunk); 65 struct sctp_chunk *chunk); 67 struct sctp_chunk *chunk, 133 struct sctp_chunk *chunk = sctp_get_ecne_prepend(asoc); in sctp_packet_config() local 135 if (chunk) in sctp_packet_config() 136 sctp_packet_append_chunk(packet, chunk); in sctp_packet_config() 173 struct sctp_chunk *chunk, *tmp; in sctp_packet_free() local 177 list_for_each_entry_safe(chunk, tmp, &packet->chunk_list, list) { in sctp_packet_free() 178 list_del_init(&chunk->list); in sctp_packet_free() [all …]
|
D | outqueue.c | 224 struct sctp_chunk *chunk, *tmp; in __sctp_outq_teardown() local 230 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 233 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 234 sctp_chunk_free(chunk); in __sctp_outq_teardown() 241 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 243 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 244 sctp_chunk_free(chunk); in __sctp_outq_teardown() 250 chunk = list_entry(lchunk, struct sctp_chunk, in __sctp_outq_teardown() 252 sctp_chunk_fail(chunk, q->error); in __sctp_outq_teardown() 253 sctp_chunk_free(chunk); in __sctp_outq_teardown() [all …]
|
D | sm_statefuns.c | 69 struct sctp_chunk *chunk, 72 struct sctp_chunk *chunk, 77 const struct sctp_chunk *chunk); 81 const struct sctp_chunk *chunk, 105 static struct sctp_sackhdr *sctp_sm_pull_sack(struct sctp_chunk *chunk); 157 struct sctp_chunk *chunk); 174 static inline bool sctp_chunk_length_valid(struct sctp_chunk *chunk, in sctp_chunk_length_valid() argument 177 __u16 chunk_length = ntohs(chunk->chunk_hdr->length); in sctp_chunk_length_valid() 180 if (unlikely(chunk->pdiscard)) in sctp_chunk_length_valid() 229 struct sctp_chunk *chunk = arg; in sctp_sf_do_4_C() local [all …]
|
D | sm_make_chunk.c | 82 static void *sctp_addto_param(struct sctp_chunk *chunk, int len, 88 struct sctp_chunk *chunk = skb_shinfo(skb)->destructor_arg; in sctp_control_release_owner() local 90 if (chunk->shkey) { in sctp_control_release_owner() 91 struct sctp_shared_key *shkey = chunk->shkey; in sctp_control_release_owner() 92 struct sctp_association *asoc = chunk->asoc; in sctp_control_release_owner() 108 sctp_auth_shkey_release(chunk->shkey); in sctp_control_release_owner() 112 static void sctp_control_set_owner_w(struct sctp_chunk *chunk) in sctp_control_set_owner_w() argument 114 struct sctp_association *asoc = chunk->asoc; in sctp_control_set_owner_w() 115 struct sk_buff *skb = chunk->skb; in sctp_control_set_owner_w() 124 if (chunk->auth) { in sctp_control_set_owner_w() [all …]
|
D | endpointola.c | 383 struct sctp_chunk *chunk; in sctp_endpoint_bh_rcv() local 398 while (NULL != (chunk = sctp_inq_pop(inqueue))) { in sctp_endpoint_bh_rcv() 399 subtype = SCTP_ST_CHUNK(chunk->chunk_hdr->type); in sctp_endpoint_bh_rcv() 404 if (first_time && (subtype.chunk == SCTP_CID_AUTH)) { in sctp_endpoint_bh_rcv() 417 chunk->auth_chunk = skb_clone(chunk->skb, in sctp_endpoint_bh_rcv() 419 chunk->auth = 1; in sctp_endpoint_bh_rcv() 430 if (NULL == chunk->asoc) { in sctp_endpoint_bh_rcv() 432 sctp_source(chunk), in sctp_endpoint_bh_rcv() 434 chunk->asoc = asoc; in sctp_endpoint_bh_rcv() 435 chunk->transport = transport; in sctp_endpoint_bh_rcv() [all …]
|
/Linux-v4.19/mm/ |
D | percpu-vm.c | 13 static struct page *pcpu_chunk_page(struct pcpu_chunk *chunk, in pcpu_chunk_page() argument 17 WARN_ON(chunk->immutable); in pcpu_chunk_page() 19 return vmalloc_to_page((void *)pcpu_chunk_addr(chunk, cpu, page_idx)); in pcpu_chunk_page() 54 static void pcpu_free_pages(struct pcpu_chunk *chunk, in pcpu_free_pages() argument 82 static int pcpu_alloc_pages(struct pcpu_chunk *chunk, in pcpu_alloc_pages() argument 127 static void pcpu_pre_unmap_flush(struct pcpu_chunk *chunk, in pcpu_pre_unmap_flush() argument 131 pcpu_chunk_addr(chunk, pcpu_low_unit_cpu, page_start), in pcpu_pre_unmap_flush() 132 pcpu_chunk_addr(chunk, pcpu_high_unit_cpu, page_end)); in pcpu_pre_unmap_flush() 153 static void pcpu_unmap_pages(struct pcpu_chunk *chunk, in pcpu_unmap_pages() argument 163 page = pcpu_chunk_page(chunk, cpu, i); in pcpu_unmap_pages() [all …]
|
D | percpu.c | 205 static bool pcpu_addr_in_chunk(struct pcpu_chunk *chunk, void *addr) in pcpu_addr_in_chunk() argument 209 if (!chunk) in pcpu_addr_in_chunk() 212 start_addr = chunk->base_addr + chunk->start_offset; in pcpu_addr_in_chunk() 213 end_addr = chunk->base_addr + chunk->nr_pages * PAGE_SIZE - in pcpu_addr_in_chunk() 214 chunk->end_offset; in pcpu_addr_in_chunk() 232 static int pcpu_chunk_slot(const struct pcpu_chunk *chunk) in pcpu_chunk_slot() argument 234 if (chunk->free_bytes < PCPU_MIN_ALLOC_SIZE || chunk->contig_bits == 0) in pcpu_chunk_slot() 237 return pcpu_size_to_slot(chunk->free_bytes); in pcpu_chunk_slot() 262 static unsigned long pcpu_chunk_addr(struct pcpu_chunk *chunk, in pcpu_chunk_addr() argument 265 return (unsigned long)chunk->base_addr + in pcpu_chunk_addr() [all …]
|
D | percpu-km.c | 36 static int pcpu_populate_chunk(struct pcpu_chunk *chunk, in pcpu_populate_chunk() argument 42 static void pcpu_depopulate_chunk(struct pcpu_chunk *chunk, in pcpu_depopulate_chunk() argument 51 struct pcpu_chunk *chunk; in pcpu_create_chunk() local 55 chunk = pcpu_alloc_chunk(gfp); in pcpu_create_chunk() 56 if (!chunk) in pcpu_create_chunk() 61 pcpu_free_chunk(chunk); in pcpu_create_chunk() 66 pcpu_set_page_chunk(nth_page(pages, i), chunk); in pcpu_create_chunk() 68 chunk->data = pages; in pcpu_create_chunk() 69 chunk->base_addr = page_address(pages) - pcpu_group_offsets[0]; in pcpu_create_chunk() 72 pcpu_chunk_populated(chunk, 0, nr_pages, false); in pcpu_create_chunk() [all …]
|
D | percpu-stats.c | 36 struct pcpu_chunk *chunk; in find_max_nr_alloc() local 41 list_for_each_entry(chunk, &pcpu_slot[slot], list) in find_max_nr_alloc() 42 max_nr_alloc = max(max_nr_alloc, chunk->nr_alloc); in find_max_nr_alloc() 53 static void chunk_map_stats(struct seq_file *m, struct pcpu_chunk *chunk, in chunk_map_stats() argument 69 last_alloc = find_last_bit(chunk->alloc_map, in chunk_map_stats() 70 pcpu_chunk_map_bits(chunk) - in chunk_map_stats() 71 chunk->end_offset / PCPU_MIN_ALLOC_SIZE - 1); in chunk_map_stats() 72 last_alloc = test_bit(last_alloc, chunk->alloc_map) ? in chunk_map_stats() 76 start = chunk->start_offset / PCPU_MIN_ALLOC_SIZE; in chunk_map_stats() 88 if (test_bit(start, chunk->alloc_map)) { in chunk_map_stats() [all …]
|
D | percpu-internal.h | 73 static inline int pcpu_chunk_nr_blocks(struct pcpu_chunk *chunk) in pcpu_chunk_nr_blocks() argument 75 return chunk->nr_pages * PAGE_SIZE / PCPU_BITMAP_BLOCK_SIZE; in pcpu_chunk_nr_blocks() 97 static inline int pcpu_chunk_map_bits(struct pcpu_chunk *chunk) in pcpu_chunk_map_bits() argument 99 return pcpu_nr_pages_to_map_bits(chunk->nr_pages); in pcpu_chunk_map_bits() 139 static inline void pcpu_stats_area_alloc(struct pcpu_chunk *chunk, size_t size) in pcpu_stats_area_alloc() argument 152 chunk->nr_alloc++; in pcpu_stats_area_alloc() 153 chunk->max_alloc_size = max(chunk->max_alloc_size, size); in pcpu_stats_area_alloc() 163 static inline void pcpu_stats_area_dealloc(struct pcpu_chunk *chunk) in pcpu_stats_area_dealloc() argument 170 chunk->nr_alloc--; in pcpu_stats_area_dealloc() 207 static inline void pcpu_stats_area_alloc(struct pcpu_chunk *chunk, size_t size) in pcpu_stats_area_alloc() argument [all …]
|
/Linux-v4.19/drivers/s390/cio/ |
D | itcw.c | 182 void *chunk; in itcw_init() local 194 chunk = fit_chunk(&start, end, sizeof(struct itcw), 1, 0); in itcw_init() 195 if (IS_ERR(chunk)) in itcw_init() 196 return chunk; in itcw_init() 197 itcw = chunk; in itcw_init() 210 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() 211 if (IS_ERR(chunk)) in itcw_init() 212 return chunk; in itcw_init() 213 itcw->tcw = chunk; in itcw_init() 218 chunk = fit_chunk(&start, end, sizeof(struct tcw), 64, 0); in itcw_init() [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | icm.c | 55 static void mlx4_free_icm_pages(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_pages() argument 59 if (chunk->nsg > 0) in mlx4_free_icm_pages() 60 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, in mlx4_free_icm_pages() 63 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_pages() 64 __free_pages(sg_page(&chunk->mem[i]), in mlx4_free_icm_pages() 65 get_order(chunk->mem[i].length)); in mlx4_free_icm_pages() 68 static void mlx4_free_icm_coherent(struct mlx4_dev *dev, struct mlx4_icm_chunk *chunk) in mlx4_free_icm_coherent() argument 72 for (i = 0; i < chunk->npages; ++i) in mlx4_free_icm_coherent() 74 chunk->mem[i].length, in mlx4_free_icm_coherent() 75 lowmem_page_address(sg_page(&chunk->mem[i])), in mlx4_free_icm_coherent() [all …]
|
D | icm.h | 64 struct mlx4_icm_chunk *chunk; member 90 iter->chunk = list_empty(&icm->chunk_list) ? in mlx4_icm_first() 98 return !iter->chunk; in mlx4_icm_last() 103 if (++iter->page_idx >= iter->chunk->nsg) { in mlx4_icm_next() 104 if (iter->chunk->list.next == &iter->icm->chunk_list) { in mlx4_icm_next() 105 iter->chunk = NULL; in mlx4_icm_next() 109 iter->chunk = list_entry(iter->chunk->list.next, in mlx4_icm_next() 117 return sg_dma_address(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_addr() 122 return sg_dma_len(&iter->chunk->mem[iter->page_idx]); in mlx4_icm_size()
|
/Linux-v4.19/lib/ |
D | genalloc.c | 39 static inline size_t chunk_size(const struct gen_pool_chunk *chunk) in chunk_size() argument 41 return chunk->end_addr - chunk->start_addr + 1; in chunk_size() 185 struct gen_pool_chunk *chunk; in gen_pool_add_virt() local 190 chunk = kzalloc_node(nbytes, GFP_KERNEL, nid); in gen_pool_add_virt() 191 if (unlikely(chunk == NULL)) in gen_pool_add_virt() 194 chunk->phys_addr = phys; in gen_pool_add_virt() 195 chunk->start_addr = virt; in gen_pool_add_virt() 196 chunk->end_addr = virt + size - 1; in gen_pool_add_virt() 197 atomic_long_set(&chunk->avail, size); in gen_pool_add_virt() 200 list_add_rcu(&chunk->next_chunk, &pool->chunks); in gen_pool_add_virt() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/i40iw/ |
D | i40iw_pble.c | 53 static void i40iw_free_vmalloc_mem(struct i40iw_hw *hw, struct i40iw_chunk *chunk); 63 struct i40iw_chunk *chunk; in i40iw_destroy_pble_pool() local 68 chunk = list_entry(clist, struct i40iw_chunk, list); in i40iw_destroy_pble_pool() 69 if (chunk->type == I40IW_VMALLOC) in i40iw_destroy_pble_pool() 70 i40iw_free_vmalloc_mem(dev->hw, chunk); in i40iw_destroy_pble_pool() 71 kfree(chunk); in i40iw_destroy_pble_pool() 138 struct i40iw_chunk *chunk = info->chunk; in add_sd_direct() local 151 chunk->type = I40IW_DMA_COHERENT; in add_sd_direct() 155 chunk->size = info->pages << I40IW_HMC_PAGED_BP_SHIFT; in add_sd_direct() 156 chunk->vaddr = ((u8 *)sd_entry->u.bp.addr.va + offset); in add_sd_direct() [all …]
|
/Linux-v4.19/kernel/ |
D | audit_tree.c | 111 static void free_chunk(struct audit_chunk *chunk) in free_chunk() argument 115 for (i = 0; i < chunk->count; i++) { in free_chunk() 116 if (chunk->owners[i].owner) in free_chunk() 117 put_tree(chunk->owners[i].owner); in free_chunk() 119 kfree(chunk); in free_chunk() 122 void audit_put_chunk(struct audit_chunk *chunk) in audit_put_chunk() argument 124 if (atomic_long_dec_and_test(&chunk->refs)) in audit_put_chunk() 125 free_chunk(chunk); in audit_put_chunk() 130 struct audit_chunk *chunk = container_of(rcu, struct audit_chunk, head); in __put_chunk() local 131 audit_put_chunk(chunk); in __put_chunk() [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlxsw/ |
D | spectrum_acl_tcam.c | 177 struct mlxsw_sp_acl_tcam_chunk *chunk; member 287 struct mlxsw_sp_acl_tcam_chunk *chunk; in mlxsw_sp_acl_tcam_region_prio() local 292 chunk = list_first_entry(®ion->chunk_list, typeof(*chunk), list); in mlxsw_sp_acl_tcam_region_prio() 293 return chunk->priority; in mlxsw_sp_acl_tcam_region_prio() 299 struct mlxsw_sp_acl_tcam_chunk *chunk; in mlxsw_sp_acl_tcam_region_max_prio() local 303 chunk = list_last_entry(®ion->chunk_list, typeof(*chunk), list); in mlxsw_sp_acl_tcam_region_max_prio() 304 return chunk->priority; in mlxsw_sp_acl_tcam_region_max_prio() 588 struct mlxsw_sp_acl_tcam_chunk *chunk) in mlxsw_sp_acl_tcam_chunk_assoc() argument 618 chunk->region = region; in mlxsw_sp_acl_tcam_chunk_assoc() 619 list_add_tail(&chunk->list, ®ion->chunk_list); in mlxsw_sp_acl_tcam_chunk_assoc() [all …]
|
/Linux-v4.19/include/net/sctp/ |
D | sm.h | 173 __u8 sctp_get_chunk_type(struct sctp_chunk *chunk); 191 const struct sctp_chunk *chunk, 194 const struct sctp_chunk *chunk); 196 const struct sctp_chunk *chunk); 199 const struct sctp_chunk *chunk); 212 const struct sctp_chunk *chunk); 214 const struct sctp_chunk *chunk); 217 const struct sctp_chunk *chunk); 218 int sctp_init_cause(struct sctp_chunk *chunk, __be16 cause, size_t paylen); 220 const struct sctp_chunk *chunk, [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_memfree.c | 64 static void mthca_free_icm_pages(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_pages() argument 68 if (chunk->nsg > 0) in mthca_free_icm_pages() 69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, in mthca_free_icm_pages() 72 for (i = 0; i < chunk->npages; ++i) in mthca_free_icm_pages() 73 __free_pages(sg_page(&chunk->mem[i]), in mthca_free_icm_pages() 74 get_order(chunk->mem[i].length)); in mthca_free_icm_pages() 77 static void mthca_free_icm_coherent(struct mthca_dev *dev, struct mthca_icm_chunk *chunk) in mthca_free_icm_coherent() argument 81 for (i = 0; i < chunk->npages; ++i) { in mthca_free_icm_coherent() 82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, in mthca_free_icm_coherent() 83 lowmem_page_address(sg_page(&chunk->mem[i])), in mthca_free_icm_coherent() [all …]
|
/Linux-v4.19/drivers/gpu/drm/qxl/ |
D | qxl_image.c | 38 struct qxl_drm_chunk *chunk; in qxl_allocate_chunk() local 41 chunk = kmalloc(sizeof(struct qxl_drm_chunk), GFP_KERNEL); in qxl_allocate_chunk() 42 if (!chunk) in qxl_allocate_chunk() 45 ret = qxl_alloc_bo_reserved(qdev, release, chunk_size, &chunk->bo); in qxl_allocate_chunk() 47 kfree(chunk); in qxl_allocate_chunk() 51 list_add_tail(&chunk->head, &image->chunk_list); in qxl_allocate_chunk() 88 struct qxl_drm_chunk *chunk, *tmp; in qxl_image_free_objects() local 90 list_for_each_entry_safe(chunk, tmp, &dimage->chunk_list, head) { in qxl_image_free_objects() 91 qxl_bo_unref(&chunk->bo); in qxl_image_free_objects() 92 kfree(chunk); in qxl_image_free_objects() [all …]
|
/Linux-v4.19/arch/mips/dec/prom/ |
D | console.c | 22 unsigned int chunk = sizeof(buf) - 1; in prom_console_write() local 25 if (chunk > c) in prom_console_write() 26 chunk = c; in prom_console_write() 27 memcpy(buf, s, chunk); in prom_console_write() 28 buf[chunk] = '\0'; in prom_console_write() 30 s += chunk; in prom_console_write() 31 c -= chunk; in prom_console_write()
|
/Linux-v4.19/sound/core/ |
D | sgbuf.c | 71 unsigned int i, pages, chunk, maxpages; in snd_malloc_sgbuf_pages() local 96 chunk = pages; in snd_malloc_sgbuf_pages() 98 if (chunk > maxpages) in snd_malloc_sgbuf_pages() 99 chunk = maxpages; in snd_malloc_sgbuf_pages() 100 chunk <<= PAGE_SHIFT; in snd_malloc_sgbuf_pages() 102 chunk, &tmpb) < 0) { in snd_malloc_sgbuf_pages() 110 chunk = tmpb.bytes >> PAGE_SHIFT; in snd_malloc_sgbuf_pages() 111 for (i = 0; i < chunk; i++) { in snd_malloc_sgbuf_pages() 115 table->addr |= chunk; /* mark head */ in snd_malloc_sgbuf_pages() 121 sgbuf->pages += chunk; in snd_malloc_sgbuf_pages() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/usnic/ |
D | usnic_vnic.c | 96 struct usnic_vnic_res_chunk *chunk; in usnic_vnic_dump() local 119 chunk = &vnic->chunks[i]; in usnic_vnic_dump() 120 for (j = 0; j < chunk->cnt; j++) { in usnic_vnic_dump() 121 res = chunk->res[j]; in usnic_vnic_dump() 274 void usnic_vnic_put_resources(struct usnic_vnic_res_chunk *chunk) in usnic_vnic_put_resources() argument 279 struct usnic_vnic *vnic = chunk->vnic; in usnic_vnic_put_resources() 281 if (chunk->cnt > 0) { in usnic_vnic_put_resources() 283 while ((i = --chunk->cnt) >= 0) { in usnic_vnic_put_resources() 284 res = chunk->res[i]; in usnic_vnic_put_resources() 285 chunk->res[i] = NULL; in usnic_vnic_put_resources() [all …]
|