Home
last modified time | relevance | path

Searched refs:num_chunks (Results 1 – 14 of 14) sorted by relevance

/Linux-v5.4/drivers/misc/habanalabs/
Dcommand_submission.c466 u32 num_chunks, u64 *cs_seq) in _hl_cs_ioctl() argument
479 if (num_chunks > HL_MAX_JOBS_PER_CS) { in _hl_cs_ioctl()
487 cs_chunk_array = kmalloc_array(num_chunks, sizeof(*cs_chunk_array), in _hl_cs_ioctl()
494 size_to_copy = num_chunks * sizeof(struct hl_cs_chunk); in _hl_cs_ioctl()
515 for (i = 0, parse_cnt = 0 ; i < num_chunks ; i++, parse_cnt++) { in _hl_cs_ioctl()
616 u32 num_chunks; in hl_cs_ioctl() local
635 num_chunks = args->in.num_chunks_restore; in hl_cs_ioctl()
663 if (num_chunks == 0) { in hl_cs_ioctl()
668 rc = _hl_cs_ioctl(hpriv, chunks, num_chunks, in hl_cs_ioctl()
682 if (num_chunks > 0) { in hl_cs_ioctl()
[all …]
/Linux-v5.4/drivers/crypto/qat/qat_common/
Dicp_qat_uclo.h255 unsigned short num_chunks; member
270 short num_chunks; member
510 unsigned short num_chunks; member
Dqat_uclo.c155 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format()
484 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk()
531 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk()
1162 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof()
/Linux-v5.4/drivers/infiniband/sw/siw/
Dsiw_mem.c375 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local
385 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get()
411 kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); in siw_umem_get()
/Linux-v5.4/drivers/gpu/drm/radeon/
Dradeon_cs.c280 if (!cs->num_chunks) { in radeon_cs_parser_init()
292 p->chunks_array = kcalloc(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); in radeon_cs_parser_init()
298 sizeof(uint64_t)*cs->num_chunks)) { in radeon_cs_parser_init()
302 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
/Linux-v5.4/drivers/mtd/nand/raw/
Dmxc_nand.c255 u16 num_chunks = mtd->writesize / 512; in copy_spare() local
262 oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; in copy_spare()
265 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
275 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_cs.c116 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init()
119 chunk_array = kmalloc_array(cs->in.num_chunks, sizeof(uint64_t), GFP_KERNEL); in amdgpu_cs_parser_init()
140 sizeof(uint64_t)*cs->in.num_chunks)) { in amdgpu_cs_parser_init()
145 p->nchunks = cs->in.num_chunks; in amdgpu_cs_parser_init()
/Linux-v5.4/include/uapi/drm/
Damdgpu_drm.h554 __u32 num_chunks; member
Dradeon_drm.h984 __u32 num_chunks; member
/Linux-v5.4/net/sctp/
Dsocket.c6974 u32 num_chunks = 0; in sctp_getsockopt_peer_auth_chunks() local
6996 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_peer_auth_chunks()
6997 if (len < num_chunks) in sctp_getsockopt_peer_auth_chunks()
7000 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_peer_auth_chunks()
7003 len = sizeof(struct sctp_authchunks) + num_chunks; in sctp_getsockopt_peer_auth_chunks()
7006 if (put_user(num_chunks, &p->gauth_number_of_chunks)) in sctp_getsockopt_peer_auth_chunks()
7019 u32 num_chunks = 0; in sctp_getsockopt_local_auth_chunks() local
7046 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_local_auth_chunks()
7047 if (len < sizeof(struct sctp_authchunks) + num_chunks) in sctp_getsockopt_local_auth_chunks()
7050 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_local_auth_chunks()
[all …]
/Linux-v5.4/drivers/net/wireless/intel/ipw2x00/
Dipw2200.c3811 if (le32_to_cpu(bd->u.data.num_chunks) > NUM_TFD_CHUNKS) { in ipw_queue_tx_free_tfd()
3813 le32_to_cpu(bd->u.data.num_chunks)); in ipw_queue_tx_free_tfd()
3819 for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) { in ipw_queue_tx_free_tfd()
10212 tfd->u.data.num_chunks = cpu_to_le32(min((u8) (NUM_TFD_CHUNKS - 2), in ipw_tx_skb()
10215 txb->nr_frags, le32_to_cpu(tfd->u.data.num_chunks)); in ipw_tx_skb()
10216 for (i = 0; i < le32_to_cpu(tfd->u.data.num_chunks); i++) { in ipw_tx_skb()
10218 i, le32_to_cpu(tfd->u.data.num_chunks), in ipw_tx_skb()
10221 i, tfd->u.data.num_chunks, in ipw_tx_skb()
10266 le32_add_cpu(&tfd->u.data.num_chunks, 1); in ipw_tx_skb()
Dipw2200.h488 __le32 num_chunks; member
/Linux-v5.4/fs/btrfs/
Dinode.c377 atomic_t num_chunks; member
1217 u64 num_chunks = DIV_ROUND_UP(end - start, SZ_512K); in cow_file_range_async() local
1226 num_chunks = 1; in cow_file_range_async()
1233 ctx = kvmalloc(struct_size(ctx, chunks, num_chunks), GFP_KERNEL); in cow_file_range_async()
1250 atomic_set(&ctx->num_chunks, num_chunks); in cow_file_range_async()
1252 for (i = 0; i < num_chunks; i++) { in cow_file_range_async()
1263 async_chunk[i].pending = &ctx->num_chunks; in cow_file_range_async()
/Linux-v5.4/drivers/staging/media/ipu3/
Dipu3-abi.h1559 u32 num_chunks; member