Home
last modified time | relevance | path

Searched refs:vcn (Results 1 – 25 of 27) sorted by relevance

12

/Linux-v5.4/fs/ntfs/
Drunlist.c151 if ((dst->vcn + dst->length) != src->vcn) in ntfs_are_rl_mergeable()
238 dst[loc].length = dst[loc + 1].vcn - dst[loc].vcn; in ntfs_rl_append()
242 dst[marker].vcn = dst[marker - 1].vcn + dst[marker - 1].length; in ntfs_rl_append()
286 disc = (src[0].vcn > 0); in ntfs_rl_insert()
296 disc = (src[0].vcn > dst[loc - 1].vcn + merged_length); in ntfs_rl_insert()
325 dst[marker].vcn = dst[marker - 1].vcn + dst[marker - 1].length; in ntfs_rl_insert()
328 dst[marker].length = dst[marker + 1].vcn - dst[marker].vcn; in ntfs_rl_insert()
333 dst[loc].vcn = dst[loc - 1].vcn + dst[loc - 1].length; in ntfs_rl_insert()
334 dst[loc].length = dst[loc + 1].vcn - dst[loc].vcn; in ntfs_rl_insert()
336 dst[loc].vcn = 0; in ntfs_rl_insert()
[all …]
Dindex.c108 VCN vcn, old_vcn; in ntfs_index_lookup() local
248 vcn = sle64_to_cpup((sle64*)((u8*)ie + le16_to_cpu(ie->length) - 8)); in ntfs_index_lookup()
264 page = ntfs_map_page(ia_mapping, vcn << in ntfs_index_lookup()
276 ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << in ntfs_index_lookup()
288 (long long)vcn, idx_ni->mft_no); in ntfs_index_lookup()
291 if (sle64_to_cpu(ia->index_block_vcn) != vcn) { in ntfs_index_lookup()
297 (unsigned long long)vcn, idx_ni->mft_no); in ntfs_index_lookup()
305 "driver bug.", (unsigned long long)vcn, in ntfs_index_lookup()
316 "driver.", (unsigned long long)vcn, in ntfs_index_lookup()
324 (unsigned long long)vcn, idx_ni->mft_no); in ntfs_index_lookup()
[all …]
Dlogfile.c715 VCN vcn, end_vcn; in ntfs_empty_logfile() local
739 vcn = 0; in ntfs_empty_logfile()
747 if (unlikely(!rl || vcn < rl->vcn || !rl->length)) { in ntfs_empty_logfile()
749 err = ntfs_map_runlist_nolock(log_ni, vcn, NULL); in ntfs_empty_logfile()
756 BUG_ON(!rl || vcn < rl->vcn || !rl->length); in ntfs_empty_logfile()
759 while (rl->length && vcn >= rl[1].vcn) in ntfs_empty_logfile()
772 vcn = rl->vcn; in ntfs_empty_logfile()
783 if (rl[1].vcn > end_vcn) in ntfs_empty_logfile()
784 len = end_vcn - rl->vcn; in ntfs_empty_logfile()
820 } while ((++rl)->vcn < end_vcn); in ntfs_empty_logfile()
Ddir.c79 VCN vcn, old_vcn; in ntfs_lookup_inode_by_name() local
291 vcn = sle64_to_cpup((sle64*)((u8*)ie + le16_to_cpu(ie->length) - 8)); in ntfs_lookup_inode_by_name()
307 page = ntfs_map_page(ia_mapping, vcn << in ntfs_lookup_inode_by_name()
319 ia = (INDEX_ALLOCATION*)(kaddr + ((vcn << in ntfs_lookup_inode_by_name()
331 (unsigned long long)vcn, dir_ni->mft_no); in ntfs_lookup_inode_by_name()
334 if (sle64_to_cpu(ia->index_block_vcn) != vcn) { in ntfs_lookup_inode_by_name()
340 (unsigned long long)vcn, dir_ni->mft_no); in ntfs_lookup_inode_by_name()
349 (unsigned long long)vcn, dir_ni->mft_no, in ntfs_lookup_inode_by_name()
359 "driver.", (unsigned long long)vcn, in ntfs_lookup_inode_by_name()
367 (unsigned long long)vcn, dir_ni->mft_no); in ntfs_lookup_inode_by_name()
[all …]
Dattrib.h49 extern int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn,
51 extern int ntfs_map_runlist(ntfs_inode *ni, VCN vcn);
53 extern LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn,
57 const VCN vcn, ntfs_attr_search_ctx *ctx);
Daops.c170 VCN vcn; in ntfs_read_block() local
245 vcn = (VCN)iblock << blocksize_bits >> in ntfs_read_block()
256 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_block()
258 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block()
286 err = ntfs_map_runlist(ni, vcn); in ntfs_read_block()
311 ni->type, (unsigned long long)vcn, in ntfs_read_block()
535 VCN vcn; in ntfs_write_block() local
702 vcn = (VCN)block << blocksize_bits; in ntfs_write_block()
703 vcn_ofs = vcn & vol->cluster_size_mask; in ntfs_write_block()
704 vcn >>= vol->cluster_size_bits; in ntfs_write_block()
[all …]
Drunlist.h29 VCN vcn; /* vcn = Starting virtual cluster number. */ member
65 extern LCN ntfs_rl_vcn_to_lcn(const runlist_element *rl, const VCN vcn);
70 const VCN vcn);
Dattrib.c70 int ntfs_map_runlist_nolock(ntfs_inode *ni, VCN vcn, ntfs_attr_search_ctx *ctx) in ntfs_map_runlist_nolock() argument
84 (unsigned long long)vcn); in ntfs_map_runlist_nolock()
120 if (vcn >= allocated_size_vcn || (a->type == ni->type && in ntfs_map_runlist_nolock()
125 <= vcn && end_vcn >= vcn)) in ntfs_map_runlist_nolock()
153 CASE_SENSITIVE, vcn, NULL, 0, ctx); in ntfs_map_runlist_nolock()
169 if (unlikely(vcn && vcn >= end_vcn)) { in ntfs_map_runlist_nolock()
284 int ntfs_map_runlist(ntfs_inode *ni, VCN vcn) in ntfs_map_runlist() argument
290 if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <= in ntfs_map_runlist()
292 err = ntfs_map_runlist_nolock(ni, vcn, NULL); in ntfs_map_runlist()
327 LCN ntfs_attr_vcn_to_lcn_nolock(ntfs_inode *ni, const VCN vcn, in ntfs_attr_vcn_to_lcn_nolock() argument
[all …]
Dcompress.c478 VCN vcn; in ntfs_read_compressed_block() local
596 for (vcn = start_vcn, start_vcn += cb_clusters; vcn < start_vcn; in ntfs_read_compressed_block()
597 vcn++) { in ntfs_read_compressed_block()
607 while (rl->length && rl[1].vcn <= vcn) in ntfs_read_compressed_block()
609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block()
613 (unsigned long long)vcn, in ntfs_read_compressed_block()
630 if (!ntfs_map_runlist(ni, vcn)) in ntfs_read_compressed_block()
730 if (vcn == start_vcn - cb_clusters) { in ntfs_read_compressed_block()
773 } else if (vcn == start_vcn) { in ntfs_read_compressed_block()
Dlcnalloc.c388 rl[rlpos].vcn = rl[rlpos - 1].vcn + in ntfs_cluster_alloc()
393 rl[rlpos].vcn = start_vcn; in ntfs_cluster_alloc()
726 rl[rlpos].vcn = rl[rlpos - 1].vcn + rl[rlpos - 1].length; in ntfs_cluster_alloc()
883 delta = start_vcn - rl->vcn; in __ntfs_cluster_free()
916 VCN vcn; in __ntfs_cluster_free() local
919 vcn = rl->vcn; in __ntfs_cluster_free()
920 rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx); in __ntfs_cluster_free()
Dmft.c525 VCN vcn; in ntfs_sync_mft_mirror() local
531 vcn = ((VCN)mft_no << vol->mft_record_size_bits) + in ntfs_sync_mft_mirror()
533 vcn_ofs = vcn & vol->cluster_size_mask; in ntfs_sync_mft_mirror()
534 vcn >>= vol->cluster_size_bits; in ntfs_sync_mft_mirror()
546 while (rl->length && rl[1].vcn <= vcn) in ntfs_sync_mft_mirror()
548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror()
718 VCN vcn; in write_mft_record_nolock() local
724 vcn = ((VCN)ni->mft_no << vol->mft_record_size_bits) + in write_mft_record_nolock()
726 vcn_ofs = vcn & vol->cluster_size_mask; in write_mft_record_nolock()
727 vcn >>= vol->cluster_size_bits; in write_mft_record_nolock()
[all …]
Dfile.c573 VCN vcn, highest_vcn = 0, cpos, cend, bh_cpos, bh_cend; in ntfs_prepare_pages_for_non_resident_write() local
628 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write()
706 cdelta = bh_cpos - vcn; in ntfs_prepare_pages_for_non_resident_write()
836 while (rl->length && rl[1].vcn <= bh_cpos) in ntfs_prepare_pages_for_non_resident_write()
845 vcn = bh_cpos; in ntfs_prepare_pages_for_non_resident_write()
846 vcn_len = rl[1].vcn - vcn; in ntfs_prepare_pages_for_non_resident_write()
857 if (likely(vcn + vcn_len >= cend)) { in ntfs_prepare_pages_for_non_resident_write()
1042 vcn = sle64_to_cpu(a->data.non_resident.lowest_vcn); in ntfs_prepare_pages_for_non_resident_write()
1043 rl2 = ntfs_rl_find_vcn_nolock(rl, vcn); in ntfs_prepare_pages_for_non_resident_write()
1060 mp_size = ntfs_get_size_for_mapping_pairs(vol, rl2, vcn, in ntfs_prepare_pages_for_non_resident_write()
[all …]
Ddebug.c143 (long long)(rl + i)->vcn, lcn_str[index], in ntfs_debug_dump_runlist()
149 (long long)(rl + i)->vcn, in ntfs_debug_dump_runlist()
Dsuper.c1166 rl2[0].vcn = 0; in check_mft_mirror()
1170 rl2[1].vcn = rl2[0].length; in check_mft_mirror()
1183 if (rl2[i].vcn != rl[i].vcn || rl2[i].lcn != rl[i].lcn || in check_mft_mirror()
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_vcn.c74 INIT_DELAYED_WORK(&adev->vcn.idle_work, amdgpu_vcn_idle_work_handler); in amdgpu_vcn_sw_init()
92 adev->vcn.indirect_sram = true; in amdgpu_vcn_sw_init()
98 adev->vcn.indirect_sram = true; in amdgpu_vcn_sw_init()
104 adev->vcn.indirect_sram = true; in amdgpu_vcn_sw_init()
110 adev->vcn.indirect_sram = true; in amdgpu_vcn_sw_init()
116 r = request_firmware(&adev->vcn.fw, fw_name, adev->dev); in amdgpu_vcn_sw_init()
123 r = amdgpu_ucode_validate(adev->vcn.fw); in amdgpu_vcn_sw_init()
127 release_firmware(adev->vcn.fw); in amdgpu_vcn_sw_init()
128 adev->vcn.fw = NULL; in amdgpu_vcn_sw_init()
132 hdr = (const struct common_firmware_header *)adev->vcn.fw->data; in amdgpu_vcn_sw_init()
[all …]
Dvcn_v2_5.c79 adev->vcn.num_vcn_inst = VCN25_MAX_HW_INSTANCES_ARCTURUS; in vcn_v2_5_early_init()
80 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in vcn_v2_5_early_init()
83 adev->vcn.harvest_config |= 1 << i; in vcn_v2_5_early_init()
86 if (adev->vcn.harvest_config == (AMDGPU_VCN_HARVEST_VCN0 | in vcn_v2_5_early_init()
91 adev->vcn.num_vcn_inst = 1; in vcn_v2_5_early_init()
93 adev->vcn.num_enc_rings = 2; in vcn_v2_5_early_init()
116 for (j = 0; j < adev->vcn.num_vcn_inst; j++) { in vcn_v2_5_sw_init()
117 if (adev->vcn.harvest_config & (1 << j)) in vcn_v2_5_sw_init()
121 VCN_2_0__SRCID__UVD_SYSTEM_MESSAGE_INTERRUPT, &adev->vcn.inst[j].irq); in vcn_v2_5_sw_init()
126 for (i = 0; i < adev->vcn.num_enc_rings; ++i) { in vcn_v2_5_sw_init()
[all …]
Dvcn_v2_0.c95 adev->vcn.num_vcn_inst = 1; in vcn_v2_0_early_init()
96 adev->vcn.num_enc_rings = 2; in vcn_v2_0_early_init()
122 &adev->vcn.inst->irq); in vcn_v2_0_sw_init()
127 for (i = 0; i < adev->vcn.num_enc_rings; ++i) { in vcn_v2_0_sw_init()
130 &adev->vcn.inst->irq); in vcn_v2_0_sw_init()
137 VCN_2_0__SRCID__JPEG_DECODE, &adev->vcn.inst->irq); in vcn_v2_0_sw_init()
147 hdr = (const struct common_firmware_header *)adev->vcn.fw->data; in vcn_v2_0_sw_init()
149 adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].fw = adev->vcn.fw; in vcn_v2_0_sw_init()
159 ring = &adev->vcn.inst->ring_dec; in vcn_v2_0_sw_init()
162 ring->doorbell_index = adev->doorbell_index.vcn.vcn_ring0_1 << 1; in vcn_v2_0_sw_init()
[all …]
Dvcn_v1_0.c66 adev->vcn.num_vcn_inst = 1; in vcn_v1_0_early_init()
67 adev->vcn.num_enc_rings = 2; in vcn_v1_0_early_init()
92 VCN_1_0__SRCID__UVD_SYSTEM_MESSAGE_INTERRUPT, &adev->vcn.inst->irq); in vcn_v1_0_sw_init()
97 for (i = 0; i < adev->vcn.num_enc_rings; ++i) { in vcn_v1_0_sw_init()
99 &adev->vcn.inst->irq); in vcn_v1_0_sw_init()
105 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_VCN, 126, &adev->vcn.inst->irq); in vcn_v1_0_sw_init()
115 hdr = (const struct common_firmware_header *)adev->vcn.fw->data; in vcn_v1_0_sw_init()
117 adev->firmware.ucode[AMDGPU_UCODE_ID_VCN].fw = adev->vcn.fw; in vcn_v1_0_sw_init()
127 ring = &adev->vcn.inst->ring_dec; in vcn_v1_0_sw_init()
129 r = amdgpu_ring_init(adev, ring, 512, &adev->vcn.inst->irq, 0); in vcn_v1_0_sw_init()
[all …]
Dvega10_reg_init.c84 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_DOORBELL64_VCN0_1; in vega10_doorbell_index_init()
85 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_DOORBELL64_VCN2_3; in vega10_doorbell_index_init()
86 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_DOORBELL64_VCN4_5; in vega10_doorbell_index_init()
87 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_DOORBELL64_VCN6_7; in vega10_doorbell_index_init()
Damdgpu_ctx.c157 for (j = 0; j < adev->vcn.num_vcn_inst; ++j) { in amdgpu_ctx_init()
158 if (adev->vcn.harvest_config & (1 << j)) in amdgpu_ctx_init()
160 rings[num_rings++] = &adev->vcn.inst[j].ring_dec; in amdgpu_ctx_init()
164 for (j = 0; j < adev->vcn.num_vcn_inst; ++j) { in amdgpu_ctx_init()
165 if (adev->vcn.harvest_config & (1 << j)) in amdgpu_ctx_init()
167 for (k = 0; k < adev->vcn.num_enc_rings; ++k) in amdgpu_ctx_init()
168 rings[num_rings++] = &adev->vcn.inst[j].ring_enc[k]; in amdgpu_ctx_init()
172 for (j = 0; j < adev->vcn.num_vcn_inst; ++j) { in amdgpu_ctx_init()
173 if (adev->vcn.harvest_config & (1 << j)) in amdgpu_ctx_init()
175 rings[num_rings++] = &adev->vcn.inst[j].ring_jpeg; in amdgpu_ctx_init()
Dvega20_reg_init.c90 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_VEGA20_DOORBELL64_VCN0_1; in vega20_doorbell_index_init()
91 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_VEGA20_DOORBELL64_VCN2_3; in vega20_doorbell_index_init()
92 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_VEGA20_DOORBELL64_VCN4_5; in vega20_doorbell_index_init()
93 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_VEGA20_DOORBELL64_VCN6_7; in vega20_doorbell_index_init()
Damdgpu_vcn.h121 *adev->vcn.dpg_sram_curr_addr++ = offset; \
122 *adev->vcn.dpg_sram_curr_addr++ = value; \
Damdgpu_kms.c218 fw_info->ver = adev->vcn.fw_version; in amdgpu_firmware_info()
375 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info()
379 if (adev->vcn.inst[i].ring_dec.sched.ready) in amdgpu_hw_ip_info()
387 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info()
391 for (j = 0; j < adev->vcn.num_enc_rings; j++) in amdgpu_hw_ip_info()
392 if (adev->vcn.inst[i].ring_enc[j].sched.ready) in amdgpu_hw_ip_info()
400 for (i = 0; i < adev->vcn.num_vcn_inst; i++) { in amdgpu_hw_ip_info()
404 if (adev->vcn.inst[i].ring_jpeg.sched.ready) in amdgpu_hw_ip_info()
Dnv.c553 adev->doorbell_index.vcn.vcn_ring0_1 = AMDGPU_NAVI10_DOORBELL64_VCN0_1; in nv_init_doorbell_index()
554 adev->doorbell_index.vcn.vcn_ring2_3 = AMDGPU_NAVI10_DOORBELL64_VCN2_3; in nv_init_doorbell_index()
555 adev->doorbell_index.vcn.vcn_ring4_5 = AMDGPU_NAVI10_DOORBELL64_VCN4_5; in nv_init_doorbell_index()
556 adev->doorbell_index.vcn.vcn_ring6_7 = AMDGPU_NAVI10_DOORBELL64_VCN6_7; in nv_init_doorbell_index()
Damdgpu_doorbell.h63 } vcn; member

12