| /Linux-v5.15/fs/ntfs3/ |
| D | attrib.c | 91 CLST svcn = le64_to_cpu(attr->nres.svcn); in attr_load_runs() local 96 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs() 99 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 104 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs() 105 vcn ? *vcn : svcn, Add2Ptr(attr, run_off), in attr_load_runs() 420 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size() local 469 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_set_size() 472 if (svcn <= vcn && vcn <= evcn) { in attr_set_size() 489 svcn = le64_to_cpu(attr->nres.svcn); in attr_set_size() 581 err = mi_pack_runs(mi, attr, run, vcn - svcn); in attr_set_size() [all …]
|
| D | frecord.c | 226 if (attr->nres.svcn) in ni_find_attr() 228 } else if (le64_to_cpu(attr->nres.svcn) > *vcn || in ni_find_attr() 325 if (le64_to_cpu(attr->nres.svcn) <= vcn && in ni_load_attr() 457 CLST svcn, struct ATTR_LIST_ENTRY **ins_le) in ni_ins_new_attr() argument 467 err = al_add_le(ni, type, name, name_len, svcn, cpu_to_le16(-1), in ni_ins_new_attr() 526 CLST svcn, evcn = 0, svcn_p, evcn_p, next_svcn; in ni_repack() local 536 svcn = le64_to_cpu(attr->nres.svcn); in ni_repack() 537 if (svcn != le64_to_cpu(le->vcn)) { in ni_repack() 542 if (!svcn) { in ni_repack() 546 } else if (svcn != evcn + 1) { in ni_repack() [all …]
|
| D | run.c | 167 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) in run_is_mapped_full() argument 173 if (!run_lookup(run, svcn, &i)) in run_is_mapped_full() 775 int run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, in run_pack() argument 780 CLST evcn1 = svcn + len; in run_pack() 787 next_vcn = vcn = svcn; in run_pack() 874 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack() argument 882 if (evcn + 1 == svcn) in run_unpack() 885 if (evcn < svcn) in run_unpack() 891 vcn64 = svcn; in run_unpack() 1001 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack_ex() argument [all …]
|
| D | attrlist.c | 61 } else if (attr->nres.svcn) { in ntfs_load_attr_list() 156 CLST svcn = attr_svcn(attr); in al_find_le() local 159 &svcn); in al_find_le() 271 u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref, in al_add_le() argument 293 le = al_find_le_to_insert(ni, type, name, name_len, svcn); in al_add_le() 318 le->vcn = cpu_to_le64(svcn); in al_add_le()
|
| D | record.c | 273 if (attr->nres.svcn || !is_attr_ext(attr)) { in mi_enum_attr() 571 CLST svcn = le64_to_cpu(attr->nres.svcn); in mi_pack_runs() local 585 err = run_pack(run, svcn, len, Add2Ptr(attr, run_off), run_size + dsize, in mi_pack_runs() 597 attr->nres.evcn = cpu_to_le64(svcn + plen - 1); in mi_pack_runs()
|
| D | ntfs_fs.h | 460 u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref, 530 const struct runs_tree *run, CLST svcn, CLST len, 785 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn); 787 int run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, 790 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, 795 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf,
|
| D | fslog.c | 2698 u64 dsize, svcn, evcn; in check_attr() local 2722 svcn = le64_to_cpu(attr->nres.svcn); in check_attr() 2726 if (svcn > evcn + 1 || run_off >= asize || in check_attr() 2732 if (run_unpack(NULL, sbi, 0, svcn, evcn, svcn, in check_attr() 3356 if (run_get_highest_vcn(le64_to_cpu(attr->nres.svcn), in do_action() 4768 CLST svcn = le64_to_cpu(attr->nres.svcn); in log_replay() local 4770 err = run_unpack(&oa->run0, sbi, inode->i_ino, svcn, in log_replay() 4771 le64_to_cpu(attr->nres.evcn), svcn, in log_replay()
|
| D | ntfs.h | 332 __le64 svcn; // 0x10: Starting VCN of this segment. member 436 return attr->non_res ? le64_to_cpu(attr->nres.svcn) : 0; in attr_svcn()
|
| D | inode.c | 367 t64 = le64_to_cpu(attr->nres.svcn); in ntfs_read_mft()
|