| /Linux-v6.1/fs/ntfs3/ |
| D | attrib.c | 92 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs() local 96 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs() 99 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 104 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs() 420 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size() local 474 evcn = le64_to_cpu(attr_b->nres.evcn); in attr_set_size() 476 if (svcn <= vcn && vcn <= evcn) { in attr_set_size() 494 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size() 601 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_set_size() 622 evcn = old_alen - 1; in attr_set_size() [all …]
|
| D | frecord.c | 230 *vcn > le64_to_cpu(attr->nres.evcn)) { in ni_find_attr() 327 vcn <= le64_to_cpu(attr->nres.evcn)) in ni_load_attr() 527 CLST svcn, evcn = 0, svcn_p, evcn_p, next_svcn; in ni_repack() local 547 } else if (svcn != evcn + 1) { in ni_repack() 552 evcn = le64_to_cpu(attr->nres.evcn); in ni_repack() 554 if (svcn > evcn + 1) { in ni_repack() 565 if (evcn + 1 == alloc) in ni_repack() 571 err = run_unpack(&run, sbi, ni->mi.rno, svcn, evcn, svcn, in ni_repack() 581 evcn_p = evcn; in ni_repack() 591 err = mi_pack_runs(mi_p, attr_p, &run, evcn + 1 - svcn_p); in ni_repack() [all …]
|
| D | run.c | 167 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) in run_is_mapped_full() argument 181 if (next_vcn > evcn) in run_is_mapped_full() 921 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack() argument 929 if (evcn + 1 == svcn) in run_unpack() 932 if (evcn < svcn) in run_unpack() 989 if (next_vcn > evcn + 1) in run_unpack() 1031 if (vcn64 != evcn + 1) { in run_unpack() 1048 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack_ex() argument 1057 ret = run_unpack(run, sbi, ino, svcn, evcn, vcn, run_buf, run_buf_size); in run_unpack_ex() 1071 next_vcn <= evcn; in run_unpack_ex()
|
| D | fslog.c | 2696 u64 dsize, svcn, evcn; in check_attr() local 2721 evcn = le64_to_cpu(attr->nres.evcn); in check_attr() 2724 if (svcn > evcn + 1 || run_off >= asize || in check_attr() 2730 if (run_unpack(NULL, sbi, 0, svcn, evcn, svcn, in check_attr() 3005 attr->nres.evcn = cpu_to_le64((u64)bytes_to_cluster(sbi, size) - 1); in attr_create_nonres_log() 3359 attr->nres.evcn = cpu_to_le64(t64); in do_action() 3362 oa2->attr->nres.evcn = attr->nres.evcn; in do_action() 4775 le64_to_cpu(attr->nres.evcn), svcn, in log_replay()
|
| D | ntfs_fs.h | 790 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn); 795 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, 800 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf,
|
| D | attrlist.c | 72 0, le64_to_cpu(attr->nres.evcn), 0, in ntfs_load_attr_list()
|
| D | inode.c | 368 err = run_unpack_ex(run, sbi, ino, t64, le64_to_cpu(attr->nres.evcn), in ntfs_read_mft() 1425 attr->nres.evcn = cpu_to_le64(-1ll); in ntfs_create_inode() 1496 attr->nres.evcn = cpu_to_le64(clst - 1); in ntfs_create_inode()
|
| D | record.c | 574 attr->nres.evcn = cpu_to_le64(svcn + plen - 1); in mi_pack_runs()
|
| D | ntfs.h | 333 __le64 evcn; // 0x18: End VCN of this segment. member
|