Lines Matching refs:evcn

65 	CLST evcn = le64_to_cpu(attr->nres.evcn);  in attr_load_runs()  local
69 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs()
72 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs()
81 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs()
414 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size() local
471 evcn = le64_to_cpu(attr_b->nres.evcn); in attr_set_size()
473 if (svcn <= vcn && vcn <= evcn) { in attr_set_size()
491 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
597 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_set_size()
618 evcn = old_alen - 1; in attr_set_size()
674 run_truncate_head(run, evcn + 1); in attr_set_size()
677 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
727 attr->nres.evcn = cpu_to_le64((u64)vcn - 1); in attr_set_size()
750 err = run_deallocate_ex(sbi, run, vcn, evcn - vcn + 1, &dlen, in attr_set_size()
772 evcn = svcn - 1; in attr_set_size()
840 svcn <= le64_to_cpu(attr_b->nres.evcn)) { in attr_set_size()
856 if (mi_pack_runs(mi, attr, run, evcn - svcn + 1)) in attr_set_size()
893 CLST alloc, evcn; in attr_data_get_block() local
946 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_data_get_block()
960 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1015 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1081 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1135 evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1139 while (end > evcn) { in attr_data_get_block()
1148 if (evcn + 1 >= alloc) { in attr_data_get_block()
1150 evcn1 = evcn + 1; in attr_data_get_block()
1165 evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1175 evcn1 = evcn + 1; in attr_data_get_block()
1184 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1298 CLST svcn, evcn; in attr_load_runs_vcn() local
1313 evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs_vcn()
1315 if (evcn < vcn || vcn < svcn) { in attr_load_runs_vcn()
1325 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, svcn, in attr_load_runs_vcn()
1639 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_allocate_frame()
1658 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1723 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1757 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame() local
1761 while (end > evcn) { in attr_allocate_frame()
1770 if (evcn + 1 >= alloc) { in attr_allocate_frame()
1772 evcn1 = evcn + 1; in attr_allocate_frame()
1788 evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame()
1798 evcn1 = evcn + 1; in attr_allocate_frame()
1808 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1909 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_collapse_range()
1928 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
1935 attr->nres.evcn = cpu_to_le64(evcn1 - 1 - len); in attr_collapse_range()
1974 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
2046 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
2171 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_punch_hole()
2191 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2231 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2263 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2405 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_insert_range()
2424 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2442 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2447 le64_add_cpu(&attr->nres.evcn, len); in attr_insert_range()
2508 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_insert_range()
2525 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_insert_range()
2540 le64_sub_cpu(&attr->nres.evcn, len); in attr_insert_range()