| /linux/fs/ntfs3/ |
| H A D | attrib.c | 65 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs() local 69 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs() 72 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 81 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs() 445 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size_ex() local 539 evcn = le64_to_cpu(attr_b->nres.evcn); in attr_set_size_ex() 541 if (svcn <= vcn && vcn <= evcn) { in attr_set_size_ex() 559 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size_ex() 666 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_set_size_ex() 687 evcn = old_alen - 1; in attr_set_size_ex() [all …]
|
| H A D | frecord.c | 237 *vcn > le64_to_cpu(attr->nres.evcn)) { in ni_find_attr() 485 CLST svcn, evcn = 0, svcn_p, evcn_p, next_svcn; in ni_repack() 505 } else if (svcn != evcn + 1) { in ni_repack() 510 evcn = le64_to_cpu(attr->nres.evcn); in ni_repack() 512 if (svcn > evcn + 1) { in ni_repack() 523 if (evcn + 1 == alloc) in ni_repack() 535 err = run_unpack(&run, sbi, ni->mi.rno, svcn, evcn, svcn, in ni_repack() 545 evcn_p = evcn; in ni_repack() 555 err = mi_pack_runs(mi_p, attr_p, &run, evcn + 1 - svcn_p); in ni_repack() 559 next_svcn = le64_to_cpu(attr_p->nres.evcn) + 1; in ni_repack() [all …]
|
| H A D | run.c | 168 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) in run_is_mapped_full() argument 182 if (next_vcn > evcn) in run_is_mapped_full() 967 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack() argument 978 if (evcn + 1 == svcn) in run_unpack() 981 if (evcn < svcn) in run_unpack() 1048 if (next_vcn > evcn + 1) in run_unpack() 1090 if (vcn64 != evcn + 1) { in run_unpack() 1107 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, in run_unpack_ex() argument 1116 ret = run_unpack(run, sbi, ino, svcn, evcn, vcn, run_buf, run_buf_size); in run_unpack_ex() 1130 next_vcn <= evcn; in run_unpack_ex()
|
| H A D | record.c | 314 if (le64_to_cpu(attr->nres.svcn) > le64_to_cpu(attr->nres.evcn) + 1) in mi_enum_attr() 657 attr->nres.evcn = cpu_to_le64(svcn + plen - 1); in mi_pack_runs()
|
| H A D | fslog.c | 2697 u64 dsize, svcn, evcn; in check_attr() local 2722 evcn = le64_to_cpu(attr->nres.evcn); in check_attr() 2725 if (svcn > evcn + 1 || run_off >= asize || in check_attr() 2734 if (run_unpack(NULL, sbi, 0, svcn, evcn, svcn, in check_attr() 3009 attr->nres.evcn = cpu_to_le64((u64)bytes_to_cluster(sbi, size) - 1); in attr_create_nonres_log() 3365 attr->nres.evcn = cpu_to_le64(t64); in do_action() 3368 oa2->attr->nres.evcn = attr->nres.evcn; in do_action() 4796 le64_to_cpu(attr->nres.evcn), svcn, in log_replay()
|
| H A D | attrlist.c | 87 0, le64_to_cpu(attr->nres.evcn), 0, in ntfs_load_attr_list()
|
| H A D | ntfs_fs.h | 870 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn); 875 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, 880 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf,
|
| H A D | inode.c | 400 err = run_unpack_ex(run, sbi, ino, t64, le64_to_cpu(attr->nres.evcn), in ntfs_read_mft() 1518 attr->nres.evcn = cpu_to_le64(-1ll); in ntfs_create_inode() 1589 attr->nres.evcn = cpu_to_le64(clst - 1); in ntfs_create_inode()
|
| H A D | ntfs.h | 340 __le64 evcn; // 0x18: End VCN of this segment. member
|