Re: [PATCH v21 05/10] fs/ntfs3: Add attrib operations
From: Mark Harmstone
Date: Fri Feb 12 2021 - 13:14:33 EST
Hi Konstantin,
It looks like you've renamed the xattrs again since I last checked this - is
there a reason for this?
As I said in my previous messages, it makes sense for these to be the same
as the de facto standard (user.DOSATTRIB and security.NTACL) used by Wine,
Samba, and my Windows btrfs driver. See also XATTR_DOS_ATTRIB in
fs/cifs/cifsglob.h, though it appears to be orphaned.
FWIW, there's a lively discussion from 2005 about this very issue on
linux-ntfs-dev, I don't know if you've seen it:
https://sourceforge.net/p/linux-ntfs/mailman/linux-ntfs-dev/?viewmonth=200501&viewday=4
Mark
On 12/2/21 4:24 pm, Konstantin Komarov wrote:
> This adds attrib operations
>
> Signed-off-by: Konstantin Komarov <almaz.alexandrovich@xxxxxxxxxxxxxxxxxxxx>
> ---
> fs/ntfs3/attrib.c | 2085 +++++++++++++++++++++++++++++++++++++++++++
> fs/ntfs3/attrlist.c | 457 ++++++++++
> fs/ntfs3/xattr.c | 1050 ++++++++++++++++++++++
> 3 files changed, 3592 insertions(+)
> create mode 100644 fs/ntfs3/attrib.c
> create mode 100644 fs/ntfs3/attrlist.c
> create mode 100644 fs/ntfs3/xattr.c
>
> diff --git a/fs/ntfs3/attrib.c b/fs/ntfs3/attrib.c
> new file mode 100644
> index 000000000000..7c85a6d33934
> --- /dev/null
> +++ b/fs/ntfs3/attrib.c
> @@ -0,0 +1,2085 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + *
> + * Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
> + *
> + * TODO: merge attr_set_size/attr_data_get_block/attr_allocate_frame?
> + */
> +
> +#include <linux/blkdev.h>
> +#include <linux/buffer_head.h>
> +#include <linux/fs.h>
> +#include <linux/hash.h>
> +#include <linux/nls.h>
> +#include <linux/ratelimit.h>
> +#include <linux/slab.h>
> +
> +#include "debug.h"
> +#include "ntfs.h"
> +#include "ntfs_fs.h"
> +
> +/*
> + * You can set external NTFS_MIN_LOG2_OF_CLUMP/NTFS_MAX_LOG2_OF_CLUMP to manage
> + * preallocate algorithm
> + */
> +#ifndef NTFS_MIN_LOG2_OF_CLUMP
> +#define NTFS_MIN_LOG2_OF_CLUMP 16
> +#endif
> +
> +#ifndef NTFS_MAX_LOG2_OF_CLUMP
> +#define NTFS_MAX_LOG2_OF_CLUMP 26
> +#endif
> +
> +// 16M
> +#define NTFS_CLUMP_MIN (1 << (NTFS_MIN_LOG2_OF_CLUMP + 8))
> +// 16G
> +#define NTFS_CLUMP_MAX (1ull << (NTFS_MAX_LOG2_OF_CLUMP + 8))
> +
> +/*
> + * get_pre_allocated
> + *
> + */
> +static inline u64 get_pre_allocated(u64 size)
> +{
> + u32 clump;
> + u8 align_shift;
> + u64 ret;
> +
> + if (size <= NTFS_CLUMP_MIN) {
> + clump = 1 << NTFS_MIN_LOG2_OF_CLUMP;
> + align_shift = NTFS_MIN_LOG2_OF_CLUMP;
> + } else if (size >= NTFS_CLUMP_MAX) {
> + clump = 1 << NTFS_MAX_LOG2_OF_CLUMP;
> + align_shift = NTFS_MAX_LOG2_OF_CLUMP;
> + } else {
> + align_shift = NTFS_MIN_LOG2_OF_CLUMP - 1 +
> + __ffs(size >> (8 + NTFS_MIN_LOG2_OF_CLUMP));
> + clump = 1u << align_shift;
> + }
> +
> + ret = (((size + clump - 1) >> align_shift)) << align_shift;
> +
> + return ret;
> +}
> +
> +/*
> + * attr_must_be_resident
> + *
> + * returns true if attribute must be resident
> + */
> +static inline bool attr_must_be_resident(struct ntfs_sb_info *sbi,
> + enum ATTR_TYPE type)
> +{
> + const struct ATTR_DEF_ENTRY *de;
> +
> + switch (type) {
> + case ATTR_STD:
> + case ATTR_NAME:
> + case ATTR_ID:
> + case ATTR_LABEL:
> + case ATTR_VOL_INFO:
> + case ATTR_ROOT:
> + case ATTR_EA_INFO:
> + return true;
> + default:
> + de = ntfs_query_def(sbi, type);
> + if (de && (de->flags & NTFS_ATTR_MUST_BE_RESIDENT))
> + return true;
> + return false;
> + }
> +}
> +
> +/*
> + * attr_load_runs
> + *
> + * load all runs stored in 'attr'
> + */
> +int attr_load_runs(struct ATTRIB *attr, struct ntfs_inode *ni,
> + struct runs_tree *run, const CLST *vcn)
> +{
> + int err;
> + CLST svcn = le64_to_cpu(attr->nres.svcn);
> + CLST evcn = le64_to_cpu(attr->nres.evcn);
> + u32 asize;
> + u16 run_off;
> +
> + if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn))
> + return 0;
> +
> + if (vcn && (evcn < *vcn || *vcn < svcn))
> + return -EINVAL;
> +
> + asize = le32_to_cpu(attr->size);
> + run_off = le16_to_cpu(attr->nres.run_off);
> + err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn,
> + vcn ? *vcn : svcn, Add2Ptr(attr, run_off),
> + asize - run_off);
> + if (err < 0)
> + return err;
> +
> + return 0;
> +}
> +
> +/*
> + * int run_deallocate_ex
> + *
> + * Deallocate clusters
> + */
> +static int run_deallocate_ex(struct ntfs_sb_info *sbi, struct runs_tree *run,
> + CLST vcn, CLST len, CLST *done, bool trim)
> +{
> + int err = 0;
> + CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0;
> + size_t idx;
> +
> + if (!len)
> + goto out;
> +
> + if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) {
> +failed:
> + run_truncate(run, vcn0);
> + err = -EINVAL;
> + goto out;
> + }
> +
> + for (;;) {
> + if (clen > len)
> + clen = len;
> +
> + if (!clen) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (lcn != SPARSE_LCN) {
> + mark_as_free_ex(sbi, lcn, clen, trim);
> + dn += clen;
> + }
> +
> + len -= clen;
> + if (!len)
> + break;
> +
> + vcn_next = vcn + clen;
> + if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) ||
> + vcn != vcn_next) {
> + // save memory - don't load entire run
> + goto failed;
> + }
> + }
> +
> +out:
> + if (done)
> + *done += dn;
> +
> + return err;
> +}
> +
> +/*
> + * attr_allocate_clusters
> + *
> + * find free space, mark it as used and store in 'run'
> + */
> +int attr_allocate_clusters(struct ntfs_sb_info *sbi, struct runs_tree *run,
> + CLST vcn, CLST lcn, CLST len, CLST *pre_alloc,
> + enum ALLOCATE_OPT opt, CLST *alen, const size_t fr,
> + CLST *new_lcn)
> +{
> + int err;
> + CLST flen, vcn0 = vcn, pre = pre_alloc ? *pre_alloc : 0;
> + struct wnd_bitmap *wnd = &sbi->used.bitmap;
> + size_t cnt = run->count;
> +
> + for (;;) {
> + err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen,
> + opt);
> +
> + if (err == -ENOSPC && pre) {
> + pre = 0;
> + if (*pre_alloc)
> + *pre_alloc = 0;
> + continue;
> + }
> +
> + if (err)
> + goto out;
> +
> + if (new_lcn && vcn == vcn0)
> + *new_lcn = lcn;
> +
> + /* Add new fragment into run storage */
> + if (!run_add_entry(run, vcn, lcn, flen, opt == ALLOCATE_MFT)) {
> + down_write_nested(&wnd->rw_lock, BITMAP_MUTEX_CLUSTERS);
> + wnd_set_free(wnd, lcn, flen);
> + up_write(&wnd->rw_lock);
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + vcn += flen;
> +
> + if (flen >= len || opt == ALLOCATE_MFT ||
> + (fr && run->count - cnt >= fr)) {
> + *alen = vcn - vcn0;
> + return 0;
> + }
> +
> + len -= flen;
> + }
> +
> +out:
> + /* undo */
> + run_deallocate_ex(sbi, run, vcn0, vcn - vcn0, NULL, false);
> + run_truncate(run, vcn0);
> +
> + return err;
> +}
> +
> +/*
> + * if page is not NULL - it is already contains resident data
> + * and locked (called from ni_write_frame)
> + */
> +int attr_make_nonresident(struct ntfs_inode *ni, struct ATTRIB *attr,
> + struct ATTR_LIST_ENTRY *le, struct mft_inode *mi,
> + u64 new_size, struct runs_tree *run,
> + struct ATTRIB **ins_attr, struct page *page)
> +{
> + struct ntfs_sb_info *sbi;
> + struct ATTRIB *attr_s;
> + struct MFT_REC *rec;
> + u32 used, asize, rsize, aoff, align;
> + bool is_data;
> + CLST len, alen;
> + char *next;
> + int err;
> +
> + if (attr->non_res) {
> + *ins_attr = attr;
> + return 0;
> + }
> +
> + sbi = mi->sbi;
> + rec = mi->mrec;
> + attr_s = NULL;
> + used = le32_to_cpu(rec->used);
> + asize = le32_to_cpu(attr->size);
> + next = Add2Ptr(attr, asize);
> + aoff = PtrOffset(rec, attr);
> + rsize = le32_to_cpu(attr->res.data_size);
> + is_data = attr->type == ATTR_DATA && !attr->name_len;
> +
> + align = sbi->cluster_size;
> + if (is_attr_compressed(attr))
> + align <<= COMPRESSION_UNIT;
> + len = (rsize + align - 1) >> sbi->cluster_bits;
> +
> + run_init(run);
> +
> + /* make a copy of original attribute */
> + attr_s = ntfs_memdup(attr, asize);
> + if (!attr_s) {
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + if (!len) {
> + /* empty resident -> empty nonresident */
> + alen = 0;
> + } else {
> + const char *data = resident_data(attr);
> +
> + err = attr_allocate_clusters(sbi, run, 0, 0, len, NULL,
> + ALLOCATE_DEF, &alen, 0, NULL);
> + if (err)
> + goto out1;
> +
> + if (!rsize) {
> + /* empty resident -> non empty nonresident */
> + } else if (!is_data) {
> + err = ntfs_sb_write_run(sbi, run, 0, data, rsize);
> + if (err)
> + goto out2;
> + } else if (!page) {
> + char *kaddr;
> +
> + page = grab_cache_page(ni->vfs_inode.i_mapping, 0);
> + if (!page) {
> + err = -ENOMEM;
> + goto out2;
> + }
> + kaddr = kmap_atomic(page);
> + memcpy(kaddr, data, rsize);
> + memset(kaddr + rsize, 0, PAGE_SIZE - rsize);
> + kunmap_atomic(kaddr);
> + flush_dcache_page(page);
> + SetPageUptodate(page);
> + set_page_dirty(page);
> + unlock_page(page);
> + put_page(page);
> + }
> + }
> +
> + /* remove original attribute */
> + used -= asize;
> + memmove(attr, Add2Ptr(attr, asize), used - aoff);
> + rec->used = cpu_to_le32(used);
> + mi->dirty = true;
> + if (le)
> + al_remove_le(ni, le);
> +
> + err = ni_insert_nonresident(ni, attr_s->type, attr_name(attr_s),
> + attr_s->name_len, run, 0, alen,
> + attr_s->flags, &attr, NULL);
> + if (err)
> + goto out3;
> +
> + ntfs_free(attr_s);
> + attr->nres.data_size = cpu_to_le64(rsize);
> + attr->nres.valid_size = attr->nres.data_size;
> +
> + *ins_attr = attr;
> +
> + if (is_data)
> + ni->ni_flags &= ~NI_FLAG_RESIDENT;
> +
> + /* Resident attribute becomes non resident */
> + return 0;
> +
> +out3:
> + attr = Add2Ptr(rec, aoff);
> + memmove(next, attr, used - aoff);
> + memcpy(attr, attr_s, asize);
> + rec->used = cpu_to_le32(used + asize);
> + mi->dirty = true;
> +out2:
> + /* undo: do not trim new allocated clusters */
> + run_deallocate(sbi, run, false);
> + run_close(run);
> +out1:
> + ntfs_free(attr_s);
> + /*reinsert le*/
> +out:
> + return err;
> +}
> +
> +/*
> + * attr_set_size_res
> + *
> + * helper for attr_set_size
> + */
> +static int attr_set_size_res(struct ntfs_inode *ni, struct ATTRIB *attr,
> + struct ATTR_LIST_ENTRY *le, struct mft_inode *mi,
> + u64 new_size, struct runs_tree *run,
> + struct ATTRIB **ins_attr)
> +{
> + struct ntfs_sb_info *sbi = mi->sbi;
> + struct MFT_REC *rec = mi->mrec;
> + u32 used = le32_to_cpu(rec->used);
> + u32 asize = le32_to_cpu(attr->size);
> + u32 aoff = PtrOffset(rec, attr);
> + u32 rsize = le32_to_cpu(attr->res.data_size);
> + u32 tail = used - aoff - asize;
> + char *next = Add2Ptr(attr, asize);
> + s64 dsize = QuadAlign(new_size) - QuadAlign(rsize);
> +
> + if (dsize < 0) {
> + memmove(next + dsize, next, tail);
> + } else if (dsize > 0) {
> + if (used + dsize > sbi->max_bytes_per_attr)
> + return attr_make_nonresident(ni, attr, le, mi, new_size,
> + run, ins_attr, NULL);
> +
> + memmove(next + dsize, next, tail);
> + memset(next, 0, dsize);
> + }
> +
> + if (new_size > rsize)
> + memset(Add2Ptr(resident_data(attr), rsize), 0,
> + new_size - rsize);
> +
> + rec->used = cpu_to_le32(used + dsize);
> + attr->size = cpu_to_le32(asize + dsize);
> + attr->res.data_size = cpu_to_le32(new_size);
> + mi->dirty = true;
> + *ins_attr = attr;
> +
> + return 0;
> +}
> +
> +/*
> + * attr_set_size
> + *
> + * change the size of attribute
> + * Extend:
> + * - sparse/compressed: no allocated clusters
> + * - normal: append allocated and preallocated new clusters
> + * Shrink:
> + * - no deallocate if keep_prealloc is set
> + */
> +int attr_set_size(struct ntfs_inode *ni, enum ATTR_TYPE type,
> + const __le16 *name, u8 name_len, struct runs_tree *run,
> + u64 new_size, const u64 *new_valid, bool keep_prealloc,
> + struct ATTRIB **ret)
> +{
> + int err = 0;
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + u8 cluster_bits = sbi->cluster_bits;
> + bool is_mft =
> + ni->mi.rno == MFT_REC_MFT && type == ATTR_DATA && !name_len;
> + u64 old_valid, old_size, old_alloc, new_alloc, new_alloc_tmp;
> + struct ATTRIB *attr = NULL, *attr_b;
> + struct ATTR_LIST_ENTRY *le, *le_b;
> + struct mft_inode *mi, *mi_b;
> + CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn;
> + CLST next_svcn, pre_alloc = -1, done = 0;
> + bool is_ext;
> + u32 align;
> + struct MFT_REC *rec;
> +
> +again:
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, type, name, name_len, NULL,
> + &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + if (!attr_b->non_res) {
> + err = attr_set_size_res(ni, attr_b, le_b, mi_b, new_size, run,
> + &attr_b);
> + if (err || !attr_b->non_res)
> + goto out;
> +
> + /* layout of records may be changed, so do a full search */
> + goto again;
> + }
> +
> + is_ext = is_attr_ext(attr_b);
> +
> +again_1:
> + align = sbi->cluster_size;
> +
> + if (is_ext) {
> + align <<= attr_b->nres.c_unit;
> + if (is_attr_sparsed(attr_b))
> + keep_prealloc = false;
> + }
> +
> + old_valid = le64_to_cpu(attr_b->nres.valid_size);
> + old_size = le64_to_cpu(attr_b->nres.data_size);
> + old_alloc = le64_to_cpu(attr_b->nres.alloc_size);
> + old_alen = old_alloc >> cluster_bits;
> +
> + new_alloc = (new_size + align - 1) & ~(u64)(align - 1);
> + new_alen = new_alloc >> cluster_bits;
> +
> + if (keep_prealloc && is_ext)
> + keep_prealloc = false;
> +
> + if (keep_prealloc && new_size < old_size) {
> + attr_b->nres.data_size = cpu_to_le64(new_size);
> + mi_b->dirty = true;
> + goto ok;
> + }
> +
> + vcn = old_alen - 1;
> +
> + svcn = le64_to_cpu(attr_b->nres.svcn);
> + evcn = le64_to_cpu(attr_b->nres.evcn);
> +
> + if (svcn <= vcn && vcn <= evcn) {
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + } else if (!le_b) {
> + err = -EINVAL;
> + goto out;
> + } else {
> + le = le_b;
> + attr = ni_find_attr(ni, attr_b, &le, type, name, name_len, &vcn,
> + &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> +next_le_1:
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn = le64_to_cpu(attr->nres.evcn);
> + }
> +
> +next_le:
> + rec = mi->mrec;
> +
> + err = attr_load_runs(attr, ni, run, NULL);
> + if (err)
> + goto out;
> +
> + if (new_size > old_size) {
> + CLST to_allocate;
> + size_t free;
> +
> + if (new_alloc <= old_alloc) {
> + attr_b->nres.data_size = cpu_to_le64(new_size);
> + mi_b->dirty = true;
> + goto ok;
> + }
> +
> + to_allocate = new_alen - old_alen;
> +add_alloc_in_same_attr_seg:
> + lcn = 0;
> + if (is_mft) {
> + /* mft allocates clusters from mftzone */
> + pre_alloc = 0;
> + } else if (is_ext) {
> + /* no preallocate for sparse/compress */
> + pre_alloc = 0;
> + } else if (pre_alloc == -1) {
> + pre_alloc = 0;
> + if (type == ATTR_DATA && !name_len &&
> + sbi->options.prealloc) {
> + CLST new_alen2 = bytes_to_cluster(
> + sbi, get_pre_allocated(new_size));
> + pre_alloc = new_alen2 - new_alen;
> + }
> +
> + /* Get the last lcn to allocate from */
> + if (old_alen &&
> + !run_lookup_entry(run, vcn, &lcn, NULL, NULL)) {
> + lcn = SPARSE_LCN;
> + }
> +
> + if (lcn == SPARSE_LCN)
> + lcn = 0;
> + else if (lcn)
> + lcn += 1;
> +
> + free = wnd_zeroes(&sbi->used.bitmap);
> + if (to_allocate > free) {
> + err = -ENOSPC;
> + goto out;
> + }
> +
> + if (pre_alloc && to_allocate + pre_alloc > free)
> + pre_alloc = 0;
> + }
> +
> + vcn = old_alen;
> +
> + if (is_ext) {
> + if (!run_add_entry(run, vcn, SPARSE_LCN, to_allocate,
> + false)) {
> + err = -ENOMEM;
> + goto out;
> + }
> + alen = to_allocate;
> + } else {
> + /* ~3 bytes per fragment */
> + err = attr_allocate_clusters(
> + sbi, run, vcn, lcn, to_allocate, &pre_alloc,
> + is_mft ? ALLOCATE_MFT : 0, &alen,
> + is_mft ? 0 :
> + (sbi->record_size -
> + le32_to_cpu(rec->used) + 8) /
> + 3 +
> + 1,
> + NULL);
> + if (err)
> + goto out;
> + }
> +
> + done += alen;
> + vcn += alen;
> + if (to_allocate > alen)
> + to_allocate -= alen;
> + else
> + to_allocate = 0;
> +
> +pack_runs:
> + err = mi_pack_runs(mi, attr, run, vcn - svcn);
> + if (err)
> + goto out;
> +
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> + new_alloc_tmp = (u64)next_svcn << cluster_bits;
> + attr_b->nres.alloc_size = cpu_to_le64(new_alloc_tmp);
> + mi_b->dirty = true;
> +
> + if (next_svcn >= vcn && !to_allocate) {
> + /* Normal way. update attribute and exit */
> + attr_b->nres.data_size = cpu_to_le64(new_size);
> + goto ok;
> + }
> +
> + /* at least two mft to avoid recursive loop*/
> + if (is_mft && next_svcn == vcn &&
> + ((u64)done << sbi->cluster_bits) >= 2 * sbi->record_size) {
> + new_size = new_alloc_tmp;
> + attr_b->nres.data_size = attr_b->nres.alloc_size;
> + goto ok;
> + }
> +
> + if (le32_to_cpu(rec->used) < sbi->record_size) {
> + old_alen = next_svcn;
> + evcn = old_alen - 1;
> + goto add_alloc_in_same_attr_seg;
> + }
> +
> + attr_b->nres.data_size = attr_b->nres.alloc_size;
> + if (new_alloc_tmp < old_valid)
> + attr_b->nres.valid_size = attr_b->nres.data_size;
> +
> + if (type == ATTR_LIST) {
> + err = ni_expand_list(ni);
> + if (err)
> + goto out;
> + if (next_svcn < vcn)
> + goto pack_runs;
> +
> + /* layout of records is changed */
> + goto again;
> + }
> +
> + if (!ni->attr_list.size) {
> + err = ni_create_attr_list(ni);
> + if (err)
> + goto out;
> + /* layout of records is changed */
> + }
> +
> + if (next_svcn >= vcn) {
> + /* this is mft data, repeat */
> + goto again;
> + }
> +
> + /* insert new attribute segment */
> + err = ni_insert_nonresident(ni, type, name, name_len, run,
> + next_svcn, vcn - next_svcn,
> + attr_b->flags, &attr, &mi);
> + if (err)
> + goto out;
> +
> + if (!is_mft)
> + run_truncate_head(run, evcn + 1);
> +
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn = le64_to_cpu(attr->nres.evcn);
> +
> + le_b = NULL;
> + /* layout of records maybe changed */
> + /* find base attribute to update*/
> + attr_b = ni_find_attr(ni, NULL, &le_b, type, name, name_len,
> + NULL, &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + attr_b->nres.alloc_size = cpu_to_le64((u64)vcn << cluster_bits);
> + attr_b->nres.data_size = attr_b->nres.alloc_size;
> + attr_b->nres.valid_size = attr_b->nres.alloc_size;
> + mi_b->dirty = true;
> + goto again_1;
> + }
> +
> + if (new_size != old_size ||
> + (new_alloc != old_alloc && !keep_prealloc)) {
> + vcn = max(svcn, new_alen);
> + new_alloc_tmp = (u64)vcn << cluster_bits;
> +
> + alen = 0;
> + err = run_deallocate_ex(sbi, run, vcn, evcn - vcn + 1, &alen,
> + true);
> + if (err)
> + goto out;
> +
> + run_truncate(run, vcn);
> +
> + if (vcn > svcn) {
> + err = mi_pack_runs(mi, attr, run, vcn - svcn);
> + if (err)
> + goto out;
> + } else if (le && le->vcn) {
> + u16 le_sz = le16_to_cpu(le->size);
> +
> + /*
> + * NOTE: list entries for one attribute are always
> + * the same size. We deal with last entry (vcn==0)
> + * and it is not first in entries array
> + * (list entry for std attribute always first)
> + * So it is safe to step back
> + */
> + mi_remove_attr(mi, attr);
> +
> + if (!al_remove_le(ni, le)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + le = (struct ATTR_LIST_ENTRY *)((u8 *)le - le_sz);
> + } else {
> + attr->nres.evcn = cpu_to_le64((u64)vcn - 1);
> + mi->dirty = true;
> + }
> +
> + attr_b->nres.alloc_size = cpu_to_le64(new_alloc_tmp);
> +
> + if (vcn == new_alen) {
> + attr_b->nres.data_size = cpu_to_le64(new_size);
> + if (new_size < old_valid)
> + attr_b->nres.valid_size =
> + attr_b->nres.data_size;
> + } else {
> + if (new_alloc_tmp <=
> + le64_to_cpu(attr_b->nres.data_size))
> + attr_b->nres.data_size =
> + attr_b->nres.alloc_size;
> + if (new_alloc_tmp <
> + le64_to_cpu(attr_b->nres.valid_size))
> + attr_b->nres.valid_size =
> + attr_b->nres.alloc_size;
> + }
> +
> + if (is_ext)
> + le64_sub_cpu(&attr_b->nres.total_size,
> + ((u64)alen << cluster_bits));
> +
> + mi_b->dirty = true;
> +
> + if (new_alloc_tmp <= new_alloc)
> + goto ok;
> +
> + old_size = new_alloc_tmp;
> + vcn = svcn - 1;
> +
> + if (le == le_b) {
> + attr = attr_b;
> + mi = mi_b;
> + evcn = svcn - 1;
> + svcn = 0;
> + goto next_le;
> + }
> +
> + if (le->type != type || le->name_len != name_len ||
> + memcmp(le_name(le), name, name_len * sizeof(short))) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + err = ni_load_mi(ni, le, &mi);
> + if (err)
> + goto out;
> +
> + attr = mi_find_attr(mi, NULL, type, name, name_len, &le->id);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + goto next_le_1;
> + }
> +
> +ok:
> + if (new_valid) {
> + __le64 valid = cpu_to_le64(min(*new_valid, new_size));
> +
> + if (attr_b->nres.valid_size != valid) {
> + attr_b->nres.valid_size = valid;
> + mi_b->dirty = true;
> + }
> + }
> +
> +out:
> + if (!err && attr_b && ret)
> + *ret = attr_b;
> +
> + /* update inode_set_bytes*/
> + if (!err && ((type == ATTR_DATA && !name_len) ||
> + (type == ATTR_ALLOC && name == I30_NAME))) {
> + bool dirty = false;
> +
> + if (ni->vfs_inode.i_size != new_size) {
> + ni->vfs_inode.i_size = new_size;
> + dirty = true;
> + }
> +
> + if (attr_b && attr_b->non_res) {
> + new_alloc = le64_to_cpu(attr_b->nres.alloc_size);
> + if (inode_get_bytes(&ni->vfs_inode) != new_alloc) {
> + inode_set_bytes(&ni->vfs_inode, new_alloc);
> + dirty = true;
> + }
> + }
> +
> + if (dirty) {
> + ni->ni_flags |= NI_FLAG_UPDATE_PARENT;
> + mark_inode_dirty(&ni->vfs_inode);
> + }
> + }
> +
> + return err;
> +}
> +
> +int attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn,
> + CLST *len, bool *new)
> +{
> + int err = 0;
> + struct runs_tree *run = &ni->file.run;
> + struct ntfs_sb_info *sbi;
> + u8 cluster_bits;
> + struct ATTRIB *attr = NULL, *attr_b;
> + struct ATTR_LIST_ENTRY *le, *le_b;
> + struct mft_inode *mi, *mi_b;
> + CLST hint, svcn, to_alloc, evcn1, next_svcn, asize, end;
> + u64 new_size, total_size;
> + u32 clst_per_frame;
> + bool ok;
> +
> + if (new)
> + *new = false;
> +
> + down_read(&ni->file.run_lock);
> + ok = run_lookup_entry(run, vcn, lcn, len, NULL);
> + up_read(&ni->file.run_lock);
> +
> + if (ok && (*lcn != SPARSE_LCN || !new)) {
> + /* normal way */
> + return 0;
> + }
> +
> + if (!clen)
> + clen = 1;
> +
> + if (ok && clen > *len)
> + clen = *len;
> +
> + sbi = ni->mi.sbi;
> + cluster_bits = sbi->cluster_bits;
> + new_size = ((u64)vcn + clen) << cluster_bits;
> +
> + ni_lock(ni);
> + down_write(&ni->file.run_lock);
> +
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL, 0, NULL, &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + if (!attr_b->non_res) {
> + *lcn = RESIDENT_LCN;
> + *len = 1;
> + goto out;
> + }
> +
> + asize = le64_to_cpu(attr_b->nres.alloc_size) >> sbi->cluster_bits;
> + if (vcn >= asize) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + clst_per_frame = 1u << attr_b->nres.c_unit;
> + to_alloc = (clen + clst_per_frame - 1) & ~(clst_per_frame - 1);
> +
> + if (vcn + to_alloc > asize)
> + to_alloc = asize - vcn;
> +
> + svcn = le64_to_cpu(attr_b->nres.svcn);
> + evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1;
> +
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> +
> + if (le_b && (vcn < svcn || evcn1 <= vcn)) {
> + attr = ni_find_attr(ni, attr_b, &le, ATTR_DATA, NULL, 0, &vcn,
> + &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + err = attr_load_runs(attr, ni, run, NULL);
> + if (err)
> + goto out;
> +
> + if (!ok) {
> + ok = run_lookup_entry(run, vcn, lcn, len, NULL);
> + if (ok && (*lcn != SPARSE_LCN || !new)) {
> + /* normal way */
> + err = 0;
> + goto ok;
> + }
> +
> + if (!ok && !new) {
> + *len = 0;
> + err = 0;
> + goto ok;
> + }
> +
> + if (ok && clen > *len) {
> + clen = *len;
> + new_size = ((u64)vcn + clen) << cluster_bits;
> + to_alloc = (clen + clst_per_frame - 1) &
> + ~(clst_per_frame - 1);
> + }
> + }
> +
> + if (!is_attr_ext(attr_b)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + /* Get the last lcn to allocate from */
> + hint = 0;
> +
> + if (vcn > evcn1) {
> + if (!run_add_entry(run, evcn1, SPARSE_LCN, vcn - evcn1,
> + false)) {
> + err = -ENOMEM;
> + goto out;
> + }
> + } else if (vcn && !run_lookup_entry(run, vcn - 1, &hint, NULL, NULL)) {
> + hint = -1;
> + }
> +
> + err = attr_allocate_clusters(
> + sbi, run, vcn, hint + 1, to_alloc, NULL, 0, len,
> + (sbi->record_size - le32_to_cpu(mi->mrec->used) + 8) / 3 + 1,
> + lcn);
> + if (err)
> + goto out;
> + *new = true;
> +
> + end = vcn + *len;
> +
> + total_size = le64_to_cpu(attr_b->nres.total_size) +
> + ((u64)*len << cluster_bits);
> +
> +repack:
> + err = mi_pack_runs(mi, attr, run, max(end, evcn1) - svcn);
> + if (err)
> + goto out;
> +
> + attr_b->nres.total_size = cpu_to_le64(total_size);
> + inode_set_bytes(&ni->vfs_inode, total_size);
> + ni->ni_flags |= NI_FLAG_UPDATE_PARENT;
> +
> + mi_b->dirty = true;
> + mark_inode_dirty(&ni->vfs_inode);
> +
> + /* stored [vcn : next_svcn) from [vcn : end) */
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> +
> + if (end <= evcn1) {
> + if (next_svcn == evcn1) {
> + /* Normal way. update attribute and exit */
> + goto ok;
> + }
> + /* add new segment [next_svcn : evcn1 - next_svcn )*/
> + if (!ni->attr_list.size) {
> + err = ni_create_attr_list(ni);
> + if (err)
> + goto out;
> + /* layout of records is changed */
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL,
> + 0, NULL, &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + goto repack;
> + }
> + }
> +
> + svcn = evcn1;
> +
> + /* Estimate next attribute */
> + attr = ni_find_attr(ni, attr, &le, ATTR_DATA, NULL, 0, &svcn, &mi);
> +
> + if (attr) {
> + CLST alloc = bytes_to_cluster(
> + sbi, le64_to_cpu(attr_b->nres.alloc_size));
> + CLST evcn = le64_to_cpu(attr->nres.evcn);
> +
> + if (end < next_svcn)
> + end = next_svcn;
> + while (end > evcn) {
> + /* remove segment [svcn : evcn)*/
> + mi_remove_attr(mi, attr);
> +
> + if (!al_remove_le(ni, le)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (evcn + 1 >= alloc) {
> + /* last attribute segment */
> + evcn1 = evcn + 1;
> + goto ins_ext;
> + }
> +
> + if (ni_load_mi(ni, le, &mi)) {
> + attr = NULL;
> + goto out;
> + }
> +
> + attr = mi_find_attr(mi, NULL, ATTR_DATA, NULL, 0,
> + &le->id);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn = le64_to_cpu(attr->nres.evcn);
> + }
> +
> + if (end < svcn)
> + end = svcn;
> +
> + err = attr_load_runs(attr, ni, run, &end);
> + if (err)
> + goto out;
> +
> + evcn1 = evcn + 1;
> + attr->nres.svcn = cpu_to_le64(next_svcn);
> + err = mi_pack_runs(mi, attr, run, evcn1 - next_svcn);
> + if (err)
> + goto out;
> +
> + le->vcn = cpu_to_le64(next_svcn);
> + ni->attr_list.dirty = true;
> + mi->dirty = true;
> +
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +ins_ext:
> + if (evcn1 > next_svcn) {
> + err = ni_insert_nonresident(ni, ATTR_DATA, NULL, 0, run,
> + next_svcn, evcn1 - next_svcn,
> + attr_b->flags, &attr, &mi);
> + if (err)
> + goto out;
> + }
> +ok:
> + run_truncate_around(run, vcn);
> +out:
> + up_write(&ni->file.run_lock);
> + ni_unlock(ni);
> +
> + return err;
> +}
> +
> +int attr_data_read_resident(struct ntfs_inode *ni, struct page *page)
> +{
> + u64 vbo;
> + struct ATTRIB *attr;
> + u32 data_size;
> +
> + attr = ni_find_attr(ni, NULL, NULL, ATTR_DATA, NULL, 0, NULL, NULL);
> + if (!attr)
> + return -EINVAL;
> +
> + if (attr->non_res)
> + return E_NTFS_NONRESIDENT;
> +
> + vbo = page->index << PAGE_SHIFT;
> + data_size = le32_to_cpu(attr->res.data_size);
> + if (vbo < data_size) {
> + const char *data = resident_data(attr);
> + char *kaddr = kmap_atomic(page);
> + u32 use = data_size - vbo;
> +
> + if (use > PAGE_SIZE)
> + use = PAGE_SIZE;
> +
> + memcpy(kaddr, data + vbo, use);
> + memset(kaddr + use, 0, PAGE_SIZE - use);
> + kunmap_atomic(kaddr);
> + flush_dcache_page(page);
> + SetPageUptodate(page);
> + } else if (!PageUptodate(page)) {
> + zero_user_segment(page, 0, PAGE_SIZE);
> + SetPageUptodate(page);
> + }
> +
> + return 0;
> +}
> +
> +int attr_data_write_resident(struct ntfs_inode *ni, struct page *page)
> +{
> + u64 vbo;
> + struct mft_inode *mi;
> + struct ATTRIB *attr;
> + u32 data_size;
> +
> + attr = ni_find_attr(ni, NULL, NULL, ATTR_DATA, NULL, 0, NULL, &mi);
> + if (!attr)
> + return -EINVAL;
> +
> + if (attr->non_res) {
> + /*return special error code to check this case*/
> + return E_NTFS_NONRESIDENT;
> + }
> +
> + vbo = page->index << PAGE_SHIFT;
> + data_size = le32_to_cpu(attr->res.data_size);
> + if (vbo < data_size) {
> + char *data = resident_data(attr);
> + char *kaddr = kmap_atomic(page);
> + u32 use = data_size - vbo;
> +
> + if (use > PAGE_SIZE)
> + use = PAGE_SIZE;
> + memcpy(data + vbo, kaddr, use);
> + kunmap_atomic(kaddr);
> + mi->dirty = true;
> + }
> + ni->i_valid = data_size;
> +
> + return 0;
> +}
> +
> +/*
> + * attr_load_runs_vcn
> + *
> + * load runs with vcn
> + */
> +int attr_load_runs_vcn(struct ntfs_inode *ni, enum ATTR_TYPE type,
> + const __le16 *name, u8 name_len, struct runs_tree *run,
> + CLST vcn)
> +{
> + struct ATTRIB *attr;
> + int err;
> + CLST svcn, evcn;
> + u16 ro;
> +
> + attr = ni_find_attr(ni, NULL, NULL, type, name, name_len, &vcn, NULL);
> + if (!attr)
> + return -ENOENT;
> +
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn = le64_to_cpu(attr->nres.evcn);
> +
> + if (evcn < vcn || vcn < svcn)
> + return -EINVAL;
> +
> + ro = le16_to_cpu(attr->nres.run_off);
> + err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, svcn,
> + Add2Ptr(attr, ro), le32_to_cpu(attr->size) - ro);
> + if (err < 0)
> + return err;
> + return 0;
> +}
> +
> +/*
> + * load runs for given range [from to)
> + */
> +int attr_load_runs_range(struct ntfs_inode *ni, enum ATTR_TYPE type,
> + const __le16 *name, u8 name_len, struct runs_tree *run,
> + u64 from, u64 to)
> +{
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + u8 cluster_bits = sbi->cluster_bits;
> + CLST vcn = from >> cluster_bits;
> + CLST vcn_last = (to - 1) >> cluster_bits;
> + CLST lcn, clen;
> + int err;
> +
> + for (vcn = from >> cluster_bits; vcn <= vcn_last; vcn += clen) {
> + if (!run_lookup_entry(run, vcn, &lcn, &clen, NULL)) {
> + err = attr_load_runs_vcn(ni, type, name, name_len, run,
> + vcn);
> + if (err)
> + return err;
> + clen = 0; /*next run_lookup_entry(vcn) must be success*/
> + }
> + }
> +
> + return 0;
> +}
> +
> +#ifdef CONFIG_NTFS3_LZX_XPRESS
> +/*
> + * attr_wof_frame_info
> + *
> + * read header of xpress/lzx file to get info about frame
> + */
> +int attr_wof_frame_info(struct ntfs_inode *ni, struct ATTRIB *attr,
> + struct runs_tree *run, u64 frame, u64 frames,
> + u8 frame_bits, u32 *ondisk_size, u64 *vbo_data)
> +{
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + u64 vbo[2], off[2], wof_size;
> + u32 voff;
> + u8 bytes_per_off;
> + char *addr;
> + struct page *page;
> + int i, err;
> + __le32 *off32;
> + __le64 *off64;
> +
> + if (ni->vfs_inode.i_size < 0x100000000ull) {
> + /* file starts with array of 32 bit offsets */
> + bytes_per_off = sizeof(__le32);
> + vbo[1] = frame << 2;
> + *vbo_data = frames << 2;
> + } else {
> + /* file starts with array of 64 bit offsets */
> + bytes_per_off = sizeof(__le64);
> + vbo[1] = frame << 3;
> + *vbo_data = frames << 3;
> + }
> +
> + /*
> + * read 4/8 bytes at [vbo - 4(8)] == offset where compressed frame starts
> + * read 4/8 bytes at [vbo] == offset where compressed frame ends
> + */
> + if (!attr->non_res) {
> + if (vbo[1] + bytes_per_off > le32_to_cpu(attr->res.data_size)) {
> + ntfs_inode_err(&ni->vfs_inode, "is corrupted");
> + return -EINVAL;
> + }
> + addr = resident_data(attr);
> +
> + if (bytes_per_off == sizeof(__le32)) {
> + off32 = Add2Ptr(addr, vbo[1]);
> + off[0] = vbo[1] ? le32_to_cpu(off32[-1]) : 0;
> + off[1] = le32_to_cpu(off32[0]);
> + } else {
> + off64 = Add2Ptr(addr, vbo[1]);
> + off[0] = vbo[1] ? le64_to_cpu(off64[-1]) : 0;
> + off[1] = le64_to_cpu(off64[0]);
> + }
> +
> + *vbo_data += off[0];
> + *ondisk_size = off[1] - off[0];
> + return 0;
> + }
> +
> + wof_size = le64_to_cpu(attr->nres.data_size);
> + down_write(&ni->file.run_lock);
> + page = ni->file.offs_page;
> + if (!page) {
> + page = alloc_page(GFP_KERNEL);
> + if (!page) {
> + err = -ENOMEM;
> + goto out;
> + }
> + page->index = -1;
> + ni->file.offs_page = page;
> + }
> + lock_page(page);
> + addr = page_address(page);
> +
> + if (vbo[1]) {
> + voff = vbo[1] & (PAGE_SIZE - 1);
> + vbo[0] = vbo[1] - bytes_per_off;
> + i = 0;
> + } else {
> + voff = 0;
> + vbo[0] = 0;
> + off[0] = 0;
> + i = 1;
> + }
> +
> + do {
> + pgoff_t index = vbo[i] >> PAGE_SHIFT;
> +
> + if (index != page->index) {
> + u64 from = vbo[i] & ~(u64)(PAGE_SIZE - 1);
> + u64 to = min(from + PAGE_SIZE, wof_size);
> +
> + err = attr_load_runs_range(ni, ATTR_DATA, WOF_NAME,
> + ARRAY_SIZE(WOF_NAME), run,
> + from, to);
> + if (err)
> + goto out1;
> +
> + err = ntfs_bio_pages(sbi, run, &page, 1, from,
> + to - from, REQ_OP_READ);
> + if (err) {
> + page->index = -1;
> + goto out1;
> + }
> + page->index = index;
> + }
> +
> + if (i) {
> + if (bytes_per_off == sizeof(__le32)) {
> + off32 = Add2Ptr(addr, voff);
> + off[1] = le32_to_cpu(*off32);
> + } else {
> + off64 = Add2Ptr(addr, voff);
> + off[1] = le64_to_cpu(*off64);
> + }
> + } else if (!voff) {
> + if (bytes_per_off == sizeof(__le32)) {
> + off32 = Add2Ptr(addr, PAGE_SIZE - sizeof(u32));
> + off[0] = le32_to_cpu(*off32);
> + } else {
> + off64 = Add2Ptr(addr, PAGE_SIZE - sizeof(u64));
> + off[0] = le64_to_cpu(*off64);
> + }
> + } else {
> + /* two values in one page*/
> + if (bytes_per_off == sizeof(__le32)) {
> + off32 = Add2Ptr(addr, voff);
> + off[0] = le32_to_cpu(off32[-1]);
> + off[1] = le32_to_cpu(off32[0]);
> + } else {
> + off64 = Add2Ptr(addr, voff);
> + off[0] = le64_to_cpu(off64[-1]);
> + off[1] = le64_to_cpu(off64[0]);
> + }
> + break;
> + }
> + } while (++i < 2);
> +
> + *vbo_data += off[0];
> + *ondisk_size = off[1] - off[0];
> +
> +out1:
> + unlock_page(page);
> +out:
> + up_write(&ni->file.run_lock);
> + return err;
> +}
> +#endif
> +
> +/*
> + * attr_is_frame_compressed
> + *
> + * This function is used to detect compressed frame
> + */
> +int attr_is_frame_compressed(struct ntfs_inode *ni, struct ATTRIB *attr,
> + CLST frame, CLST *clst_data)
> +{
> + int err;
> + u32 clst_frame;
> + CLST clen, lcn, vcn, alen, slen, vcn_next;
> + size_t idx;
> + struct runs_tree *run;
> +
> + *clst_data = 0;
> +
> + if (!is_attr_compressed(attr))
> + return 0;
> +
> + if (!attr->non_res)
> + return 0;
> +
> + clst_frame = 1u << attr->nres.c_unit;
> + vcn = frame * clst_frame;
> + run = &ni->file.run;
> +
> + if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) {
> + err = attr_load_runs_vcn(ni, attr->type, attr_name(attr),
> + attr->name_len, run, vcn);
> + if (err)
> + return err;
> +
> + if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx))
> + return -EINVAL;
> + }
> +
> + if (lcn == SPARSE_LCN) {
> + /* sparsed frame */
> + return 0;
> + }
> +
> + if (clen >= clst_frame) {
> + /*
> + * The frame is not compressed 'cause
> + * it does not contain any sparse clusters
> + */
> + *clst_data = clst_frame;
> + return 0;
> + }
> +
> + alen = bytes_to_cluster(ni->mi.sbi, le64_to_cpu(attr->nres.alloc_size));
> + slen = 0;
> + *clst_data = clen;
> +
> + /*
> + * The frame is compressed if *clst_data + slen >= clst_frame
> + * Check next fragments
> + */
> + while ((vcn += clen) < alen) {
> + vcn_next = vcn;
> +
> + if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) ||
> + vcn_next != vcn) {
> + err = attr_load_runs_vcn(ni, attr->type,
> + attr_name(attr),
> + attr->name_len, run, vcn_next);
> + if (err)
> + return err;
> + vcn = vcn_next;
> +
> + if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx))
> + return -EINVAL;
> + }
> +
> + if (lcn == SPARSE_LCN) {
> + slen += clen;
> + } else {
> + if (slen) {
> + /*
> + * data_clusters + sparse_clusters =
> + * not enough for frame
> + */
> + return -EINVAL;
> + }
> + *clst_data += clen;
> + }
> +
> + if (*clst_data + slen >= clst_frame) {
> + if (!slen) {
> + /*
> + * There is no sparsed clusters in this frame
> + * So it is not compressed
> + */
> + *clst_data = clst_frame;
> + } else {
> + /*frame is compressed*/
> + }
> + break;
> + }
> + }
> +
> + return 0;
> +}
> +
> +/*
> + * attr_allocate_frame
> + *
> + * allocate/free clusters for 'frame'
> + * assumed: down_write(&ni->file.run_lock);
> + */
> +int attr_allocate_frame(struct ntfs_inode *ni, CLST frame, size_t compr_size,
> + u64 new_valid)
> +{
> + int err = 0;
> + struct runs_tree *run = &ni->file.run;
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + struct ATTRIB *attr = NULL, *attr_b;
> + struct ATTR_LIST_ENTRY *le, *le_b;
> + struct mft_inode *mi, *mi_b;
> + CLST svcn, evcn1, next_svcn, lcn, len;
> + CLST vcn, end, clst_data;
> + u64 total_size, valid_size, data_size;
> +
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL, 0, NULL, &mi_b);
> + if (!attr_b)
> + return -ENOENT;
> +
> + if (!is_attr_ext(attr_b))
> + return -EINVAL;
> +
> + vcn = frame << NTFS_LZNT_CUNIT;
> + total_size = le64_to_cpu(attr_b->nres.total_size);
> +
> + svcn = le64_to_cpu(attr_b->nres.svcn);
> + evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1;
> + data_size = le64_to_cpu(attr_b->nres.data_size);
> +
> + if (svcn <= vcn && vcn < evcn1) {
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + } else if (!le_b) {
> + err = -EINVAL;
> + goto out;
> + } else {
> + le = le_b;
> + attr = ni_find_attr(ni, attr_b, &le, ATTR_DATA, NULL, 0, &vcn,
> + &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + err = attr_load_runs(attr, ni, run, NULL);
> + if (err)
> + goto out;
> +
> + err = attr_is_frame_compressed(ni, attr_b, frame, &clst_data);
> + if (err)
> + goto out;
> +
> + total_size -= (u64)clst_data << sbi->cluster_bits;
> +
> + len = bytes_to_cluster(sbi, compr_size);
> +
> + if (len == clst_data)
> + goto out;
> +
> + if (len < clst_data) {
> + err = run_deallocate_ex(sbi, run, vcn + len, clst_data - len,
> + NULL, true);
> + if (err)
> + goto out;
> +
> + if (!run_add_entry(run, vcn + len, SPARSE_LCN, clst_data - len,
> + false)) {
> + err = -ENOMEM;
> + goto out;
> + }
> + end = vcn + clst_data;
> + /* run contains updated range [vcn + len : end) */
> + } else {
> + CLST alen, hint = 0;
> + /* Get the last lcn to allocate from */
> + if (vcn + clst_data &&
> + !run_lookup_entry(run, vcn + clst_data - 1, &hint, NULL,
> + NULL)) {
> + hint = -1;
> + }
> +
> + err = attr_allocate_clusters(sbi, run, vcn + clst_data,
> + hint + 1, len - clst_data, NULL, 0,
> + &alen, 0, &lcn);
> + if (err)
> + goto out;
> +
> + end = vcn + len;
> + /* run contains updated range [vcn + clst_data : end) */
> + }
> +
> + total_size += (u64)len << sbi->cluster_bits;
> +
> +repack:
> + err = mi_pack_runs(mi, attr, run, max(end, evcn1) - svcn);
> + if (err)
> + goto out;
> +
> + attr_b->nres.total_size = cpu_to_le64(total_size);
> + inode_set_bytes(&ni->vfs_inode, total_size);
> +
> + mi_b->dirty = true;
> + mark_inode_dirty(&ni->vfs_inode);
> +
> + /* stored [vcn : next_svcn) from [vcn : end) */
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> +
> + if (end <= evcn1) {
> + if (next_svcn == evcn1) {
> + /* Normal way. update attribute and exit */
> + goto ok;
> + }
> + /* add new segment [next_svcn : evcn1 - next_svcn )*/
> + if (!ni->attr_list.size) {
> + err = ni_create_attr_list(ni);
> + if (err)
> + goto out;
> + /* layout of records is changed */
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL,
> + 0, NULL, &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + goto repack;
> + }
> + }
> +
> + svcn = evcn1;
> +
> + /* Estimate next attribute */
> + attr = ni_find_attr(ni, attr, &le, ATTR_DATA, NULL, 0, &svcn, &mi);
> +
> + if (attr) {
> + CLST alloc = bytes_to_cluster(
> + sbi, le64_to_cpu(attr_b->nres.alloc_size));
> + CLST evcn = le64_to_cpu(attr->nres.evcn);
> +
> + if (end < next_svcn)
> + end = next_svcn;
> + while (end > evcn) {
> + /* remove segment [svcn : evcn)*/
> + mi_remove_attr(mi, attr);
> +
> + if (!al_remove_le(ni, le)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (evcn + 1 >= alloc) {
> + /* last attribute segment */
> + evcn1 = evcn + 1;
> + goto ins_ext;
> + }
> +
> + if (ni_load_mi(ni, le, &mi)) {
> + attr = NULL;
> + goto out;
> + }
> +
> + attr = mi_find_attr(mi, NULL, ATTR_DATA, NULL, 0,
> + &le->id);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn = le64_to_cpu(attr->nres.evcn);
> + }
> +
> + if (end < svcn)
> + end = svcn;
> +
> + err = attr_load_runs(attr, ni, run, &end);
> + if (err)
> + goto out;
> +
> + evcn1 = evcn + 1;
> + attr->nres.svcn = cpu_to_le64(next_svcn);
> + err = mi_pack_runs(mi, attr, run, evcn1 - next_svcn);
> + if (err)
> + goto out;
> +
> + le->vcn = cpu_to_le64(next_svcn);
> + ni->attr_list.dirty = true;
> + mi->dirty = true;
> +
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +ins_ext:
> + if (evcn1 > next_svcn) {
> + err = ni_insert_nonresident(ni, ATTR_DATA, NULL, 0, run,
> + next_svcn, evcn1 - next_svcn,
> + attr_b->flags, &attr, &mi);
> + if (err)
> + goto out;
> + }
> +ok:
> + run_truncate_around(run, vcn);
> +out:
> + if (new_valid > data_size)
> + new_valid = data_size;
> +
> + valid_size = le64_to_cpu(attr_b->nres.valid_size);
> + if (new_valid != valid_size) {
> + attr_b->nres.valid_size = cpu_to_le64(valid_size);
> + mi_b->dirty = true;
> + }
> +
> + return err;
> +}
> +
> +/* Collapse range in file */
> +int attr_collapse_range(struct ntfs_inode *ni, u64 vbo, u64 bytes)
> +{
> + int err = 0;
> + struct runs_tree *run = &ni->file.run;
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + struct ATTRIB *attr = NULL, *attr_b;
> + struct ATTR_LIST_ENTRY *le, *le_b;
> + struct mft_inode *mi, *mi_b;
> + CLST svcn, evcn1, len, dealloc, alen;
> + CLST vcn, end;
> + u64 valid_size, data_size, alloc_size, total_size;
> + u32 mask;
> + __le16 a_flags;
> +
> + if (!bytes)
> + return 0;
> +
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL, 0, NULL, &mi_b);
> + if (!attr_b)
> + return -ENOENT;
> +
> + if (!attr_b->non_res) {
> + /* Attribute is resident. Nothing to do? */
> + return 0;
> + }
> +
> + data_size = le64_to_cpu(attr_b->nres.data_size);
> + alloc_size = le64_to_cpu(attr_b->nres.alloc_size);
> + a_flags = attr_b->flags;
> +
> + if (is_attr_ext(attr_b)) {
> + total_size = le64_to_cpu(attr_b->nres.total_size);
> + mask = (sbi->cluster_size << attr_b->nres.c_unit) - 1;
> + } else {
> + total_size = alloc_size;
> + mask = sbi->cluster_mask;
> + }
> +
> + if (vbo & mask)
> + return -EINVAL;
> +
> + if (bytes & mask)
> + return -EINVAL;
> +
> + if (vbo > data_size)
> + return -EINVAL;
> +
> + down_write(&ni->file.run_lock);
> +
> + if (vbo + bytes >= data_size) {
> + u64 new_valid = min(ni->i_valid, vbo);
> +
> + /* Simple truncate file at 'vbo' */
> + truncate_setsize(&ni->vfs_inode, vbo);
> + err = attr_set_size(ni, ATTR_DATA, NULL, 0, &ni->file.run, vbo,
> + &new_valid, true, NULL);
> +
> + if (!err && new_valid < ni->i_valid)
> + ni->i_valid = new_valid;
> +
> + goto out;
> + }
> +
> + /*
> + * Enumerate all attribute segments and collapse
> + */
> + alen = alloc_size >> sbi->cluster_bits;
> + vcn = vbo >> sbi->cluster_bits;
> + len = bytes >> sbi->cluster_bits;
> + end = vcn + len;
> + dealloc = 0;
> +
> + svcn = le64_to_cpu(attr_b->nres.svcn);
> + evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1;
> +
> + if (svcn <= vcn && vcn < evcn1) {
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + } else if (!le_b) {
> + err = -EINVAL;
> + goto out;
> + } else {
> + le = le_b;
> + attr = ni_find_attr(ni, attr_b, &le, ATTR_DATA, NULL, 0, &vcn,
> + &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + for (;;) {
> + if (svcn >= end) {
> + /* shift vcn */
> + attr->nres.svcn = cpu_to_le64(svcn - len);
> + attr->nres.evcn = cpu_to_le64(evcn1 - 1 - len);
> + if (le) {
> + le->vcn = attr->nres.svcn;
> + ni->attr_list.dirty = true;
> + }
> + mi->dirty = true;
> + } else if (svcn < vcn || end < evcn1) {
> + CLST vcn1, eat, next_svcn;
> +
> + /* collapse a part of this attribute segment */
> + err = attr_load_runs(attr, ni, run, &svcn);
> + if (err)
> + goto out;
> + vcn1 = max(vcn, svcn);
> + eat = min(end, evcn1) - vcn1;
> +
> + err = run_deallocate_ex(sbi, run, vcn1, eat, &dealloc,
> + true);
> + if (err)
> + goto out;
> +
> + if (!run_collapse_range(run, vcn1, eat)) {
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + if (svcn >= vcn) {
> + /* shift vcn */
> + attr->nres.svcn = cpu_to_le64(vcn);
> + if (le) {
> + le->vcn = attr->nres.svcn;
> + ni->attr_list.dirty = true;
> + }
> + }
> +
> + err = mi_pack_runs(mi, attr, run, evcn1 - svcn - eat);
> + if (err)
> + goto out;
> +
> + next_svcn = le64_to_cpu(attr->nres.evcn) + 1;
> + if (next_svcn + eat < evcn1) {
> + err = ni_insert_nonresident(
> + ni, ATTR_DATA, NULL, 0, run, next_svcn,
> + evcn1 - eat - next_svcn, a_flags, &attr,
> + &mi);
> + if (err)
> + goto out;
> +
> + /* layout of records maybe changed */
> + attr_b = NULL;
> + le = al_find_ex(ni, NULL, ATTR_DATA, NULL, 0,
> + &next_svcn);
> + if (!le) {
> + err = -EINVAL;
> + goto out;
> + }
> + }
> +
> + /* free all allocated memory */
> + run_truncate(run, 0);
> + } else {
> + u16 le_sz;
> + u16 roff = le16_to_cpu(attr->nres.run_off);
> +
> + /*run==1 means unpack and deallocate*/
> + run_unpack_ex(RUN_DEALLOCATE, sbi, ni->mi.rno, svcn,
> + evcn1 - 1, svcn, Add2Ptr(attr, roff),
> + le32_to_cpu(attr->size) - roff);
> +
> + /* delete this attribute segment */
> + mi_remove_attr(mi, attr);
> + if (!le)
> + break;
> +
> + le_sz = le16_to_cpu(le->size);
> + if (!al_remove_le(ni, le)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (evcn1 >= alen)
> + break;
> +
> + if (!svcn) {
> + /* Load next record that contains this attribute */
> + if (ni_load_mi(ni, le, &mi)) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + /* Look for required attribute */
> + attr = mi_find_attr(mi, NULL, ATTR_DATA, NULL,
> + 0, &le->id);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> + goto next_attr;
> + }
> + le = (struct ATTR_LIST_ENTRY *)((u8 *)le - le_sz);
> + }
> +
> + if (evcn1 >= alen)
> + break;
> +
> + attr = ni_enum_attr_ex(ni, attr, &le, &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> +next_attr:
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + if (!attr_b) {
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL, 0, NULL,
> + &mi_b);
> + if (!attr_b) {
> + err = -ENOENT;
> + goto out;
> + }
> + }
> +
> + data_size -= bytes;
> + valid_size = ni->i_valid;
> + if (vbo + bytes <= valid_size)
> + valid_size -= bytes;
> + else if (vbo < valid_size)
> + valid_size = vbo;
> +
> + attr_b->nres.alloc_size = cpu_to_le64(alloc_size - bytes);
> + attr_b->nres.data_size = cpu_to_le64(data_size);
> + attr_b->nres.valid_size = cpu_to_le64(min(valid_size, data_size));
> + total_size -= (u64)dealloc << sbi->cluster_bits;
> + if (is_attr_ext(attr_b))
> + attr_b->nres.total_size = cpu_to_le64(total_size);
> + mi_b->dirty = true;
> +
> + /*update inode size*/
> + ni->i_valid = valid_size;
> + ni->vfs_inode.i_size = data_size;
> + inode_set_bytes(&ni->vfs_inode, total_size);
> + ni->ni_flags |= NI_FLAG_UPDATE_PARENT;
> + mark_inode_dirty(&ni->vfs_inode);
> +
> +out:
> + up_write(&ni->file.run_lock);
> + if (err)
> + make_bad_inode(&ni->vfs_inode);
> +
> + return err;
> +}
> +
> +/* not for normal files */
> +int attr_punch_hole(struct ntfs_inode *ni, u64 vbo, u64 bytes)
> +{
> + int err = 0;
> + struct runs_tree *run = &ni->file.run;
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + struct ATTRIB *attr = NULL, *attr_b;
> + struct ATTR_LIST_ENTRY *le, *le_b;
> + struct mft_inode *mi, *mi_b;
> + CLST svcn, evcn1, vcn, len, end, alen, dealloc;
> + u64 total_size, alloc_size;
> +
> + if (!bytes)
> + return 0;
> +
> + le_b = NULL;
> + attr_b = ni_find_attr(ni, NULL, &le_b, ATTR_DATA, NULL, 0, NULL, &mi_b);
> + if (!attr_b)
> + return -ENOENT;
> +
> + if (!attr_b->non_res) {
> + u32 data_size = le32_to_cpu(attr->res.data_size);
> + u32 from, to;
> +
> + if (vbo > data_size)
> + return 0;
> +
> + from = vbo;
> + to = (vbo + bytes) < data_size ? (vbo + bytes) : data_size;
> + memset(Add2Ptr(resident_data(attr_b), from), 0, to - from);
> + return 0;
> + }
> +
> + /* TODO: add support for normal files too */
> + if (!is_attr_ext(attr_b))
> + return -EOPNOTSUPP;
> +
> + alloc_size = le64_to_cpu(attr_b->nres.alloc_size);
> + total_size = le64_to_cpu(attr_b->nres.total_size);
> +
> + if (vbo >= alloc_size) {
> + // NOTE: it is allowed
> + return 0;
> + }
> +
> + if (vbo + bytes > alloc_size)
> + bytes = alloc_size - vbo;
> +
> + down_write(&ni->file.run_lock);
> + /*
> + * Enumerate all attribute segments and punch hole where necessary
> + */
> + alen = alloc_size >> sbi->cluster_bits;
> + vcn = vbo >> sbi->cluster_bits;
> + len = bytes >> sbi->cluster_bits;
> + end = vcn + len;
> + dealloc = 0;
> +
> + svcn = le64_to_cpu(attr_b->nres.svcn);
> + evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1;
> +
> + if (svcn <= vcn && vcn < evcn1) {
> + attr = attr_b;
> + le = le_b;
> + mi = mi_b;
> + } else if (!le_b) {
> + err = -EINVAL;
> + goto out;
> + } else {
> + le = le_b;
> + attr = ni_find_attr(ni, attr_b, &le, ATTR_DATA, NULL, 0, &vcn,
> + &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + while (svcn < end) {
> + CLST vcn1, zero, dealloc2;
> +
> + err = attr_load_runs(attr, ni, run, &svcn);
> + if (err)
> + goto out;
> + vcn1 = max(vcn, svcn);
> + zero = min(end, evcn1) - vcn1;
> +
> + dealloc2 = dealloc;
> + err = run_deallocate_ex(sbi, run, vcn1, zero, &dealloc, true);
> + if (err)
> + goto out;
> +
> + if (dealloc2 == dealloc) {
> + /* looks like the required range is already sparsed */
> + } else {
> + if (!run_add_entry(run, vcn1, SPARSE_LCN, zero,
> + false)) {
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + err = mi_pack_runs(mi, attr, run, evcn1 - svcn);
> + if (err)
> + goto out;
> + }
> + /* free all allocated memory */
> + run_truncate(run, 0);
> +
> + if (evcn1 >= alen)
> + break;
> +
> + attr = ni_enum_attr_ex(ni, attr, &le, &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + svcn = le64_to_cpu(attr->nres.svcn);
> + evcn1 = le64_to_cpu(attr->nres.evcn) + 1;
> + }
> +
> + total_size -= (u64)dealloc << sbi->cluster_bits;
> + attr_b->nres.total_size = cpu_to_le64(total_size);
> + mi_b->dirty = true;
> +
> + /*update inode size*/
> + inode_set_bytes(&ni->vfs_inode, total_size);
> + ni->ni_flags |= NI_FLAG_UPDATE_PARENT;
> + mark_inode_dirty(&ni->vfs_inode);
> +
> +out:
> + up_write(&ni->file.run_lock);
> + if (err)
> + make_bad_inode(&ni->vfs_inode);
> +
> + return err;
> +}
> diff --git a/fs/ntfs3/attrlist.c b/fs/ntfs3/attrlist.c
> new file mode 100644
> index 000000000000..04a30b58cce5
> --- /dev/null
> +++ b/fs/ntfs3/attrlist.c
> @@ -0,0 +1,457 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + *
> + * Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
> + *
> + */
> +
> +#include <linux/blkdev.h>
> +#include <linux/buffer_head.h>
> +#include <linux/fs.h>
> +#include <linux/nls.h>
> +
> +#include "debug.h"
> +#include "ntfs.h"
> +#include "ntfs_fs.h"
> +
> +/* Returns true if le is valid */
> +static inline bool al_is_valid_le(const struct ntfs_inode *ni,
> + struct ATTR_LIST_ENTRY *le)
> +{
> + if (!le || !ni->attr_list.le || !ni->attr_list.size)
> + return false;
> +
> + return PtrOffset(ni->attr_list.le, le) + le16_to_cpu(le->size) <=
> + ni->attr_list.size;
> +}
> +
> +void al_destroy(struct ntfs_inode *ni)
> +{
> + run_close(&ni->attr_list.run);
> + ntfs_free(ni->attr_list.le);
> + ni->attr_list.le = NULL;
> + ni->attr_list.size = 0;
> + ni->attr_list.dirty = false;
> +}
> +
> +/*
> + * ntfs_load_attr_list
> + *
> + * This method makes sure that the ATTRIB list, if present,
> + * has been properly set up.
> + */
> +int ntfs_load_attr_list(struct ntfs_inode *ni, struct ATTRIB *attr)
> +{
> + int err;
> + size_t lsize;
> + void *le = NULL;
> +
> + if (ni->attr_list.size)
> + return 0;
> +
> + if (!attr->non_res) {
> + lsize = le32_to_cpu(attr->res.data_size);
> + le = ntfs_malloc(al_aligned(lsize));
> + if (!le) {
> + err = -ENOMEM;
> + goto out;
> + }
> + memcpy(le, resident_data(attr), lsize);
> + } else if (attr->nres.svcn) {
> + err = -EINVAL;
> + goto out;
> + } else {
> + u16 run_off = le16_to_cpu(attr->nres.run_off);
> +
> + lsize = le64_to_cpu(attr->nres.data_size);
> +
> + run_init(&ni->attr_list.run);
> +
> + err = run_unpack_ex(&ni->attr_list.run, ni->mi.sbi, ni->mi.rno,
> + 0, le64_to_cpu(attr->nres.evcn), 0,
> + Add2Ptr(attr, run_off),
> + le32_to_cpu(attr->size) - run_off);
> + if (err < 0)
> + goto out;
> +
> + le = ntfs_malloc(al_aligned(lsize));
> + if (!le) {
> + err = -ENOMEM;
> + goto out;
> + }
> +
> + err = ntfs_read_run_nb(ni->mi.sbi, &ni->attr_list.run, 0, le,
> + lsize, NULL);
> + if (err)
> + goto out;
> + }
> +
> + ni->attr_list.size = lsize;
> + ni->attr_list.le = le;
> +
> + return 0;
> +
> +out:
> + ni->attr_list.le = le;
> + al_destroy(ni);
> +
> + return err;
> +}
> +
> +/*
> + * al_enumerate
> + *
> + * Returns the next list le
> + * if le is NULL then returns the first le
> + */
> +struct ATTR_LIST_ENTRY *al_enumerate(struct ntfs_inode *ni,
> + struct ATTR_LIST_ENTRY *le)
> +{
> + size_t off;
> + u16 sz;
> +
> + if (!le) {
> + le = ni->attr_list.le;
> + } else {
> + sz = le16_to_cpu(le->size);
> + if (sz < sizeof(struct ATTR_LIST_ENTRY)) {
> + /* Impossible 'cause we should not return such le */
> + return NULL;
> + }
> + le = Add2Ptr(le, sz);
> + }
> +
> + /* Check boundary */
> + off = PtrOffset(ni->attr_list.le, le);
> + if (off + sizeof(struct ATTR_LIST_ENTRY) > ni->attr_list.size) {
> + // The regular end of list
> + return NULL;
> + }
> +
> + sz = le16_to_cpu(le->size);
> +
> + /* Check le for errors */
> + if (sz < sizeof(struct ATTR_LIST_ENTRY) ||
> + off + sz > ni->attr_list.size ||
> + sz < le->name_off + le->name_len * sizeof(short)) {
> + return NULL;
> + }
> +
> + return le;
> +}
> +
> +/*
> + * al_find_le
> + *
> + * finds the first le in the list which matches type, name and vcn
> + * Returns NULL if not found
> + */
> +struct ATTR_LIST_ENTRY *al_find_le(struct ntfs_inode *ni,
> + struct ATTR_LIST_ENTRY *le,
> + const struct ATTRIB *attr)
> +{
> + CLST svcn = attr_svcn(attr);
> +
> + return al_find_ex(ni, le, attr->type, attr_name(attr), attr->name_len,
> + &svcn);
> +}
> +
> +/*
> + * al_find_ex
> + *
> + * finds the first le in the list which matches type, name and vcn
> + * Returns NULL if not found
> + */
> +struct ATTR_LIST_ENTRY *al_find_ex(struct ntfs_inode *ni,
> + struct ATTR_LIST_ENTRY *le,
> + enum ATTR_TYPE type, const __le16 *name,
> + u8 name_len, const CLST *vcn)
> +{
> + struct ATTR_LIST_ENTRY *ret = NULL;
> + u32 type_in = le32_to_cpu(type);
> +
> + while ((le = al_enumerate(ni, le))) {
> + u64 le_vcn;
> + int diff = le32_to_cpu(le->type) - type_in;
> +
> + /* List entries are sorted by type, name and vcn */
> + if (diff < 0)
> + continue;
> +
> + if (diff > 0)
> + return ret;
> +
> + if (le->name_len != name_len)
> + continue;
> +
> + le_vcn = le64_to_cpu(le->vcn);
> + if (!le_vcn) {
> + /*
> + * compare entry names only for entry with vcn == 0
> + */
> + diff = ntfs_cmp_names(le_name(le), name_len, name,
> + name_len, ni->mi.sbi->upcase,
> + true);
> + if (diff < 0)
> + continue;
> +
> + if (diff > 0)
> + return ret;
> + }
> +
> + if (!vcn)
> + return le;
> +
> + if (*vcn == le_vcn)
> + return le;
> +
> + if (*vcn < le_vcn)
> + return ret;
> +
> + ret = le;
> + }
> +
> + return ret;
> +}
> +
> +/*
> + * al_find_le_to_insert
> + *
> + * finds the first list entry which matches type, name and vcn
> + */
> +static struct ATTR_LIST_ENTRY *al_find_le_to_insert(struct ntfs_inode *ni,
> + enum ATTR_TYPE type,
> + const __le16 *name,
> + u8 name_len, CLST vcn)
> +{
> + struct ATTR_LIST_ENTRY *le = NULL, *prev;
> + u32 type_in = le32_to_cpu(type);
> +
> + /* List entries are sorted by type, name, vcn */
> + while ((le = al_enumerate(ni, prev = le))) {
> + int diff = le32_to_cpu(le->type) - type_in;
> +
> + if (diff < 0)
> + continue;
> +
> + if (diff > 0)
> + return le;
> +
> + if (!le->vcn) {
> + /*
> + * compare entry names only for entry with vcn == 0
> + */
> + diff = ntfs_cmp_names(le_name(le), le->name_len, name,
> + name_len, ni->mi.sbi->upcase,
> + true);
> + if (diff < 0)
> + continue;
> +
> + if (diff > 0)
> + return le;
> + }
> +
> + if (le64_to_cpu(le->vcn) >= vcn)
> + return le;
> + }
> +
> + return prev ? Add2Ptr(prev, le16_to_cpu(prev->size)) : ni->attr_list.le;
> +}
> +
> +/*
> + * al_add_le
> + *
> + * adds an "attribute list entry" to the list.
> + */
> +int al_add_le(struct ntfs_inode *ni, enum ATTR_TYPE type, const __le16 *name,
> + u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref,
> + struct ATTR_LIST_ENTRY **new_le)
> +{
> + int err;
> + struct ATTRIB *attr;
> + struct ATTR_LIST_ENTRY *le;
> + size_t off;
> + u16 sz;
> + size_t asize, new_asize;
> + u64 new_size;
> + typeof(ni->attr_list) *al = &ni->attr_list;
> +
> + /*
> + * Compute the size of the new le and the new length of the
> + * list with al le added.
> + */
> + sz = le_size(name_len);
> + new_size = al->size + sz;
> + asize = al_aligned(al->size);
> + new_asize = al_aligned(new_size);
> +
> + /* Scan forward to the point at which the new le should be inserted. */
> + le = al_find_le_to_insert(ni, type, name, name_len, svcn);
> + off = PtrOffset(al->le, le);
> +
> + if (new_size > asize) {
> + void *ptr = ntfs_malloc(new_asize);
> +
> + if (!ptr)
> + return -ENOMEM;
> +
> + memcpy(ptr, al->le, off);
> + memcpy(Add2Ptr(ptr, off + sz), le, al->size - off);
> + le = Add2Ptr(ptr, off);
> + ntfs_free(al->le);
> + al->le = ptr;
> + } else {
> + memmove(Add2Ptr(le, sz), le, al->size - off);
> + }
> +
> + al->size = new_size;
> +
> + le->type = type;
> + le->size = cpu_to_le16(sz);
> + le->name_len = name_len;
> + le->name_off = offsetof(struct ATTR_LIST_ENTRY, name);
> + le->vcn = cpu_to_le64(svcn);
> + le->ref = *ref;
> + le->id = id;
> + memcpy(le->name, name, sizeof(short) * name_len);
> +
> + al->dirty = true;
> +
> + err = attr_set_size(ni, ATTR_LIST, NULL, 0, &al->run, new_size,
> + &new_size, true, &attr);
> + if (err)
> + return err;
> +
> + if (attr && attr->non_res) {
> + err = ntfs_sb_write_run(ni->mi.sbi, &al->run, 0, al->le,
> + al->size);
> + if (err)
> + return err;
> + }
> +
> + al->dirty = false;
> + *new_le = le;
> +
> + return 0;
> +}
> +
> +/*
> + * al_remove_le
> + *
> + * removes 'le' from attribute list
> + */
> +bool al_remove_le(struct ntfs_inode *ni, struct ATTR_LIST_ENTRY *le)
> +{
> + u16 size;
> + size_t off;
> + typeof(ni->attr_list) *al = &ni->attr_list;
> +
> + if (!al_is_valid_le(ni, le))
> + return false;
> +
> + /* Save on stack the size of le */
> + size = le16_to_cpu(le->size);
> + off = PtrOffset(al->le, le);
> +
> + memmove(le, Add2Ptr(le, size), al->size - (off + size));
> +
> + al->size -= size;
> + al->dirty = true;
> +
> + return true;
> +}
> +
> +/*
> + * al_delete_le
> + *
> + * deletes from the list the first le which matches its parameters.
> + */
> +bool al_delete_le(struct ntfs_inode *ni, enum ATTR_TYPE type, CLST vcn,
> + const __le16 *name, size_t name_len,
> + const struct MFT_REF *ref)
> +{
> + u16 size;
> + struct ATTR_LIST_ENTRY *le;
> + size_t off;
> + typeof(ni->attr_list) *al = &ni->attr_list;
> +
> + /* Scan forward to the first le that matches the input */
> + le = al_find_ex(ni, NULL, type, name, name_len, &vcn);
> + if (!le)
> + return false;
> +
> + off = PtrOffset(al->le, le);
> +
> +next:
> + if (off >= al->size)
> + return false;
> + if (le->type != type)
> + return false;
> + if (le->name_len != name_len)
> + return false;
> + if (name_len && ntfs_cmp_names(le_name(le), name_len, name, name_len,
> + ni->mi.sbi->upcase, true))
> + return false;
> + if (le64_to_cpu(le->vcn) != vcn)
> + return false;
> +
> + /*
> + * The caller specified a segment reference, so we have to
> + * scan through the matching entries until we find that segment
> + * reference or we run of matching entries.
> + */
> + if (ref && memcmp(ref, &le->ref, sizeof(*ref))) {
> + off += le16_to_cpu(le->size);
> + le = Add2Ptr(al->le, off);
> + goto next;
> + }
> +
> + /* Save on stack the size of le */
> + size = le16_to_cpu(le->size);
> + /* Delete the le. */
> + memmove(le, Add2Ptr(le, size), al->size - (off + size));
> +
> + al->size -= size;
> + al->dirty = true;
> +
> + return true;
> +}
> +
> +/*
> + * al_update
> + */
> +int al_update(struct ntfs_inode *ni)
> +{
> + int err;
> + struct ATTRIB *attr;
> + typeof(ni->attr_list) *al = &ni->attr_list;
> +
> + if (!al->dirty || !al->size)
> + return 0;
> +
> + /*
> + * attribute list increased on demand in al_add_le
> + * attribute list decreased here
> + */
> + err = attr_set_size(ni, ATTR_LIST, NULL, 0, &al->run, al->size, NULL,
> + false, &attr);
> + if (err)
> + goto out;
> +
> + if (!attr->non_res) {
> + memcpy(resident_data(attr), al->le, al->size);
> + } else {
> + err = ntfs_sb_write_run(ni->mi.sbi, &al->run, 0, al->le,
> + al->size);
> + if (err)
> + goto out;
> +
> + attr->nres.valid_size = attr->nres.data_size;
> + }
> +
> + ni->mi.dirty = true;
> + al->dirty = false;
> +
> +out:
> + return err;
> +}
> diff --git a/fs/ntfs3/xattr.c b/fs/ntfs3/xattr.c
> new file mode 100644
> index 000000000000..ffe329315b12
> --- /dev/null
> +++ b/fs/ntfs3/xattr.c
> @@ -0,0 +1,1050 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + *
> + * Copyright (C) 2019-2021 Paragon Software GmbH, All rights reserved.
> + *
> + */
> +
> +#include <linux/blkdev.h>
> +#include <linux/buffer_head.h>
> +#include <linux/fs.h>
> +#include <linux/nls.h>
> +#include <linux/posix_acl.h>
> +#include <linux/posix_acl_xattr.h>
> +#include <linux/xattr.h>
> +
> +#include "debug.h"
> +#include "ntfs.h"
> +#include "ntfs_fs.h"
> +
> +// clang-format off
> +#define SYSTEM_DOS_ATTRIB "system.dos_attrib"
> +#define SYSTEM_NTFS_ATTRIB "system.ntfs_attrib"
> +#define SYSTEM_NTFS_SECURITY "system.ntfs_security"
> +// clang-format on
> +
> +static inline size_t unpacked_ea_size(const struct EA_FULL *ea)
> +{
> + return !ea->size ? DwordAlign(offsetof(struct EA_FULL, name) + 1 +
> + ea->name_len + le16_to_cpu(ea->elength)) :
> + le32_to_cpu(ea->size);
> +}
> +
> +static inline size_t packed_ea_size(const struct EA_FULL *ea)
> +{
> + return offsetof(struct EA_FULL, name) + 1 -
> + offsetof(struct EA_FULL, flags) + ea->name_len +
> + le16_to_cpu(ea->elength);
> +}
> +
> +/*
> + * find_ea
> + *
> + * assume there is at least one xattr in the list
> + */
> +static inline bool find_ea(const struct EA_FULL *ea_all, u32 bytes,
> + const char *name, u8 name_len, u32 *off)
> +{
> + *off = 0;
> +
> + if (!ea_all || !bytes)
> + return false;
> +
> + for (;;) {
> + const struct EA_FULL *ea = Add2Ptr(ea_all, *off);
> + u32 next_off = *off + unpacked_ea_size(ea);
> +
> + if (next_off > bytes)
> + return false;
> +
> + if (ea->name_len == name_len &&
> + !memcmp(ea->name, name, name_len))
> + return true;
> +
> + *off = next_off;
> + if (next_off >= bytes)
> + return false;
> + }
> +}
> +
> +/*
> + * ntfs_read_ea
> + *
> + * reads all extended attributes
> + * ea - new allocated memory
> + * info - pointer into resident data
> + */
> +static int ntfs_read_ea(struct ntfs_inode *ni, struct EA_FULL **ea,
> + size_t add_bytes, const struct EA_INFO **info)
> +{
> + int err;
> + struct ATTR_LIST_ENTRY *le = NULL;
> + struct ATTRIB *attr_info, *attr_ea;
> + void *ea_p;
> + u32 size;
> +
> + static_assert(le32_to_cpu(ATTR_EA_INFO) < le32_to_cpu(ATTR_EA));
> +
> + *ea = NULL;
> + *info = NULL;
> +
> + attr_info =
> + ni_find_attr(ni, NULL, &le, ATTR_EA_INFO, NULL, 0, NULL, NULL);
> + attr_ea =
> + ni_find_attr(ni, attr_info, &le, ATTR_EA, NULL, 0, NULL, NULL);
> +
> + if (!attr_ea || !attr_info)
> + return 0;
> +
> + *info = resident_data_ex(attr_info, sizeof(struct EA_INFO));
> + if (!*info)
> + return -EINVAL;
> +
> + /* Check Ea limit */
> + size = le32_to_cpu((*info)->size);
> + if (size > ni->mi.sbi->ea_max_size)
> + return -EFBIG;
> +
> + if (attr_size(attr_ea) > ni->mi.sbi->ea_max_size)
> + return -EFBIG;
> +
> + /* Allocate memory for packed Ea */
> + ea_p = ntfs_malloc(size + add_bytes);
> + if (!ea_p)
> + return -ENOMEM;
> +
> + if (attr_ea->non_res) {
> + struct runs_tree run;
> +
> + run_init(&run);
> +
> + err = attr_load_runs(attr_ea, ni, &run, NULL);
> + if (!err)
> + err = ntfs_read_run_nb(ni->mi.sbi, &run, 0, ea_p, size,
> + NULL);
> + run_close(&run);
> +
> + if (err)
> + goto out;
> + } else {
> + void *p = resident_data_ex(attr_ea, size);
> +
> + if (!p) {
> + err = -EINVAL;
> + goto out;
> + }
> + memcpy(ea_p, p, size);
> + }
> +
> + memset(Add2Ptr(ea_p, size), 0, add_bytes);
> + *ea = ea_p;
> + return 0;
> +
> +out:
> + ntfs_free(ea_p);
> + *ea = NULL;
> + return err;
> +}
> +
> +/*
> + * ntfs_list_ea
> + *
> + * copy a list of xattrs names into the buffer
> + * provided, or compute the buffer size required
> + */
> +static int ntfs_list_ea(struct ntfs_inode *ni, char *buffer,
> + size_t bytes_per_buffer, size_t *bytes)
> +{
> + const struct EA_INFO *info;
> + struct EA_FULL *ea_all = NULL;
> + const struct EA_FULL *ea;
> + u32 off, size;
> + int err;
> +
> + *bytes = 0;
> +
> + err = ntfs_read_ea(ni, &ea_all, 0, &info);
> + if (err)
> + return err;
> +
> + if (!info || !ea_all)
> + return 0;
> +
> + size = le32_to_cpu(info->size);
> +
> + /* Enumerate all xattrs */
> + for (off = 0; off < size; off += unpacked_ea_size(ea)) {
> + ea = Add2Ptr(ea_all, off);
> +
> + if (buffer) {
> + if (*bytes + ea->name_len + 1 > bytes_per_buffer) {
> + err = -ERANGE;
> + goto out;
> + }
> +
> + memcpy(buffer + *bytes, ea->name, ea->name_len);
> + buffer[*bytes + ea->name_len] = 0;
> + }
> +
> + *bytes += ea->name_len + 1;
> + }
> +
> +out:
> + ntfs_free(ea_all);
> + return err;
> +}
> +
> +static int ntfs_get_ea(struct inode *inode, const char *name, size_t name_len,
> + void *buffer, size_t size, size_t *required)
> +{
> + struct ntfs_inode *ni = ntfs_i(inode);
> + const struct EA_INFO *info;
> + struct EA_FULL *ea_all = NULL;
> + const struct EA_FULL *ea;
> + u32 off, len;
> + int err;
> +
> + if (!(ni->ni_flags & NI_FLAG_EA))
> + return -ENODATA;
> +
> + if (!required)
> + ni_lock(ni);
> +
> + len = 0;
> +
> + if (name_len > 255) {
> + err = -ENAMETOOLONG;
> + goto out;
> + }
> +
> + err = ntfs_read_ea(ni, &ea_all, 0, &info);
> + if (err)
> + goto out;
> +
> + if (!info)
> + goto out;
> +
> + /* Enumerate all xattrs */
> + if (!find_ea(ea_all, le32_to_cpu(info->size), name, name_len, &off)) {
> + err = -ENODATA;
> + goto out;
> + }
> + ea = Add2Ptr(ea_all, off);
> +
> + len = le16_to_cpu(ea->elength);
> + if (!buffer) {
> + err = 0;
> + goto out;
> + }
> +
> + if (len > size) {
> + err = -ERANGE;
> + if (required)
> + *required = len;
> + goto out;
> + }
> +
> + memcpy(buffer, ea->name + ea->name_len + 1, len);
> + err = 0;
> +
> +out:
> + ntfs_free(ea_all);
> + if (!required)
> + ni_unlock(ni);
> +
> + return err ? err : len;
> +}
> +
> +static noinline int ntfs_set_ea(struct inode *inode, const char *name,
> + size_t name_len, const void *value,
> + size_t val_size, int flags, int locked)
> +{
> + struct ntfs_inode *ni = ntfs_i(inode);
> + struct ntfs_sb_info *sbi = ni->mi.sbi;
> + int err;
> + struct EA_INFO ea_info;
> + const struct EA_INFO *info;
> + struct EA_FULL *new_ea;
> + struct EA_FULL *ea_all = NULL;
> + size_t add, new_pack;
> + u32 off, size;
> + __le16 size_pack;
> + struct ATTRIB *attr;
> + struct ATTR_LIST_ENTRY *le;
> + struct mft_inode *mi;
> + struct runs_tree ea_run;
> + u64 new_sz;
> + void *p;
> +
> + if (!locked)
> + ni_lock(ni);
> +
> + run_init(&ea_run);
> +
> + if (name_len > 255) {
> + err = -ENAMETOOLONG;
> + goto out;
> + }
> +
> + add = DwordAlign(offsetof(struct EA_FULL, name) + 1 + name_len +
> + val_size);
> +
> + err = ntfs_read_ea(ni, &ea_all, add, &info);
> + if (err)
> + goto out;
> +
> + if (!info) {
> + memset(&ea_info, 0, sizeof(ea_info));
> + size = 0;
> + size_pack = 0;
> + } else {
> + memcpy(&ea_info, info, sizeof(ea_info));
> + size = le32_to_cpu(ea_info.size);
> + size_pack = ea_info.size_pack;
> + }
> +
> + if (info && find_ea(ea_all, size, name, name_len, &off)) {
> + struct EA_FULL *ea;
> + size_t ea_sz;
> +
> + if (flags & XATTR_CREATE) {
> + err = -EEXIST;
> + goto out;
> + }
> +
> + /* Remove current xattr */
> + ea = Add2Ptr(ea_all, off);
> + if (ea->flags & FILE_NEED_EA)
> + le16_add_cpu(&ea_info.count, -1);
> +
> + ea_sz = unpacked_ea_size(ea);
> +
> + le16_add_cpu(&ea_info.size_pack, 0 - packed_ea_size(ea));
> +
> + memmove(ea, Add2Ptr(ea, ea_sz), size - off - ea_sz);
> +
> + size -= ea_sz;
> + memset(Add2Ptr(ea_all, size), 0, ea_sz);
> +
> + ea_info.size = cpu_to_le32(size);
> +
> + if ((flags & XATTR_REPLACE) && !val_size)
> + goto update_ea;
> + } else {
> + if (flags & XATTR_REPLACE) {
> + err = -ENODATA;
> + goto out;
> + }
> +
> + if (!ea_all) {
> + ea_all = ntfs_zalloc(add);
> + if (!ea_all) {
> + err = -ENOMEM;
> + goto out;
> + }
> + }
> + }
> +
> + /* append new xattr */
> + new_ea = Add2Ptr(ea_all, size);
> + new_ea->size = cpu_to_le32(add);
> + new_ea->flags = 0;
> + new_ea->name_len = name_len;
> + new_ea->elength = cpu_to_le16(val_size);
> + memcpy(new_ea->name, name, name_len);
> + new_ea->name[name_len] = 0;
> + memcpy(new_ea->name + name_len + 1, value, val_size);
> + new_pack = le16_to_cpu(ea_info.size_pack) + packed_ea_size(new_ea);
> +
> + /* should fit into 16 bits */
> + if (new_pack > 0xffff) {
> + err = -EFBIG; // -EINVAL?
> + goto out;
> + }
> + ea_info.size_pack = cpu_to_le16(new_pack);
> +
> + /* new size of ATTR_EA */
> + size += add;
> + if (size > sbi->ea_max_size) {
> + err = -EFBIG; // -EINVAL?
> + goto out;
> + }
> + ea_info.size = cpu_to_le32(size);
> +
> +update_ea:
> +
> + if (!info) {
> + /* Create xattr */
> + if (!size) {
> + err = 0;
> + goto out;
> + }
> +
> + err = ni_insert_resident(ni, sizeof(struct EA_INFO),
> + ATTR_EA_INFO, NULL, 0, NULL, NULL);
> + if (err)
> + goto out;
> +
> + err = ni_insert_resident(ni, 0, ATTR_EA, NULL, 0, NULL, NULL);
> + if (err)
> + goto out;
> + }
> +
> + new_sz = size;
> + err = attr_set_size(ni, ATTR_EA, NULL, 0, &ea_run, new_sz, &new_sz,
> + false, NULL);
> + if (err)
> + goto out;
> +
> + le = NULL;
> + attr = ni_find_attr(ni, NULL, &le, ATTR_EA_INFO, NULL, 0, NULL, &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (!size) {
> + /* delete xattr, ATTR_EA_INFO */
> + err = ni_remove_attr_le(ni, attr, le);
> + if (err)
> + goto out;
> + } else {
> + p = resident_data_ex(attr, sizeof(struct EA_INFO));
> + if (!p) {
> + err = -EINVAL;
> + goto out;
> + }
> + memcpy(p, &ea_info, sizeof(struct EA_INFO));
> + mi->dirty = true;
> + }
> +
> + le = NULL;
> + attr = ni_find_attr(ni, NULL, &le, ATTR_EA, NULL, 0, NULL, &mi);
> + if (!attr) {
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (!size) {
> + /* delete xattr, ATTR_EA */
> + err = ni_remove_attr_le(ni, attr, le);
> + if (err)
> + goto out;
> + } else if (attr->non_res) {
> + err = ntfs_sb_write_run(sbi, &ea_run, 0, ea_all, size);
> + if (err)
> + goto out;
> + } else {
> + p = resident_data_ex(attr, size);
> + if (!p) {
> + err = -EINVAL;
> + goto out;
> + }
> + memcpy(p, ea_all, size);
> + mi->dirty = true;
> + }
> +
> + if (ea_info.size_pack != size_pack)
> + ni->ni_flags |= NI_FLAG_UPDATE_PARENT;
> + mark_inode_dirty(&ni->vfs_inode);
> +
> + /* Check if we delete the last xattr */
> + if (val_size || flags != XATTR_REPLACE ||
> + ntfs_list_ea(ni, NULL, 0, &val_size) || val_size) {
> + ni->ni_flags |= NI_FLAG_EA;
> + } else {
> + ni->ni_flags &= ~NI_FLAG_EA;
> + }
> +
> +out:
> + if (!locked)
> + ni_unlock(ni);
> +
> + run_close(&ea_run);
> + ntfs_free(ea_all);
> +
> + return err;
> +}
> +
> +#ifdef CONFIG_NTFS3_FS_POSIX_ACL
> +static inline void ntfs_posix_acl_release(struct posix_acl *acl)
> +{
> + if (acl && refcount_dec_and_test(&acl->a_refcount))
> + kfree(acl);
> +}
> +
> +static struct posix_acl *ntfs_get_acl_ex(struct user_namespace *mnt_userns,
> + struct inode *inode, int type,
> + int locked)
> +{
> + struct ntfs_inode *ni = ntfs_i(inode);
> + const char *name;
> + size_t name_len;
> + struct posix_acl *acl;
> + size_t req;
> + int err;
> + void *buf;
> +
> + /* allocate PATH_MAX bytes */
> + buf = __getname();
> + if (!buf)
> + return ERR_PTR(-ENOMEM);
> +
> + /* Possible values of 'type' was already checked above */
> + if (type == ACL_TYPE_ACCESS) {
> + name = XATTR_NAME_POSIX_ACL_ACCESS;
> + name_len = sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1;
> + } else {
> + name = XATTR_NAME_POSIX_ACL_DEFAULT;
> + name_len = sizeof(XATTR_NAME_POSIX_ACL_DEFAULT) - 1;
> + }
> +
> + if (!locked)
> + ni_lock(ni);
> +
> + err = ntfs_get_ea(inode, name, name_len, buf, PATH_MAX, &req);
> +
> + if (!locked)
> + ni_unlock(ni);
> +
> + /* Translate extended attribute to acl */
> + if (err > 0) {
> + acl = posix_acl_from_xattr(mnt_userns, buf, err);
> + if (!IS_ERR(acl))
> + set_cached_acl(inode, type, acl);
> + } else {
> + acl = err == -ENODATA ? NULL : ERR_PTR(err);
> + }
> +
> + __putname(buf);
> +
> + return acl;
> +}
> +
> +/*
> + * ntfs_get_acl
> + *
> + * inode_operations::get_acl
> + */
> +struct posix_acl *ntfs_get_acl(struct inode *inode, int type)
> +{
> + /* TODO: init_user_ns? */
> + return ntfs_get_acl_ex(&init_user_ns, inode, type, 0);
> +}
> +
> +static noinline int ntfs_set_acl_ex(struct user_namespace *mnt_userns,
> + struct inode *inode, struct posix_acl *acl,
> + int type, int locked)
> +{
> + const char *name;
> + size_t size, name_len;
> + void *value = NULL;
> + int err = 0;
> +
> + if (S_ISLNK(inode->i_mode))
> + return -EOPNOTSUPP;
> +
> + switch (type) {
> + case ACL_TYPE_ACCESS:
> + if (acl) {
> + umode_t mode = inode->i_mode;
> +
> + err = posix_acl_equiv_mode(acl, &mode);
> + if (err < 0)
> + return err;
> +
> + if (inode->i_mode != mode) {
> + inode->i_mode = mode;
> + mark_inode_dirty(inode);
> + }
> +
> + if (!err) {
> + /*
> + * acl can be exactly represented in the
> + * traditional file mode permission bits
> + */
> + acl = NULL;
> + goto out;
> + }
> + }
> + name = XATTR_NAME_POSIX_ACL_ACCESS;
> + name_len = sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1;
> + break;
> +
> + case ACL_TYPE_DEFAULT:
> + if (!S_ISDIR(inode->i_mode))
> + return acl ? -EACCES : 0;
> + name = XATTR_NAME_POSIX_ACL_DEFAULT;
> + name_len = sizeof(XATTR_NAME_POSIX_ACL_DEFAULT) - 1;
> + break;
> +
> + default:
> + return -EINVAL;
> + }
> +
> + if (!acl)
> + goto out;
> +
> + size = posix_acl_xattr_size(acl->a_count);
> + value = ntfs_malloc(size);
> + if (!value)
> + return -ENOMEM;
> +
> + err = posix_acl_to_xattr(mnt_userns, acl, value, size);
> + if (err)
> + goto out;
> +
> + err = ntfs_set_ea(inode, name, name_len, value, size, 0, locked);
> + if (err)
> + goto out;
> +
> + inode->i_flags &= ~S_NOSEC;
> +
> +out:
> + if (!err)
> + set_cached_acl(inode, type, acl);
> +
> + kfree(value);
> +
> + return err;
> +}
> +
> +/*
> + * ntfs_set_acl
> + *
> + * inode_operations::set_acl
> + */
> +int ntfs_set_acl(struct user_namespace *mnt_userns, struct inode *inode,
> + struct posix_acl *acl, int type)
> +{
> + return ntfs_set_acl_ex(mnt_userns, inode, acl, type, 0);
> +}
> +
> +static int ntfs_xattr_get_acl(struct user_namespace *mnt_userns,
> + struct inode *inode, int type, void *buffer,
> + size_t size)
> +{
> + struct posix_acl *acl;
> + int err;
> +
> + if (!(inode->i_sb->s_flags & SB_POSIXACL))
> + return -EOPNOTSUPP;
> +
> + acl = ntfs_get_acl(inode, type);
> + if (IS_ERR(acl))
> + return PTR_ERR(acl);
> +
> + if (!acl)
> + return -ENODATA;
> +
> + err = posix_acl_to_xattr(mnt_userns, acl, buffer, size);
> + ntfs_posix_acl_release(acl);
> +
> + return err;
> +}
> +
> +static int ntfs_xattr_set_acl(struct user_namespace *mnt_userns,
> + struct inode *inode, int type, const void *value,
> + size_t size)
> +{
> + struct posix_acl *acl;
> + int err;
> +
> + if (!(inode->i_sb->s_flags & SB_POSIXACL))
> + return -EOPNOTSUPP;
> +
> + if (!inode_owner_or_capable(mnt_userns, inode))
> + return -EPERM;
> +
> + if (!value)
> + return 0;
> +
> + acl = posix_acl_from_xattr(mnt_userns, value, size);
> + if (IS_ERR(acl))
> + return PTR_ERR(acl);
> +
> + if (acl) {
> + err = posix_acl_valid(mnt_userns, acl);
> + if (err)
> + goto release_and_out;
> + }
> +
> + err = ntfs_set_acl(mnt_userns, inode, acl, type);
> +
> +release_and_out:
> + ntfs_posix_acl_release(acl);
> + return err;
> +}
> +
> +/*
> + * Initialize the ACLs of a new inode. Called from ntfs_create_inode.
> + */
> +int ntfs_init_acl(struct user_namespace *mnt_userns, struct inode *inode,
> + struct inode *dir)
> +{
> + struct posix_acl *default_acl, *acl;
> + int err;
> +
> + /*
> + * TODO refactoring lock
> + * ni_lock(dir) ... -> posix_acl_create(dir,...) -> ntfs_get_acl -> ni_lock(dir)
> + */
> + inode->i_default_acl = NULL;
> +
> + default_acl = ntfs_get_acl_ex(mnt_userns, dir, ACL_TYPE_DEFAULT, 1);
> +
> + if (!default_acl || default_acl == ERR_PTR(-EOPNOTSUPP)) {
> + inode->i_mode &= ~current_umask();
> + err = 0;
> + goto out;
> + }
> +
> + if (IS_ERR(default_acl)) {
> + err = PTR_ERR(default_acl);
> + goto out;
> + }
> +
> + acl = default_acl;
> + err = __posix_acl_create(&acl, GFP_NOFS, &inode->i_mode);
> + if (err < 0)
> + goto out1;
> + if (!err) {
> + posix_acl_release(acl);
> + acl = NULL;
> + }
> +
> + if (!S_ISDIR(inode->i_mode)) {
> + posix_acl_release(default_acl);
> + default_acl = NULL;
> + }
> +
> + if (default_acl)
> + err = ntfs_set_acl_ex(mnt_userns, inode, default_acl,
> + ACL_TYPE_DEFAULT, 1);
> +
> + if (!acl)
> + inode->i_acl = NULL;
> + else if (!err)
> + err = ntfs_set_acl_ex(mnt_userns, inode, acl, ACL_TYPE_ACCESS,
> + 1);
> +
> + posix_acl_release(acl);
> +out1:
> + posix_acl_release(default_acl);
> +
> +out:
> + return err;
> +}
> +#endif
> +
> +/*
> + * ntfs_acl_chmod
> + *
> + * helper for 'ntfs3_setattr'
> + */
> +int ntfs_acl_chmod(struct user_namespace *mnt_userns, struct inode *inode)
> +{
> + struct super_block *sb = inode->i_sb;
> +
> + if (!(sb->s_flags & SB_POSIXACL))
> + return 0;
> +
> + if (S_ISLNK(inode->i_mode))
> + return -EOPNOTSUPP;
> +
> + return posix_acl_chmod(mnt_userns, inode, inode->i_mode);
> +}
> +
> +/*
> + * ntfs_permission
> + *
> + * inode_operations::permission
> + */
> +int ntfs_permission(struct user_namespace *mnt_userns, struct inode *inode,
> + int mask)
> +{
> + if (ntfs_sb(inode->i_sb)->options.no_acs_rules) {
> + /* "no access rules" mode - allow all changes */
> + return 0;
> + }
> +
> + return generic_permission(mnt_userns, inode, mask);
> +}
> +
> +/*
> + * ntfs_listxattr
> + *
> + * inode_operations::listxattr
> + */
> +ssize_t ntfs_listxattr(struct dentry *dentry, char *buffer, size_t size)
> +{
> + struct inode *inode = d_inode(dentry);
> + struct ntfs_inode *ni = ntfs_i(inode);
> + ssize_t ret = -1;
> + int err;
> +
> + if (!(ni->ni_flags & NI_FLAG_EA)) {
> + ret = 0;
> + goto out;
> + }
> +
> + ni_lock(ni);
> +
> + err = ntfs_list_ea(ni, buffer, size, (size_t *)&ret);
> +
> + ni_unlock(ni);
> +
> + if (err)
> + ret = err;
> +out:
> + return ret;
> +}
> +
> +static int ntfs_getxattr(const struct xattr_handler *handler, struct dentry *de,
> + struct inode *inode, const char *name, void *buffer,
> + size_t size)
> +{
> + int err;
> + struct ntfs_inode *ni = ntfs_i(inode);
> + size_t name_len = strlen(name);
> +
> + /* Dispatch request */
> + if (name_len == sizeof(SYSTEM_DOS_ATTRIB) - 1 &&
> + !memcmp(name, SYSTEM_DOS_ATTRIB, sizeof(SYSTEM_DOS_ATTRIB))) {
> + /* system.dos_attrib */
> + if (!buffer) {
> + err = sizeof(u8);
> + } else if (size < sizeof(u8)) {
> + err = -ENODATA;
> + } else {
> + err = sizeof(u8);
> + *(u8 *)buffer = le32_to_cpu(ni->std_fa);
> + }
> + goto out;
> + }
> +
> + if (name_len == sizeof(SYSTEM_NTFS_ATTRIB) - 1 &&
> + !memcmp(name, SYSTEM_NTFS_ATTRIB, sizeof(SYSTEM_NTFS_ATTRIB))) {
> + /* system.ntfs_attrib */
> + if (!buffer) {
> + err = sizeof(u32);
> + } else if (size < sizeof(u32)) {
> + err = -ENODATA;
> + } else {
> + err = sizeof(u32);
> + *(u32 *)buffer = le32_to_cpu(ni->std_fa);
> + }
> + goto out;
> + }
> +
> + if (name_len == sizeof(SYSTEM_NTFS_SECURITY) - 1 &&
> + !memcmp(name, SYSTEM_NTFS_SECURITY, sizeof(SYSTEM_NTFS_SECURITY))) {
> + /* system.ntfs_security*/
> + struct SECURITY_DESCRIPTOR_RELATIVE *sd = NULL;
> + size_t sd_size = 0;
> +
> + if (!is_ntfs3(ni->mi.sbi)) {
> + /* we should get nt4 security */
> + err = -EINVAL;
> + goto out;
> + } else if (le32_to_cpu(ni->std_security_id) <
> + SECURITY_ID_FIRST) {
> + err = -ENOENT;
> + goto out;
> + }
> +
> + err = ntfs_get_security_by_id(ni->mi.sbi, ni->std_security_id,
> + &sd, &sd_size);
> + if (err)
> + goto out;
> +
> + if (!is_sd_valid(sd, sd_size)) {
> + ntfs_inode_warn(
> + inode,
> + "looks like you get incorrect security descriptor id=%u",
> + ni->std_security_id);
> + }
> +
> + if (!buffer) {
> + err = sd_size;
> + } else if (size < sd_size) {
> + err = -ENODATA;
> + } else {
> + err = sd_size;
> + memcpy(buffer, sd, sd_size);
> + }
> + ntfs_free(sd);
> + goto out;
> + }
> +
> +#ifdef CONFIG_NTFS3_FS_POSIX_ACL
> + if ((name_len == sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1 &&
> + !memcmp(name, XATTR_NAME_POSIX_ACL_ACCESS,
> + sizeof(XATTR_NAME_POSIX_ACL_ACCESS))) ||
> + (name_len == sizeof(XATTR_NAME_POSIX_ACL_DEFAULT) - 1 &&
> + !memcmp(name, XATTR_NAME_POSIX_ACL_DEFAULT,
> + sizeof(XATTR_NAME_POSIX_ACL_DEFAULT)))) {
> + /* TODO: init_user_ns? */
> + err = ntfs_xattr_get_acl(
> + &init_user_ns, inode,
> + name_len == sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1 ?
> + ACL_TYPE_ACCESS :
> + ACL_TYPE_DEFAULT,
> + buffer, size);
> + goto out;
> + }
> +#endif
> + /* deal with ntfs extended attribute */
> + err = ntfs_get_ea(inode, name, name_len, buffer, size, NULL);
> +
> +out:
> + return err;
> +}
> +
> +/*
> + * ntfs_setxattr
> + *
> + * inode_operations::setxattr
> + */
> +static noinline int ntfs_setxattr(const struct xattr_handler *handler,
> + struct user_namespace *mnt_userns,
> + struct dentry *de, struct inode *inode,
> + const char *name, const void *value,
> + size_t size, int flags)
> +{
> + int err = -EINVAL;
> + struct ntfs_inode *ni = ntfs_i(inode);
> + size_t name_len = strlen(name);
> + enum FILE_ATTRIBUTE new_fa;
> +
> + /* Dispatch request */
> + if (name_len == sizeof(SYSTEM_DOS_ATTRIB) - 1 &&
> + !memcmp(name, SYSTEM_DOS_ATTRIB, sizeof(SYSTEM_DOS_ATTRIB))) {
> + if (sizeof(u8) != size)
> + goto out;
> + new_fa = cpu_to_le32(*(u8 *)value);
> + goto set_new_fa;
> + }
> +
> + if (name_len == sizeof(SYSTEM_NTFS_ATTRIB) - 1 &&
> + !memcmp(name, SYSTEM_NTFS_ATTRIB, sizeof(SYSTEM_NTFS_ATTRIB))) {
> + if (size != sizeof(u32))
> + goto out;
> + new_fa = cpu_to_le32(*(u32 *)value);
> +
> + if (S_ISREG(inode->i_mode)) {
> + /* Process compressed/sparsed in special way*/
> + ni_lock(ni);
> + err = ni_new_attr_flags(ni, new_fa);
> + ni_unlock(ni);
> + if (err)
> + goto out;
> + }
> +set_new_fa:
> + /*
> + * Thanks Mark Harmstone:
> + * keep directory bit consistency
> + */
> + if (S_ISDIR(inode->i_mode))
> + new_fa |= FILE_ATTRIBUTE_DIRECTORY;
> + else
> + new_fa &= ~FILE_ATTRIBUTE_DIRECTORY;
> +
> + if (ni->std_fa != new_fa) {
> + ni->std_fa = new_fa;
> + if (new_fa & FILE_ATTRIBUTE_READONLY)
> + inode->i_mode &= ~0222;
> + else
> + inode->i_mode |= 0222;
> + /* std attribute always in primary record */
> + ni->mi.dirty = true;
> + mark_inode_dirty(inode);
> + }
> + err = 0;
> +
> + goto out;
> + }
> +
> + if (name_len == sizeof(SYSTEM_NTFS_SECURITY) - 1 &&
> + !memcmp(name, SYSTEM_NTFS_SECURITY, sizeof(SYSTEM_NTFS_SECURITY))) {
> + /* system.ntfs_security*/
> + __le32 security_id;
> + bool inserted;
> + struct ATTR_STD_INFO5 *std;
> +
> + if (!is_ntfs3(ni->mi.sbi)) {
> + /*
> + * we should replace ATTR_SECURE
> + * Skip this way cause it is nt4 feature
> + */
> + err = -EINVAL;
> + goto out;
> + }
> +
> + if (!is_sd_valid(value, size)) {
> + err = -EINVAL;
> + ntfs_inode_warn(
> + inode,
> + "you try to set invalid security descriptor");
> + goto out;
> + }
> +
> + err = ntfs_insert_security(ni->mi.sbi, value, size,
> + &security_id, &inserted);
> + if (err)
> + goto out;
> +
> + ni_lock(ni);
> + std = ni_std5(ni);
> + if (!std) {
> + err = -EINVAL;
> + } else if (std->security_id != security_id) {
> + std->security_id = ni->std_security_id = security_id;
> + /* std attribute always in primary record */
> + ni->mi.dirty = true;
> + mark_inode_dirty(&ni->vfs_inode);
> + }
> + ni_unlock(ni);
> + goto out;
> + }
> +
> +#ifdef CONFIG_NTFS3_FS_POSIX_ACL
> + if ((name_len == sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1 &&
> + !memcmp(name, XATTR_NAME_POSIX_ACL_ACCESS,
> + sizeof(XATTR_NAME_POSIX_ACL_ACCESS))) ||
> + (name_len == sizeof(XATTR_NAME_POSIX_ACL_DEFAULT) - 1 &&
> + !memcmp(name, XATTR_NAME_POSIX_ACL_DEFAULT,
> + sizeof(XATTR_NAME_POSIX_ACL_DEFAULT)))) {
> + /* TODO: init_user_ns? */
> + err = ntfs_xattr_set_acl(
> + &init_user_ns, inode,
> + name_len == sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1 ?
> + ACL_TYPE_ACCESS :
> + ACL_TYPE_DEFAULT,
> + value, size);
> + goto out;
> + }
> +#endif
> + /* deal with ntfs extended attribute */
> + err = ntfs_set_ea(inode, name, name_len, value, size, flags, 0);
> +
> +out:
> + return err;
> +}
> +
> +static bool ntfs_xattr_user_list(struct dentry *dentry)
> +{
> + return 1;
> +}
> +
> +static const struct xattr_handler ntfs_xattr_handler = {
> + .prefix = "",
> + .get = ntfs_getxattr,
> + .set = ntfs_setxattr,
> + .list = ntfs_xattr_user_list,
> +};
> +
> +const struct xattr_handler *ntfs_xattr_handlers[] = {
> + &ntfs_xattr_handler,
> + NULL,
> +};