mirror of
https://github.com/torvalds/linux.git
synced 2026-04-18 06:44:00 -04:00
Add bound validation in ntfs_external_attr_find to prevent out-of-bounds memory accesses. This ensures that the attribute record's length, name offset, and both resident and non-resident value offsets strictly fall within the safe boundaries of the MFT record. Signed-off-by: Hyunchul Lee <hyc.lee@gmail.com> Signed-off-by: Namjae Jeon <linkinjeon@kernel.org>
5540 lines
163 KiB
C
5540 lines
163 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* NTFS attribute operations.
|
|
*
|
|
* Copyright (c) 2001-2012 Anton Altaparmakov and Tuxera Inc.
|
|
* Copyright (c) 2002 Richard Russon
|
|
* Copyright (c) 2025 LG Electronics Co., Ltd.
|
|
*
|
|
* Part of this file is based on code from the NTFS-3G.
|
|
* and is copyrighted by the respective authors below:
|
|
* Copyright (c) 2000-2010 Anton Altaparmakov
|
|
* Copyright (c) 2002-2005 Richard Russon
|
|
* Copyright (c) 2002-2008 Szabolcs Szakacsits
|
|
* Copyright (c) 2004-2007 Yura Pakhuchiy
|
|
* Copyright (c) 2007-2021 Jean-Pierre Andre
|
|
* Copyright (c) 2010 Erik Larsson
|
|
*/
|
|
|
|
#include <linux/writeback.h>
|
|
#include <linux/iomap.h>
|
|
|
|
#include "attrib.h"
|
|
#include "attrlist.h"
|
|
#include "lcnalloc.h"
|
|
#include "debug.h"
|
|
#include "mft.h"
|
|
#include "ntfs.h"
|
|
#include "iomap.h"
|
|
|
|
__le16 AT_UNNAMED[] = { cpu_to_le16('\0') };
|
|
|
|
/*
|
|
* ntfs_map_runlist_nolock - map (a part of) a runlist of an ntfs inode
|
|
* @ni: ntfs inode for which to map (part of) a runlist
|
|
* @vcn: map runlist part containing this vcn
|
|
* @ctx: active attribute search context if present or NULL if not
|
|
*
|
|
* Map the part of a runlist containing the @vcn of the ntfs inode @ni.
|
|
*
|
|
* If @ctx is specified, it is an active search context of @ni and its base mft
|
|
* record. This is needed when ntfs_map_runlist_nolock() encounters unmapped
|
|
* runlist fragments and allows their mapping. If you do not have the mft
|
|
* record mapped, you can specify @ctx as NULL and ntfs_map_runlist_nolock()
|
|
* will perform the necessary mapping and unmapping.
|
|
*
|
|
* Note, ntfs_map_runlist_nolock() saves the state of @ctx on entry and
|
|
* restores it before returning. Thus, @ctx will be left pointing to the same
|
|
* attribute on return as on entry. However, the actual pointers in @ctx may
|
|
* point to different memory locations on return, so you must remember to reset
|
|
* any cached pointers from the @ctx, i.e. after the call to
|
|
* ntfs_map_runlist_nolock(), you will probably want to do:
|
|
* m = ctx->mrec;
|
|
* a = ctx->attr;
|
|
* Assuming you cache ctx->attr in a variable @a of type struct attr_record *
|
|
* and that you cache ctx->mrec in a variable @m of type struct mft_record *.
|
|
*
|
|
* Return 0 on success and -errno on error. There is one special error code
|
|
* which is not an error as such. This is -ENOENT. It means that @vcn is out
|
|
* of bounds of the runlist.
|
|
*
|
|
* Note the runlist can be NULL after this function returns if @vcn is zero and
|
|
* the attribute has zero allocated size, i.e. there simply is no runlist.
|
|
*
|
|
* WARNING: If @ctx is supplied, regardless of whether success or failure is
|
|
* returned, you need to check IS_ERR(@ctx->mrec) and if 'true' the @ctx
|
|
* is no longer valid, i.e. you need to either call
|
|
* ntfs_attr_reinit_search_ctx() or ntfs_attr_put_search_ctx() on it.
|
|
* In that case PTR_ERR(@ctx->mrec) will give you the error code for
|
|
* why the mapping of the old inode failed.
|
|
*
|
|
* Locking: - The runlist described by @ni must be locked for writing on entry
|
|
* and is locked on return. Note the runlist will be modified.
|
|
* - If @ctx is NULL, the base mft record of @ni must not be mapped on
|
|
* entry and it will be left unmapped on return.
|
|
* - If @ctx is not NULL, the base mft record must be mapped on entry
|
|
* and it will be left mapped on return.
|
|
*/
|
|
int ntfs_map_runlist_nolock(struct ntfs_inode *ni, s64 vcn, struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
s64 end_vcn;
|
|
unsigned long flags;
|
|
struct ntfs_inode *base_ni;
|
|
struct mft_record *m;
|
|
struct attr_record *a;
|
|
struct runlist_element *rl;
|
|
struct folio *put_this_folio = NULL;
|
|
int err = 0;
|
|
bool ctx_is_temporary = false, ctx_needs_reset;
|
|
struct ntfs_attr_search_ctx old_ctx = { NULL, };
|
|
size_t new_rl_count;
|
|
|
|
ntfs_debug("Mapping runlist part containing vcn 0x%llx.",
|
|
(unsigned long long)vcn);
|
|
if (!NInoAttr(ni))
|
|
base_ni = ni;
|
|
else
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
if (!ctx) {
|
|
ctx_is_temporary = ctx_needs_reset = true;
|
|
m = map_mft_record(base_ni);
|
|
if (IS_ERR(m))
|
|
return PTR_ERR(m);
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, m);
|
|
if (unlikely(!ctx)) {
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
} else {
|
|
s64 allocated_size_vcn;
|
|
|
|
WARN_ON(IS_ERR(ctx->mrec));
|
|
a = ctx->attr;
|
|
if (!a->non_resident) {
|
|
err = -EIO;
|
|
goto err_out;
|
|
}
|
|
end_vcn = le64_to_cpu(a->data.non_resident.highest_vcn);
|
|
read_lock_irqsave(&ni->size_lock, flags);
|
|
allocated_size_vcn =
|
|
ntfs_bytes_to_cluster(ni->vol, ni->allocated_size);
|
|
read_unlock_irqrestore(&ni->size_lock, flags);
|
|
if (!a->data.non_resident.lowest_vcn && end_vcn <= 0)
|
|
end_vcn = allocated_size_vcn - 1;
|
|
/*
|
|
* If we already have the attribute extent containing @vcn in
|
|
* @ctx, no need to look it up again. We slightly cheat in
|
|
* that if vcn exceeds the allocated size, we will refuse to
|
|
* map the runlist below, so there is definitely no need to get
|
|
* the right attribute extent.
|
|
*/
|
|
if (vcn >= allocated_size_vcn || (a->type == ni->type &&
|
|
a->name_length == ni->name_len &&
|
|
!memcmp((u8 *)a + le16_to_cpu(a->name_offset),
|
|
ni->name, ni->name_len) &&
|
|
le64_to_cpu(a->data.non_resident.lowest_vcn)
|
|
<= vcn && end_vcn >= vcn))
|
|
ctx_needs_reset = false;
|
|
else {
|
|
/* Save the old search context. */
|
|
old_ctx = *ctx;
|
|
/*
|
|
* If the currently mapped (extent) inode is not the
|
|
* base inode we will unmap it when we reinitialize the
|
|
* search context which means we need to get a
|
|
* reference to the page containing the mapped mft
|
|
* record so we do not accidentally drop changes to the
|
|
* mft record when it has not been marked dirty yet.
|
|
*/
|
|
if (old_ctx.base_ntfs_ino && old_ctx.ntfs_ino !=
|
|
old_ctx.base_ntfs_ino) {
|
|
put_this_folio = old_ctx.ntfs_ino->folio;
|
|
folio_get(put_this_folio);
|
|
}
|
|
/*
|
|
* Reinitialize the search context so we can lookup the
|
|
* needed attribute extent.
|
|
*/
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
ctx_needs_reset = true;
|
|
}
|
|
}
|
|
if (ctx_needs_reset) {
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, vcn, NULL, 0, ctx);
|
|
if (unlikely(err)) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
goto err_out;
|
|
}
|
|
WARN_ON(!ctx->attr->non_resident);
|
|
}
|
|
a = ctx->attr;
|
|
/*
|
|
* Only decompress the mapping pairs if @vcn is inside it. Otherwise
|
|
* we get into problems when we try to map an out of bounds vcn because
|
|
* we then try to map the already mapped runlist fragment and
|
|
* ntfs_mapping_pairs_decompress() fails.
|
|
*/
|
|
end_vcn = le64_to_cpu(a->data.non_resident.highest_vcn) + 1;
|
|
if (unlikely(vcn && vcn >= end_vcn)) {
|
|
err = -ENOENT;
|
|
goto err_out;
|
|
}
|
|
rl = ntfs_mapping_pairs_decompress(ni->vol, a, &ni->runlist, &new_rl_count);
|
|
if (IS_ERR(rl))
|
|
err = PTR_ERR(rl);
|
|
else {
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_count;
|
|
}
|
|
err_out:
|
|
if (ctx_is_temporary) {
|
|
if (likely(ctx))
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
unmap_mft_record(base_ni);
|
|
} else if (ctx_needs_reset) {
|
|
/*
|
|
* If there is no attribute list, restoring the search context
|
|
* is accomplished simply by copying the saved context back over
|
|
* the caller supplied context. If there is an attribute list,
|
|
* things are more complicated as we need to deal with mapping
|
|
* of mft records and resulting potential changes in pointers.
|
|
*/
|
|
if (NInoAttrList(base_ni)) {
|
|
/*
|
|
* If the currently mapped (extent) inode is not the
|
|
* one we had before, we need to unmap it and map the
|
|
* old one.
|
|
*/
|
|
if (ctx->ntfs_ino != old_ctx.ntfs_ino) {
|
|
/*
|
|
* If the currently mapped inode is not the
|
|
* base inode, unmap it.
|
|
*/
|
|
if (ctx->base_ntfs_ino && ctx->ntfs_ino !=
|
|
ctx->base_ntfs_ino) {
|
|
unmap_extent_mft_record(ctx->ntfs_ino);
|
|
ctx->mrec = ctx->base_mrec;
|
|
WARN_ON(!ctx->mrec);
|
|
}
|
|
/*
|
|
* If the old mapped inode is not the base
|
|
* inode, map it.
|
|
*/
|
|
if (old_ctx.base_ntfs_ino &&
|
|
old_ctx.ntfs_ino != old_ctx.base_ntfs_ino) {
|
|
retry_map:
|
|
ctx->mrec = map_mft_record(old_ctx.ntfs_ino);
|
|
/*
|
|
* Something bad has happened. If out
|
|
* of memory retry till it succeeds.
|
|
* Any other errors are fatal and we
|
|
* return the error code in ctx->mrec.
|
|
* Let the caller deal with it... We
|
|
* just need to fudge things so the
|
|
* caller can reinit and/or put the
|
|
* search context safely.
|
|
*/
|
|
if (IS_ERR(ctx->mrec)) {
|
|
if (PTR_ERR(ctx->mrec) == -ENOMEM) {
|
|
schedule();
|
|
goto retry_map;
|
|
} else
|
|
old_ctx.ntfs_ino =
|
|
old_ctx.base_ntfs_ino;
|
|
}
|
|
}
|
|
}
|
|
/* Update the changed pointers in the saved context. */
|
|
if (ctx->mrec != old_ctx.mrec) {
|
|
if (!IS_ERR(ctx->mrec))
|
|
old_ctx.attr = (struct attr_record *)(
|
|
(u8 *)ctx->mrec +
|
|
((u8 *)old_ctx.attr -
|
|
(u8 *)old_ctx.mrec));
|
|
old_ctx.mrec = ctx->mrec;
|
|
}
|
|
}
|
|
/* Restore the search context to the saved one. */
|
|
*ctx = old_ctx;
|
|
/*
|
|
* We drop the reference on the page we took earlier. In the
|
|
* case that IS_ERR(ctx->mrec) is true this means we might lose
|
|
* some changes to the mft record that had been made between
|
|
* the last time it was marked dirty/written out and now. This
|
|
* at this stage is not a problem as the mapping error is fatal
|
|
* enough that the mft record cannot be written out anyway and
|
|
* the caller is very likely to shutdown the whole inode
|
|
* immediately and mark the volume dirty for chkdsk to pick up
|
|
* the pieces anyway.
|
|
*/
|
|
if (put_this_folio)
|
|
folio_put(put_this_folio);
|
|
}
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_map_runlist - map (a part of) a runlist of an ntfs inode
|
|
* @ni: ntfs inode for which to map (part of) a runlist
|
|
* @vcn: map runlist part containing this vcn
|
|
*
|
|
* Map the part of a runlist containing the @vcn of the ntfs inode @ni.
|
|
*
|
|
* Return 0 on success and -errno on error. There is one special error code
|
|
* which is not an error as such. This is -ENOENT. It means that @vcn is out
|
|
* of bounds of the runlist.
|
|
*
|
|
* Locking: - The runlist must be unlocked on entry and is unlocked on return.
|
|
* - This function takes the runlist lock for writing and may modify
|
|
* the runlist.
|
|
*/
|
|
int ntfs_map_runlist(struct ntfs_inode *ni, s64 vcn)
|
|
{
|
|
int err = 0;
|
|
|
|
down_write(&ni->runlist.lock);
|
|
/* Make sure someone else didn't do the work while we were sleeping. */
|
|
if (likely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) <=
|
|
LCN_RL_NOT_MAPPED))
|
|
err = ntfs_map_runlist_nolock(ni, vcn, NULL);
|
|
up_write(&ni->runlist.lock);
|
|
return err;
|
|
}
|
|
|
|
struct runlist_element *ntfs_attr_vcn_to_rl(struct ntfs_inode *ni, s64 vcn, s64 *lcn)
|
|
{
|
|
struct runlist_element *rl = ni->runlist.rl;
|
|
int err;
|
|
bool is_retry = false;
|
|
|
|
if (!rl) {
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err)
|
|
return ERR_PTR(-ENOENT);
|
|
rl = ni->runlist.rl;
|
|
}
|
|
|
|
remap_rl:
|
|
/* Seek to element containing target vcn. */
|
|
while (rl->length && rl[1].vcn <= vcn)
|
|
rl++;
|
|
*lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
|
|
|
|
if (*lcn <= LCN_RL_NOT_MAPPED && is_retry == false) {
|
|
is_retry = true;
|
|
if (!ntfs_map_runlist_nolock(ni, vcn, NULL)) {
|
|
rl = ni->runlist.rl;
|
|
goto remap_rl;
|
|
}
|
|
}
|
|
|
|
return rl;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_vcn_to_lcn_nolock - convert a vcn into a lcn given an ntfs inode
|
|
* @ni: ntfs inode of the attribute whose runlist to search
|
|
* @vcn: vcn to convert
|
|
* @write_locked: true if the runlist is locked for writing
|
|
*
|
|
* Find the virtual cluster number @vcn in the runlist of the ntfs attribute
|
|
* described by the ntfs inode @ni and return the corresponding logical cluster
|
|
* number (lcn).
|
|
*
|
|
* If the @vcn is not mapped yet, the attempt is made to map the attribute
|
|
* extent containing the @vcn and the vcn to lcn conversion is retried.
|
|
*
|
|
* If @write_locked is true the caller has locked the runlist for writing and
|
|
* if false for reading.
|
|
*
|
|
* Since lcns must be >= 0, we use negative return codes with special meaning:
|
|
*
|
|
* Return code Meaning / Description
|
|
* ==========================================
|
|
* LCN_HOLE Hole / not allocated on disk.
|
|
* LCN_ENOENT There is no such vcn in the runlist, i.e. @vcn is out of bounds.
|
|
* LCN_ENOMEM Not enough memory to map runlist.
|
|
* LCN_EIO Critical error (runlist/file is corrupt, i/o error, etc).
|
|
*
|
|
* Locking: - The runlist must be locked on entry and is left locked on return.
|
|
* - If @write_locked is 'false', i.e. the runlist is locked for reading,
|
|
* the lock may be dropped inside the function so you cannot rely on
|
|
* the runlist still being the same when this function returns.
|
|
*/
|
|
s64 ntfs_attr_vcn_to_lcn_nolock(struct ntfs_inode *ni, const s64 vcn,
|
|
const bool write_locked)
|
|
{
|
|
s64 lcn;
|
|
unsigned long flags;
|
|
bool is_retry = false;
|
|
|
|
ntfs_debug("Entering for i_ino 0x%llx, vcn 0x%llx, %s_locked.",
|
|
ni->mft_no, (unsigned long long)vcn,
|
|
write_locked ? "write" : "read");
|
|
if (!ni->runlist.rl) {
|
|
read_lock_irqsave(&ni->size_lock, flags);
|
|
if (!ni->allocated_size) {
|
|
read_unlock_irqrestore(&ni->size_lock, flags);
|
|
return LCN_ENOENT;
|
|
}
|
|
read_unlock_irqrestore(&ni->size_lock, flags);
|
|
}
|
|
retry_remap:
|
|
/* Convert vcn to lcn. If that fails map the runlist and retry once. */
|
|
lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn);
|
|
if (likely(lcn >= LCN_HOLE)) {
|
|
ntfs_debug("Done, lcn 0x%llx.", (long long)lcn);
|
|
return lcn;
|
|
}
|
|
if (lcn != LCN_RL_NOT_MAPPED) {
|
|
if (lcn != LCN_ENOENT)
|
|
lcn = LCN_EIO;
|
|
} else if (!is_retry) {
|
|
int err;
|
|
|
|
if (!write_locked) {
|
|
up_read(&ni->runlist.lock);
|
|
down_write(&ni->runlist.lock);
|
|
if (unlikely(ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn) !=
|
|
LCN_RL_NOT_MAPPED)) {
|
|
up_write(&ni->runlist.lock);
|
|
down_read(&ni->runlist.lock);
|
|
goto retry_remap;
|
|
}
|
|
}
|
|
err = ntfs_map_runlist_nolock(ni, vcn, NULL);
|
|
if (!write_locked) {
|
|
up_write(&ni->runlist.lock);
|
|
down_read(&ni->runlist.lock);
|
|
}
|
|
if (likely(!err)) {
|
|
is_retry = true;
|
|
goto retry_remap;
|
|
}
|
|
if (err == -ENOENT)
|
|
lcn = LCN_ENOENT;
|
|
else if (err == -ENOMEM)
|
|
lcn = LCN_ENOMEM;
|
|
else
|
|
lcn = LCN_EIO;
|
|
}
|
|
if (lcn != LCN_ENOENT)
|
|
ntfs_error(ni->vol->sb, "Failed with error code %lli.",
|
|
(long long)lcn);
|
|
return lcn;
|
|
}
|
|
|
|
struct runlist_element *__ntfs_attr_find_vcn_nolock(struct runlist *runlist, const s64 vcn)
|
|
{
|
|
size_t lower_idx, upper_idx, idx;
|
|
struct runlist_element *run;
|
|
int rh = runlist->rl_hint;
|
|
|
|
if (runlist->count <= 1)
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
if (runlist->count - 1 > rh && runlist->rl[rh].vcn <= vcn) {
|
|
if (vcn < runlist->rl[rh].vcn + runlist->rl[rh].length)
|
|
return &runlist->rl[rh];
|
|
if (runlist->count - 2 == rh)
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
lower_idx = rh + 1;
|
|
} else {
|
|
run = &runlist->rl[0];
|
|
if (vcn < run->vcn)
|
|
return ERR_PTR(-ENOENT);
|
|
else if (vcn < run->vcn + run->length) {
|
|
runlist->rl_hint = 0;
|
|
return run;
|
|
}
|
|
|
|
lower_idx = 1;
|
|
}
|
|
|
|
run = &runlist->rl[runlist->count - 2];
|
|
if (vcn >= run->vcn && vcn < run->vcn + run->length) {
|
|
runlist->rl_hint = runlist->count - 2;
|
|
return run;
|
|
}
|
|
if (vcn >= run->vcn + run->length)
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
upper_idx = runlist->count - 2;
|
|
|
|
while (lower_idx <= upper_idx) {
|
|
idx = (lower_idx + upper_idx) >> 1;
|
|
run = &runlist->rl[idx];
|
|
|
|
if (vcn < run->vcn)
|
|
upper_idx = idx - 1;
|
|
else if (vcn >= run->vcn + run->length)
|
|
lower_idx = idx + 1;
|
|
else {
|
|
runlist->rl_hint = idx;
|
|
return run;
|
|
}
|
|
}
|
|
|
|
return ERR_PTR(-ENOENT);
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_find_vcn_nolock - find a vcn in the runlist of an ntfs inode
|
|
* @ni: ntfs inode describing the runlist to search
|
|
* @vcn: vcn to find
|
|
* @ctx: active attribute search context if present or NULL if not
|
|
*
|
|
* Find the virtual cluster number @vcn in the runlist described by the ntfs
|
|
* inode @ni and return the address of the runlist element containing the @vcn.
|
|
*
|
|
* If the @vcn is not mapped yet, the attempt is made to map the attribute
|
|
* extent containing the @vcn and the vcn to lcn conversion is retried.
|
|
*
|
|
* If @ctx is specified, it is an active search context of @ni and its base mft
|
|
* record. This is needed when ntfs_attr_find_vcn_nolock() encounters unmapped
|
|
* runlist fragments and allows their mapping. If you do not have the mft
|
|
* record mapped, you can specify @ctx as NULL and ntfs_attr_find_vcn_nolock()
|
|
* will perform the necessary mapping and unmapping.
|
|
*
|
|
* Note, ntfs_attr_find_vcn_nolock() saves the state of @ctx on entry and
|
|
* restores it before returning. Thus, @ctx will be left pointing to the same
|
|
* attribute on return as on entry. However, the actual pointers in @ctx may
|
|
* point to different memory locations on return, so you must remember to reset
|
|
* any cached pointers from the @ctx, i.e. after the call to
|
|
* ntfs_attr_find_vcn_nolock(), you will probably want to do:
|
|
* m = ctx->mrec;
|
|
* a = ctx->attr;
|
|
* Assuming you cache ctx->attr in a variable @a of type attr_record * and that
|
|
* you cache ctx->mrec in a variable @m of type struct mft_record *.
|
|
* Note you need to distinguish between the lcn of the returned runlist element
|
|
* being >= 0 and LCN_HOLE. In the later case you have to return zeroes on
|
|
* read and allocate clusters on write.
|
|
*/
|
|
struct runlist_element *ntfs_attr_find_vcn_nolock(struct ntfs_inode *ni, const s64 vcn,
|
|
struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
unsigned long flags;
|
|
struct runlist_element *rl;
|
|
int err = 0;
|
|
bool is_retry = false;
|
|
|
|
ntfs_debug("Entering for i_ino 0x%llx, vcn 0x%llx, with%s ctx.",
|
|
ni->mft_no, (unsigned long long)vcn, ctx ? "" : "out");
|
|
if (!ni->runlist.rl) {
|
|
read_lock_irqsave(&ni->size_lock, flags);
|
|
if (!ni->allocated_size) {
|
|
read_unlock_irqrestore(&ni->size_lock, flags);
|
|
return ERR_PTR(-ENOENT);
|
|
}
|
|
read_unlock_irqrestore(&ni->size_lock, flags);
|
|
}
|
|
|
|
retry_remap:
|
|
rl = ni->runlist.rl;
|
|
if (likely(rl && vcn >= rl[0].vcn)) {
|
|
rl = __ntfs_attr_find_vcn_nolock(&ni->runlist, vcn);
|
|
if (IS_ERR(rl))
|
|
err = PTR_ERR(rl);
|
|
else if (rl->lcn >= LCN_HOLE)
|
|
return rl;
|
|
else if (rl->lcn <= LCN_ENOENT)
|
|
err = -EIO;
|
|
}
|
|
if (!err && !is_retry) {
|
|
/*
|
|
* If the search context is invalid we cannot map the unmapped
|
|
* region.
|
|
*/
|
|
if (ctx && IS_ERR(ctx->mrec))
|
|
err = PTR_ERR(ctx->mrec);
|
|
else {
|
|
/*
|
|
* The @vcn is in an unmapped region, map the runlist
|
|
* and retry.
|
|
*/
|
|
err = ntfs_map_runlist_nolock(ni, vcn, ctx);
|
|
if (likely(!err)) {
|
|
is_retry = true;
|
|
goto retry_remap;
|
|
}
|
|
}
|
|
if (err == -EINVAL)
|
|
err = -EIO;
|
|
} else if (!err)
|
|
err = -EIO;
|
|
if (err != -ENOENT)
|
|
ntfs_error(ni->vol->sb, "Failed with error code %i.", err);
|
|
return ERR_PTR(err);
|
|
}
|
|
|
|
static u32 ntfs_resident_attr_min_value_length(const __le32 type)
|
|
{
|
|
switch (type) {
|
|
case AT_STANDARD_INFORMATION:
|
|
return offsetof(struct standard_information, ver) +
|
|
sizeof(((struct standard_information *)0)->ver.v1.reserved12);
|
|
case AT_ATTRIBUTE_LIST:
|
|
return offsetof(struct attr_list_entry, name);
|
|
case AT_FILE_NAME:
|
|
return offsetof(struct file_name_attr, file_name);
|
|
case AT_OBJECT_ID:
|
|
return sizeof(struct guid);
|
|
case AT_SECURITY_DESCRIPTOR:
|
|
return sizeof(struct security_descriptor_relative);
|
|
case AT_VOLUME_INFORMATION:
|
|
return sizeof(struct volume_information);
|
|
case AT_INDEX_ROOT:
|
|
return sizeof(struct index_root);
|
|
case AT_REPARSE_POINT:
|
|
return offsetof(struct reparse_point, reparse_data);
|
|
case AT_EA_INFORMATION:
|
|
return sizeof(struct ea_information);
|
|
case AT_EA:
|
|
return offsetof(struct ea_attr, ea_name) + 1;
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_find - find (next) attribute in mft record
|
|
* @type: attribute type to find
|
|
* @name: attribute name to find (optional, i.e. NULL means don't care)
|
|
* @name_len: attribute name length (only needed if @name present)
|
|
* @ic: IGNORE_CASE or CASE_SENSITIVE (ignored if @name not present)
|
|
* @val: attribute value to find (optional, resident attributes only)
|
|
* @val_len: attribute value length
|
|
* @ctx: search context with mft record and attribute to search from
|
|
*
|
|
* You should not need to call this function directly. Use ntfs_attr_lookup()
|
|
* instead.
|
|
*
|
|
* ntfs_attr_find() takes a search context @ctx as parameter and searches the
|
|
* mft record specified by @ctx->mrec, beginning at @ctx->attr, for an
|
|
* attribute of @type, optionally @name and @val.
|
|
*
|
|
* If the attribute is found, ntfs_attr_find() returns 0 and @ctx->attr will
|
|
* point to the found attribute.
|
|
*
|
|
* If the attribute is not found, ntfs_attr_find() returns -ENOENT and
|
|
* @ctx->attr will point to the attribute before which the attribute being
|
|
* searched for would need to be inserted if such an action were to be desired.
|
|
*
|
|
* On actual error, ntfs_attr_find() returns -EIO. In this case @ctx->attr is
|
|
* undefined and in particular do not rely on it not changing.
|
|
*
|
|
* If @ctx->is_first is 'true', the search begins with @ctx->attr itself. If it
|
|
* is 'false', the search begins after @ctx->attr.
|
|
*
|
|
* If @ic is IGNORE_CASE, the @name comparisson is not case sensitive and
|
|
* @ctx->ntfs_ino must be set to the ntfs inode to which the mft record
|
|
* @ctx->mrec belongs. This is so we can get at the ntfs volume and hence at
|
|
* the upcase table. If @ic is CASE_SENSITIVE, the comparison is case
|
|
* sensitive. When @name is present, @name_len is the @name length in Unicode
|
|
* characters.
|
|
*
|
|
* If @name is not present (NULL), we assume that the unnamed attribute is
|
|
* being searched for.
|
|
*
|
|
* Finally, the resident attribute value @val is looked for, if present. If
|
|
* @val is not present (NULL), @val_len is ignored.
|
|
*
|
|
* ntfs_attr_find() only searches the specified mft record and it ignores the
|
|
* presence of an attribute list attribute (unless it is the one being searched
|
|
* for, obviously). If you need to take attribute lists into consideration,
|
|
* use ntfs_attr_lookup() instead (see below). This also means that you cannot
|
|
* use ntfs_attr_find() to search for extent records of non-resident
|
|
* attributes, as extents with lowest_vcn != 0 are usually described by the
|
|
* attribute list attribute only. - Note that it is possible that the first
|
|
* extent is only in the attribute list while the last extent is in the base
|
|
* mft record, so do not rely on being able to find the first extent in the
|
|
* base mft record.
|
|
*
|
|
* Warning: Never use @val when looking for attribute types which can be
|
|
* non-resident as this most likely will result in a crash!
|
|
*/
|
|
static int ntfs_attr_find(const __le32 type, const __le16 *name,
|
|
const u32 name_len, const u32 ic,
|
|
const u8 *val, const u32 val_len, struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
struct attr_record *a;
|
|
struct ntfs_volume *vol = ctx->ntfs_ino->vol;
|
|
__le16 *upcase = vol->upcase;
|
|
u32 upcase_len = vol->upcase_len;
|
|
unsigned int space;
|
|
|
|
/*
|
|
* Iterate over attributes in mft record starting at @ctx->attr, or the
|
|
* attribute following that, if @ctx->is_first is 'true'.
|
|
*/
|
|
if (ctx->is_first) {
|
|
a = ctx->attr;
|
|
ctx->is_first = false;
|
|
} else
|
|
a = (struct attr_record *)((u8 *)ctx->attr +
|
|
le32_to_cpu(ctx->attr->length));
|
|
for (;; a = (struct attr_record *)((u8 *)a + le32_to_cpu(a->length))) {
|
|
if ((u8 *)a < (u8 *)ctx->mrec || (u8 *)a > (u8 *)ctx->mrec +
|
|
le32_to_cpu(ctx->mrec->bytes_allocated))
|
|
break;
|
|
|
|
space = le32_to_cpu(ctx->mrec->bytes_in_use) - ((u8 *)a - (u8 *)ctx->mrec);
|
|
if ((space < offsetof(struct attr_record, data.resident.reserved) + 1 ||
|
|
space < le32_to_cpu(a->length)) && (space < 4 || a->type != AT_END))
|
|
break;
|
|
|
|
ctx->attr = a;
|
|
if (((type != AT_UNUSED) && (le32_to_cpu(a->type) > le32_to_cpu(type))) ||
|
|
a->type == AT_END)
|
|
return -ENOENT;
|
|
if (unlikely(!a->length))
|
|
break;
|
|
if (type == AT_UNUSED)
|
|
return 0;
|
|
if (a->type != type)
|
|
continue;
|
|
/*
|
|
* If @name is present, compare the two names. If @name is
|
|
* missing, assume we want an unnamed attribute.
|
|
*/
|
|
if (!name || name == AT_UNNAMED) {
|
|
/* The search failed if the found attribute is named. */
|
|
if (a->name_length)
|
|
return -ENOENT;
|
|
} else {
|
|
if (a->name_length && ((le16_to_cpu(a->name_offset) +
|
|
a->name_length * sizeof(__le16)) >
|
|
le32_to_cpu(a->length))) {
|
|
ntfs_error(vol->sb, "Corrupt attribute name in MFT record %llu\n",
|
|
ctx->ntfs_ino->mft_no);
|
|
break;
|
|
}
|
|
|
|
if (!ntfs_are_names_equal(name, name_len,
|
|
(__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length, ic, upcase, upcase_len)) {
|
|
register int rc;
|
|
|
|
rc = ntfs_collate_names(name, name_len,
|
|
(__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length, 1, IGNORE_CASE,
|
|
upcase, upcase_len);
|
|
/*
|
|
* If @name collates before a->name, there is no
|
|
* matching attribute.
|
|
*/
|
|
if (rc == -1)
|
|
return -ENOENT;
|
|
/* If the strings are not equal, continue search. */
|
|
if (rc)
|
|
continue;
|
|
rc = ntfs_collate_names(name, name_len,
|
|
(__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length, 1, CASE_SENSITIVE,
|
|
upcase, upcase_len);
|
|
if (rc == -1)
|
|
return -ENOENT;
|
|
if (rc)
|
|
continue;
|
|
}
|
|
}
|
|
|
|
/* Validate attribute's value offset/length */
|
|
if (!a->non_resident) {
|
|
u32 min_len;
|
|
u32 value_length = le32_to_cpu(a->data.resident.value_length);
|
|
u16 value_offset = le16_to_cpu(a->data.resident.value_offset);
|
|
|
|
if (value_length > le32_to_cpu(a->length) ||
|
|
value_offset > le32_to_cpu(a->length) - value_length)
|
|
break;
|
|
|
|
min_len = ntfs_resident_attr_min_value_length(a->type);
|
|
if (min_len && value_length < min_len) {
|
|
ntfs_error(vol->sb,
|
|
"Too small %#x resident attribute value in MFT record %lld\n",
|
|
le32_to_cpu(a->type), (long long)ctx->ntfs_ino->mft_no);
|
|
break;
|
|
}
|
|
} else {
|
|
u32 min_len;
|
|
u16 mp_offset;
|
|
|
|
min_len = offsetof(struct attr_record, data.non_resident.initialized_size) +
|
|
sizeof(a->data.non_resident.initialized_size);
|
|
if (le32_to_cpu(a->length) < min_len)
|
|
break;
|
|
|
|
mp_offset = le16_to_cpu(a->data.non_resident.mapping_pairs_offset);
|
|
if (mp_offset < min_len ||
|
|
mp_offset > le32_to_cpu(a->length))
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* The names match or @name not present and attribute is
|
|
* unnamed. If no @val specified, we have found the attribute
|
|
* and are done.
|
|
*/
|
|
if (!val || a->non_resident)
|
|
return 0;
|
|
/* @val is present; compare values. */
|
|
else {
|
|
u32 value_length = le32_to_cpu(a->data.resident.value_length);
|
|
int rc;
|
|
|
|
rc = memcmp(val, (u8 *)a + le16_to_cpu(
|
|
a->data.resident.value_offset),
|
|
min_t(u32, val_len, value_length));
|
|
/*
|
|
* If @val collates before the current attribute's
|
|
* value, there is no matching attribute.
|
|
*/
|
|
if (!rc) {
|
|
if (val_len == value_length)
|
|
return 0;
|
|
if (val_len < value_length)
|
|
return -ENOENT;
|
|
} else if (rc < 0)
|
|
return -ENOENT;
|
|
}
|
|
}
|
|
ntfs_error(vol->sb, "mft %#llx, type %#x is corrupt. Run chkdsk.",
|
|
(long long)ctx->ntfs_ino->mft_no, le32_to_cpu(type));
|
|
NVolSetErrors(vol);
|
|
return -EIO;
|
|
}
|
|
|
|
void ntfs_attr_name_free(unsigned char **name)
|
|
{
|
|
if (*name) {
|
|
kfree(*name);
|
|
*name = NULL;
|
|
}
|
|
}
|
|
|
|
char *ntfs_attr_name_get(const struct ntfs_volume *vol, const __le16 *uname,
|
|
const int uname_len)
|
|
{
|
|
unsigned char *name = NULL;
|
|
int name_len;
|
|
|
|
name_len = ntfs_ucstonls(vol, uname, uname_len, &name, 0);
|
|
if (name_len < 0) {
|
|
ntfs_error(vol->sb, "ntfs_ucstonls error");
|
|
/* This function when returns -1, memory for name might
|
|
* be allocated. So lets free this memory.
|
|
*/
|
|
ntfs_attr_name_free(&name);
|
|
return NULL;
|
|
|
|
} else if (name_len > 0)
|
|
return name;
|
|
|
|
ntfs_attr_name_free(&name);
|
|
return NULL;
|
|
}
|
|
|
|
int load_attribute_list(struct ntfs_inode *base_ni, u8 *al_start, const s64 size)
|
|
{
|
|
struct inode *attr_vi = NULL;
|
|
u8 *al;
|
|
struct attr_list_entry *ale;
|
|
|
|
if (!al_start || size <= 0)
|
|
return -EINVAL;
|
|
|
|
attr_vi = ntfs_attr_iget(VFS_I(base_ni), AT_ATTRIBUTE_LIST, AT_UNNAMED, 0);
|
|
if (IS_ERR(attr_vi)) {
|
|
ntfs_error(base_ni->vol->sb,
|
|
"Failed to open an inode for Attribute list, mft = %llu",
|
|
base_ni->mft_no);
|
|
return PTR_ERR(attr_vi);
|
|
}
|
|
|
|
if (ntfs_inode_attr_pread(attr_vi, 0, size, al_start) != size) {
|
|
iput(attr_vi);
|
|
ntfs_error(base_ni->vol->sb,
|
|
"Failed to read attribute list, mft = %llu",
|
|
base_ni->mft_no);
|
|
return -EIO;
|
|
}
|
|
iput(attr_vi);
|
|
|
|
for (al = al_start; al < al_start + size; al += le16_to_cpu(ale->length)) {
|
|
ale = (struct attr_list_entry *)al;
|
|
if (ale->name_offset != sizeof(struct attr_list_entry))
|
|
break;
|
|
if (le16_to_cpu(ale->length) <= ale->name_offset + ale->name_length ||
|
|
al + le16_to_cpu(ale->length) > al_start + size)
|
|
break;
|
|
if (ale->type == AT_UNUSED)
|
|
break;
|
|
if (MSEQNO_LE(ale->mft_reference) == 0)
|
|
break;
|
|
}
|
|
if (al != al_start + size) {
|
|
ntfs_error(base_ni->vol->sb, "Corrupt attribute list, mft = %llu",
|
|
base_ni->mft_no);
|
|
return -EIO;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_external_attr_find - find an attribute in the attribute list of an inode
|
|
* @type: attribute type to find
|
|
* @name: attribute name to find (optional, i.e. NULL means don't care)
|
|
* @name_len: attribute name length (only needed if @name present)
|
|
* @ic: IGNORE_CASE or CASE_SENSITIVE (ignored if @name not present)
|
|
* @lowest_vcn: lowest vcn to find (optional, non-resident attributes only)
|
|
* @val: attribute value to find (optional, resident attributes only)
|
|
* @val_len: attribute value length
|
|
* @ctx: search context with mft record and attribute to search from
|
|
*
|
|
* You should not need to call this function directly. Use ntfs_attr_lookup()
|
|
* instead.
|
|
*
|
|
* Find an attribute by searching the attribute list for the corresponding
|
|
* attribute list entry. Having found the entry, map the mft record if the
|
|
* attribute is in a different mft record/inode, ntfs_attr_find() the attribute
|
|
* in there and return it.
|
|
*
|
|
* On first search @ctx->ntfs_ino must be the base mft record and @ctx must
|
|
* have been obtained from a call to ntfs_attr_get_search_ctx(). On subsequent
|
|
* calls @ctx->ntfs_ino can be any extent inode, too (@ctx->base_ntfs_ino is
|
|
* then the base inode).
|
|
*
|
|
* After finishing with the attribute/mft record you need to call
|
|
* ntfs_attr_put_search_ctx() to cleanup the search context (unmapping any
|
|
* mapped inodes, etc).
|
|
*
|
|
* If the attribute is found, ntfs_external_attr_find() returns 0 and
|
|
* @ctx->attr will point to the found attribute. @ctx->mrec will point to the
|
|
* mft record in which @ctx->attr is located and @ctx->al_entry will point to
|
|
* the attribute list entry for the attribute.
|
|
*
|
|
* If the attribute is not found, ntfs_external_attr_find() returns -ENOENT and
|
|
* @ctx->attr will point to the attribute in the base mft record before which
|
|
* the attribute being searched for would need to be inserted if such an action
|
|
* were to be desired. @ctx->mrec will point to the mft record in which
|
|
* @ctx->attr is located and @ctx->al_entry will point to the attribute list
|
|
* entry of the attribute before which the attribute being searched for would
|
|
* need to be inserted if such an action were to be desired.
|
|
*
|
|
* Thus to insert the not found attribute, one wants to add the attribute to
|
|
* @ctx->mrec (the base mft record) and if there is not enough space, the
|
|
* attribute should be placed in a newly allocated extent mft record. The
|
|
* attribute list entry for the inserted attribute should be inserted in the
|
|
* attribute list attribute at @ctx->al_entry.
|
|
*
|
|
* On actual error, ntfs_external_attr_find() returns -EIO. In this case
|
|
* @ctx->attr is undefined and in particular do not rely on it not changing.
|
|
*/
|
|
static int ntfs_external_attr_find(const __le32 type,
|
|
const __le16 *name, const u32 name_len,
|
|
const u32 ic, const s64 lowest_vcn,
|
|
const u8 *val, const u32 val_len, struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
struct ntfs_inode *base_ni = ctx->base_ntfs_ino, *ni = ctx->ntfs_ino;
|
|
struct ntfs_volume *vol;
|
|
struct attr_list_entry *al_entry, *next_al_entry;
|
|
u8 *al_start, *al_end;
|
|
struct attr_record *a;
|
|
__le16 *al_name;
|
|
u32 al_name_len;
|
|
u32 attr_len, mft_free_len;
|
|
bool is_first_search = false;
|
|
int err = 0;
|
|
static const char *es = " Unmount and run chkdsk.";
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, type 0x%x.", ni->mft_no, type);
|
|
if (!base_ni) {
|
|
/* First call happens with the base mft record. */
|
|
base_ni = ctx->base_ntfs_ino = ctx->ntfs_ino;
|
|
ctx->base_mrec = ctx->mrec;
|
|
ctx->mapped_base_mrec = ctx->mapped_mrec;
|
|
}
|
|
if (ni == base_ni)
|
|
ctx->base_attr = ctx->attr;
|
|
if (type == AT_END)
|
|
goto not_found;
|
|
vol = base_ni->vol;
|
|
al_start = base_ni->attr_list;
|
|
al_end = al_start + base_ni->attr_list_size;
|
|
if (!ctx->al_entry) {
|
|
ctx->al_entry = (struct attr_list_entry *)al_start;
|
|
is_first_search = true;
|
|
}
|
|
/*
|
|
* Iterate over entries in attribute list starting at @ctx->al_entry,
|
|
* or the entry following that, if @ctx->is_first is 'true'.
|
|
*/
|
|
if (ctx->is_first) {
|
|
al_entry = ctx->al_entry;
|
|
ctx->is_first = false;
|
|
/*
|
|
* If an enumeration and the first attribute is higher than
|
|
* the attribute list itself, need to return the attribute list
|
|
* attribute.
|
|
*/
|
|
if ((type == AT_UNUSED) && is_first_search &&
|
|
le32_to_cpu(al_entry->type) >
|
|
le32_to_cpu(AT_ATTRIBUTE_LIST))
|
|
goto find_attr_list_attr;
|
|
} else {
|
|
/* Check for small entry */
|
|
if (((al_end - (u8 *)ctx->al_entry) <
|
|
(long)offsetof(struct attr_list_entry, name)) ||
|
|
(le16_to_cpu(ctx->al_entry->length) & 7) ||
|
|
(le16_to_cpu(ctx->al_entry->length) < offsetof(struct attr_list_entry, name)))
|
|
goto corrupt;
|
|
|
|
al_entry = (struct attr_list_entry *)((u8 *)ctx->al_entry +
|
|
le16_to_cpu(ctx->al_entry->length));
|
|
|
|
if ((u8 *)al_entry == al_end)
|
|
goto not_found;
|
|
|
|
/* Preliminary check for small entry */
|
|
if ((al_end - (u8 *)al_entry) <
|
|
(long)offsetof(struct attr_list_entry, name))
|
|
goto corrupt;
|
|
|
|
/*
|
|
* If this is an enumeration and the attribute list attribute
|
|
* is the next one in the enumeration sequence, just return the
|
|
* attribute list attribute from the base mft record as it is
|
|
* not listed in the attribute list itself.
|
|
*/
|
|
if ((type == AT_UNUSED) && le32_to_cpu(ctx->al_entry->type) <
|
|
le32_to_cpu(AT_ATTRIBUTE_LIST) &&
|
|
le32_to_cpu(al_entry->type) >
|
|
le32_to_cpu(AT_ATTRIBUTE_LIST)) {
|
|
find_attr_list_attr:
|
|
|
|
/* Check for bogus calls. */
|
|
if (name || name_len || val || val_len || lowest_vcn)
|
|
return -EINVAL;
|
|
|
|
/* We want the base record. */
|
|
if (ctx->ntfs_ino != base_ni)
|
|
unmap_mft_record(ctx->ntfs_ino);
|
|
ctx->ntfs_ino = base_ni;
|
|
ctx->mapped_mrec = ctx->mapped_base_mrec;
|
|
ctx->mrec = ctx->base_mrec;
|
|
ctx->is_first = true;
|
|
|
|
/* Sanity checks are performed elsewhere. */
|
|
ctx->attr = (struct attr_record *)((u8 *)ctx->mrec +
|
|
le16_to_cpu(ctx->mrec->attrs_offset));
|
|
|
|
/* Find the attribute list attribute. */
|
|
err = ntfs_attr_find(AT_ATTRIBUTE_LIST, NULL, 0,
|
|
IGNORE_CASE, NULL, 0, ctx);
|
|
|
|
/*
|
|
* Setup the search context so the correct
|
|
* attribute is returned next time round.
|
|
*/
|
|
ctx->al_entry = al_entry;
|
|
ctx->is_first = true;
|
|
|
|
/* Got it. Done. */
|
|
if (!err)
|
|
return 0;
|
|
|
|
/* Error! If other than not found return it. */
|
|
if (err != -ENOENT)
|
|
return err;
|
|
|
|
/* Not found?!? Absurd! */
|
|
ntfs_error(ctx->ntfs_ino->vol->sb, "Attribute list wasn't found");
|
|
return -EIO;
|
|
}
|
|
}
|
|
for (;; al_entry = next_al_entry) {
|
|
/* Out of bounds check. */
|
|
if ((u8 *)al_entry < base_ni->attr_list ||
|
|
(u8 *)al_entry > al_end)
|
|
break; /* Inode is corrupt. */
|
|
ctx->al_entry = al_entry;
|
|
/* Catch the end of the attribute list. */
|
|
if ((u8 *)al_entry == al_end)
|
|
goto not_found;
|
|
|
|
if ((((u8 *)al_entry + offsetof(struct attr_list_entry, name)) > al_end) ||
|
|
((u8 *)al_entry + le16_to_cpu(al_entry->length) > al_end) ||
|
|
(le16_to_cpu(al_entry->length) & 7) ||
|
|
(le16_to_cpu(al_entry->length) <
|
|
offsetof(struct attr_list_entry, name_length)) ||
|
|
(al_entry->name_length && ((u8 *)al_entry + al_entry->name_offset +
|
|
al_entry->name_length * sizeof(__le16)) > al_end))
|
|
break; /* corrupt */
|
|
|
|
next_al_entry = (struct attr_list_entry *)((u8 *)al_entry +
|
|
le16_to_cpu(al_entry->length));
|
|
if (type != AT_UNUSED) {
|
|
if (le32_to_cpu(al_entry->type) > le32_to_cpu(type))
|
|
goto not_found;
|
|
if (type != al_entry->type)
|
|
continue;
|
|
}
|
|
/*
|
|
* If @name is present, compare the two names. If @name is
|
|
* missing, assume we want an unnamed attribute.
|
|
*/
|
|
al_name_len = al_entry->name_length;
|
|
al_name = (__le16 *)((u8 *)al_entry + al_entry->name_offset);
|
|
|
|
/*
|
|
* If !@type we want the attribute represented by this
|
|
* attribute list entry.
|
|
*/
|
|
if (type == AT_UNUSED)
|
|
goto is_enumeration;
|
|
|
|
if (!name || name == AT_UNNAMED) {
|
|
if (al_name_len)
|
|
goto not_found;
|
|
} else if (!ntfs_are_names_equal(al_name, al_name_len, name,
|
|
name_len, ic, vol->upcase, vol->upcase_len)) {
|
|
register int rc;
|
|
|
|
rc = ntfs_collate_names(name, name_len, al_name,
|
|
al_name_len, 1, IGNORE_CASE,
|
|
vol->upcase, vol->upcase_len);
|
|
/*
|
|
* If @name collates before al_name, there is no
|
|
* matching attribute.
|
|
*/
|
|
if (rc == -1)
|
|
goto not_found;
|
|
/* If the strings are not equal, continue search. */
|
|
if (rc)
|
|
continue;
|
|
|
|
rc = ntfs_collate_names(name, name_len, al_name,
|
|
al_name_len, 1, CASE_SENSITIVE,
|
|
vol->upcase, vol->upcase_len);
|
|
if (rc == -1)
|
|
goto not_found;
|
|
if (rc)
|
|
continue;
|
|
}
|
|
/*
|
|
* The names match or @name not present and attribute is
|
|
* unnamed. Now check @lowest_vcn. Continue search if the
|
|
* next attribute list entry still fits @lowest_vcn. Otherwise
|
|
* we have reached the right one or the search has failed.
|
|
*/
|
|
if (lowest_vcn && (u8 *)next_al_entry >= al_start &&
|
|
(u8 *)next_al_entry + 6 < al_end &&
|
|
(u8 *)next_al_entry + le16_to_cpu(
|
|
next_al_entry->length) <= al_end &&
|
|
le64_to_cpu(next_al_entry->lowest_vcn) <=
|
|
lowest_vcn &&
|
|
next_al_entry->type == al_entry->type &&
|
|
next_al_entry->name_length == al_name_len &&
|
|
ntfs_are_names_equal((__le16 *)((u8 *)
|
|
next_al_entry +
|
|
next_al_entry->name_offset),
|
|
next_al_entry->name_length,
|
|
al_name, al_name_len, CASE_SENSITIVE,
|
|
vol->upcase, vol->upcase_len))
|
|
continue;
|
|
|
|
is_enumeration:
|
|
if (MREF_LE(al_entry->mft_reference) == ni->mft_no) {
|
|
if (MSEQNO_LE(al_entry->mft_reference) != ni->seq_no) {
|
|
ntfs_error(vol->sb,
|
|
"Found stale mft reference in attribute list of base inode 0x%llx.%s",
|
|
base_ni->mft_no, es);
|
|
err = -EIO;
|
|
break;
|
|
}
|
|
} else { /* Mft references do not match. */
|
|
/* If there is a mapped record unmap it first. */
|
|
if (ni != base_ni)
|
|
unmap_extent_mft_record(ni);
|
|
/* Do we want the base record back? */
|
|
if (MREF_LE(al_entry->mft_reference) ==
|
|
base_ni->mft_no) {
|
|
ni = ctx->ntfs_ino = base_ni;
|
|
ctx->mrec = ctx->base_mrec;
|
|
ctx->mapped_mrec = ctx->mapped_base_mrec;
|
|
} else {
|
|
/* We want an extent record. */
|
|
ctx->mrec = map_extent_mft_record(base_ni,
|
|
le64_to_cpu(
|
|
al_entry->mft_reference), &ni);
|
|
if (IS_ERR(ctx->mrec)) {
|
|
ntfs_error(vol->sb,
|
|
"Failed to map extent mft record 0x%lx of base inode 0x%llx.%s",
|
|
MREF_LE(al_entry->mft_reference),
|
|
base_ni->mft_no, es);
|
|
err = PTR_ERR(ctx->mrec);
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
/* Cause @ctx to be sanitized below. */
|
|
ni = NULL;
|
|
break;
|
|
}
|
|
ctx->ntfs_ino = ni;
|
|
ctx->mapped_mrec = true;
|
|
|
|
}
|
|
}
|
|
a = ctx->attr = (struct attr_record *)((u8 *)ctx->mrec +
|
|
le16_to_cpu(ctx->mrec->attrs_offset));
|
|
/*
|
|
* ctx->vfs_ino, ctx->mrec, and ctx->attr now point to the
|
|
* mft record containing the attribute represented by the
|
|
* current al_entry.
|
|
*/
|
|
/*
|
|
* We could call into ntfs_attr_find() to find the right
|
|
* attribute in this mft record but this would be less
|
|
* efficient and not quite accurate as ntfs_attr_find() ignores
|
|
* the attribute instance numbers for example which become
|
|
* important when one plays with attribute lists. Also,
|
|
* because a proper match has been found in the attribute list
|
|
* entry above, the comparison can now be optimized. So it is
|
|
* worth re-implementing a simplified ntfs_attr_find() here.
|
|
*/
|
|
/*
|
|
* Use a manual loop so we can still use break and continue
|
|
* with the same meanings as above.
|
|
*/
|
|
do_next_attr_loop:
|
|
if ((u8 *)a < (u8 *)ctx->mrec ||
|
|
(u8 *)a >= (u8 *)ctx->mrec + le32_to_cpu(ctx->mrec->bytes_allocated) ||
|
|
(u8 *)a >= (u8 *)ctx->mrec + le32_to_cpu(ctx->mrec->bytes_in_use))
|
|
break;
|
|
|
|
mft_free_len = le32_to_cpu(ctx->mrec->bytes_in_use) -
|
|
((u8 *)a - (u8 *)ctx->mrec);
|
|
if (mft_free_len >= sizeof(a->type) && a->type == AT_END)
|
|
continue;
|
|
|
|
attr_len = le32_to_cpu(a->length);
|
|
if (!attr_len ||
|
|
attr_len < offsetof(struct attr_record, data.resident.reserved) +
|
|
sizeof(a->data.resident.reserved) ||
|
|
attr_len > mft_free_len)
|
|
break;
|
|
|
|
if (al_entry->instance != a->instance)
|
|
goto do_next_attr;
|
|
/*
|
|
* If the type and/or the name are mismatched between the
|
|
* attribute list entry and the attribute record, there is
|
|
* corruption so we break and return error EIO.
|
|
*/
|
|
if (al_entry->type != a->type)
|
|
break;
|
|
if (a->name_length && ((le16_to_cpu(a->name_offset) +
|
|
a->name_length * sizeof(__le16)) > attr_len))
|
|
break;
|
|
if (!ntfs_are_names_equal((__le16 *)((u8 *)a +
|
|
le16_to_cpu(a->name_offset)), a->name_length,
|
|
al_name, al_name_len, CASE_SENSITIVE,
|
|
vol->upcase, vol->upcase_len))
|
|
break;
|
|
|
|
ctx->attr = a;
|
|
|
|
if (a->non_resident) {
|
|
u32 min_len;
|
|
u16 mp_offset;
|
|
|
|
min_len = offsetof(struct attr_record,
|
|
data.non_resident.initialized_size) +
|
|
sizeof(a->data.non_resident.initialized_size);
|
|
|
|
if (le32_to_cpu(a->length) < min_len)
|
|
break;
|
|
|
|
mp_offset =
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset);
|
|
if (mp_offset < min_len || mp_offset > attr_len)
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* If no @val specified or @val specified and it matches, we
|
|
* have found it!
|
|
*/
|
|
if ((type == AT_UNUSED) || !val)
|
|
goto attr_found;
|
|
if (!a->non_resident) {
|
|
u32 value_length = le32_to_cpu(a->data.resident.value_length);
|
|
u16 value_offset = le16_to_cpu(a->data.resident.value_offset);
|
|
|
|
if (attr_len < offsetof(struct attr_record, data.resident.reserved) +
|
|
sizeof(a->data.resident.reserved))
|
|
break;
|
|
if (value_length > attr_len || value_offset > attr_len - value_length)
|
|
break;
|
|
|
|
value_length = ntfs_resident_attr_min_value_length(a->type);
|
|
if (value_length && le32_to_cpu(a->data.resident.value_length) <
|
|
value_length) {
|
|
pr_err("Too small resident attribute value in MFT record %lld, type %#x\n",
|
|
(long long)ctx->ntfs_ino->mft_no, a->type);
|
|
break;
|
|
}
|
|
if (value_length == val_len &&
|
|
!memcmp((u8 *)a + value_offset, val, val_len)) {
|
|
attr_found:
|
|
ntfs_debug("Done, found.");
|
|
return 0;
|
|
}
|
|
}
|
|
do_next_attr:
|
|
/* Proceed to the next attribute in the current mft record. */
|
|
a = (struct attr_record *)((u8 *)a + attr_len);
|
|
goto do_next_attr_loop;
|
|
}
|
|
|
|
corrupt:
|
|
if (ni != base_ni) {
|
|
if (ni)
|
|
unmap_extent_mft_record(ni);
|
|
ctx->ntfs_ino = base_ni;
|
|
ctx->mrec = ctx->base_mrec;
|
|
ctx->attr = ctx->base_attr;
|
|
ctx->mapped_mrec = ctx->mapped_base_mrec;
|
|
}
|
|
|
|
if (!err) {
|
|
u64 mft_no = ctx->al_entry ? MREF_LE(ctx->al_entry->mft_reference) : 0;
|
|
u32 type = ctx->al_entry ? le32_to_cpu(ctx->al_entry->type) : 0;
|
|
|
|
ntfs_error(vol->sb,
|
|
"Base inode 0x%llx contains corrupt attribute, mft %#llx, type %#x. %s",
|
|
(long long)base_ni->mft_no, (long long)mft_no, type,
|
|
"Unmount and run chkdsk.");
|
|
err = -EIO;
|
|
}
|
|
|
|
if (err != -ENOMEM)
|
|
NVolSetErrors(vol);
|
|
return err;
|
|
not_found:
|
|
/*
|
|
* If we were looking for AT_END, we reset the search context @ctx and
|
|
* use ntfs_attr_find() to seek to the end of the base mft record.
|
|
*/
|
|
if (type == AT_UNUSED || type == AT_END) {
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
return ntfs_attr_find(AT_END, name, name_len, ic, val, val_len,
|
|
ctx);
|
|
}
|
|
/*
|
|
* The attribute was not found. Before we return, we want to ensure
|
|
* @ctx->mrec and @ctx->attr indicate the position at which the
|
|
* attribute should be inserted in the base mft record. Since we also
|
|
* want to preserve @ctx->al_entry we cannot reinitialize the search
|
|
* context using ntfs_attr_reinit_search_ctx() as this would set
|
|
* @ctx->al_entry to NULL. Thus we do the necessary bits manually (see
|
|
* ntfs_attr_init_search_ctx() below). Note, we _only_ preserve
|
|
* @ctx->al_entry as the remaining fields (base_*) are identical to
|
|
* their non base_ counterparts and we cannot set @ctx->base_attr
|
|
* correctly yet as we do not know what @ctx->attr will be set to by
|
|
* the call to ntfs_attr_find() below.
|
|
*/
|
|
if (ni != base_ni)
|
|
unmap_extent_mft_record(ni);
|
|
ctx->mrec = ctx->base_mrec;
|
|
ctx->attr = (struct attr_record *)((u8 *)ctx->mrec +
|
|
le16_to_cpu(ctx->mrec->attrs_offset));
|
|
ctx->is_first = true;
|
|
ctx->ntfs_ino = base_ni;
|
|
ctx->base_ntfs_ino = NULL;
|
|
ctx->base_mrec = NULL;
|
|
ctx->base_attr = NULL;
|
|
ctx->mapped_mrec = ctx->mapped_base_mrec;
|
|
/*
|
|
* In case there are multiple matches in the base mft record, need to
|
|
* keep enumerating until we get an attribute not found response (or
|
|
* another error), otherwise we would keep returning the same attribute
|
|
* over and over again and all programs using us for enumeration would
|
|
* lock up in a tight loop.
|
|
*/
|
|
do {
|
|
err = ntfs_attr_find(type, name, name_len, ic, val, val_len,
|
|
ctx);
|
|
} while (!err);
|
|
ntfs_debug("Done, not found.");
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_lookup - find an attribute in an ntfs inode
|
|
* @type: attribute type to find
|
|
* @name: attribute name to find (optional, i.e. NULL means don't care)
|
|
* @name_len: attribute name length (only needed if @name present)
|
|
* @ic: IGNORE_CASE or CASE_SENSITIVE (ignored if @name not present)
|
|
* @lowest_vcn: lowest vcn to find (optional, non-resident attributes only)
|
|
* @val: attribute value to find (optional, resident attributes only)
|
|
* @val_len: attribute value length
|
|
* @ctx: search context with mft record and attribute to search from
|
|
*
|
|
* Find an attribute in an ntfs inode. On first search @ctx->ntfs_ino must
|
|
* be the base mft record and @ctx must have been obtained from a call to
|
|
* ntfs_attr_get_search_ctx().
|
|
*
|
|
* This function transparently handles attribute lists and @ctx is used to
|
|
* continue searches where they were left off at.
|
|
*
|
|
* After finishing with the attribute/mft record you need to call
|
|
* ntfs_attr_put_search_ctx() to cleanup the search context (unmapping any
|
|
* mapped inodes, etc).
|
|
*
|
|
* Return 0 if the search was successful and -errno if not.
|
|
*
|
|
* When 0, @ctx->attr is the found attribute and it is in mft record
|
|
* @ctx->mrec. If an attribute list attribute is present, @ctx->al_entry is
|
|
* the attribute list entry of the found attribute.
|
|
*
|
|
* When -ENOENT, @ctx->attr is the attribute which collates just after the
|
|
* attribute being searched for, i.e. if one wants to add the attribute to the
|
|
* mft record this is the correct place to insert it into. If an attribute
|
|
* list attribute is present, @ctx->al_entry is the attribute list entry which
|
|
* collates just after the attribute list entry of the attribute being searched
|
|
* for, i.e. if one wants to add the attribute to the mft record this is the
|
|
* correct place to insert its attribute list entry into.
|
|
*/
|
|
int ntfs_attr_lookup(const __le32 type, const __le16 *name,
|
|
const u32 name_len, const u32 ic,
|
|
const s64 lowest_vcn, const u8 *val, const u32 val_len,
|
|
struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
struct ntfs_inode *base_ni;
|
|
|
|
ntfs_debug("Entering.");
|
|
if (ctx->base_ntfs_ino)
|
|
base_ni = ctx->base_ntfs_ino;
|
|
else
|
|
base_ni = ctx->ntfs_ino;
|
|
/* Sanity check, just for debugging really. */
|
|
if (!base_ni || !NInoAttrList(base_ni) || type == AT_ATTRIBUTE_LIST)
|
|
return ntfs_attr_find(type, name, name_len, ic, val, val_len,
|
|
ctx);
|
|
return ntfs_external_attr_find(type, name, name_len, ic, lowest_vcn,
|
|
val, val_len, ctx);
|
|
}
|
|
|
|
/**
|
|
* ntfs_attr_init_search_ctx - initialize an attribute search context
|
|
* @ctx: attribute search context to initialize
|
|
* @ni: ntfs inode with which to initialize the search context
|
|
* @mrec: mft record with which to initialize the search context
|
|
*
|
|
* Initialize the attribute search context @ctx with @ni and @mrec.
|
|
*/
|
|
static bool ntfs_attr_init_search_ctx(struct ntfs_attr_search_ctx *ctx,
|
|
struct ntfs_inode *ni, struct mft_record *mrec)
|
|
{
|
|
if (!mrec) {
|
|
mrec = map_mft_record(ni);
|
|
if (IS_ERR(mrec))
|
|
return false;
|
|
ctx->mapped_mrec = true;
|
|
} else {
|
|
ctx->mapped_mrec = false;
|
|
}
|
|
|
|
ctx->mrec = mrec;
|
|
/* Sanity checks are performed elsewhere. */
|
|
ctx->attr = (struct attr_record *)((u8 *)mrec + le16_to_cpu(mrec->attrs_offset));
|
|
ctx->is_first = true;
|
|
ctx->ntfs_ino = ni;
|
|
ctx->al_entry = NULL;
|
|
ctx->base_ntfs_ino = NULL;
|
|
ctx->base_mrec = NULL;
|
|
ctx->base_attr = NULL;
|
|
ctx->mapped_base_mrec = false;
|
|
return true;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_reinit_search_ctx - reinitialize an attribute search context
|
|
* @ctx: attribute search context to reinitialize
|
|
*
|
|
* Reinitialize the attribute search context @ctx, unmapping an associated
|
|
* extent mft record if present, and initialize the search context again.
|
|
*
|
|
* This is used when a search for a new attribute is being started to reset
|
|
* the search context to the beginning.
|
|
*/
|
|
void ntfs_attr_reinit_search_ctx(struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
bool mapped_mrec;
|
|
|
|
if (likely(!ctx->base_ntfs_ino)) {
|
|
/* No attribute list. */
|
|
ctx->is_first = true;
|
|
/* Sanity checks are performed elsewhere. */
|
|
ctx->attr = (struct attr_record *)((u8 *)ctx->mrec +
|
|
le16_to_cpu(ctx->mrec->attrs_offset));
|
|
/*
|
|
* This needs resetting due to ntfs_external_attr_find() which
|
|
* can leave it set despite having zeroed ctx->base_ntfs_ino.
|
|
*/
|
|
ctx->al_entry = NULL;
|
|
return;
|
|
} /* Attribute list. */
|
|
if (ctx->ntfs_ino != ctx->base_ntfs_ino && ctx->ntfs_ino)
|
|
unmap_extent_mft_record(ctx->ntfs_ino);
|
|
|
|
mapped_mrec = ctx->mapped_base_mrec;
|
|
ntfs_attr_init_search_ctx(ctx, ctx->base_ntfs_ino, ctx->base_mrec);
|
|
ctx->mapped_mrec = mapped_mrec;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_get_search_ctx - allocate/initialize a new attribute search context
|
|
* @ni: ntfs inode with which to initialize the search context
|
|
* @mrec: mft record with which to initialize the search context
|
|
*
|
|
* Allocate a new attribute search context, initialize it with @ni and @mrec,
|
|
* and return it. Return NULL if allocation failed.
|
|
*/
|
|
struct ntfs_attr_search_ctx *ntfs_attr_get_search_ctx(struct ntfs_inode *ni,
|
|
struct mft_record *mrec)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
bool init;
|
|
|
|
ctx = kmem_cache_alloc(ntfs_attr_ctx_cache, GFP_NOFS);
|
|
if (ctx) {
|
|
init = ntfs_attr_init_search_ctx(ctx, ni, mrec);
|
|
if (init == false) {
|
|
kmem_cache_free(ntfs_attr_ctx_cache, ctx);
|
|
ctx = NULL;
|
|
}
|
|
}
|
|
|
|
return ctx;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_put_search_ctx - release an attribute search context
|
|
* @ctx: attribute search context to free
|
|
*
|
|
* Release the attribute search context @ctx, unmapping an associated extent
|
|
* mft record if present.
|
|
*/
|
|
void ntfs_attr_put_search_ctx(struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
if (ctx->mapped_mrec)
|
|
unmap_mft_record(ctx->ntfs_ino);
|
|
|
|
if (ctx->mapped_base_mrec && ctx->base_ntfs_ino &&
|
|
ctx->ntfs_ino != ctx->base_ntfs_ino)
|
|
unmap_extent_mft_record(ctx->base_ntfs_ino);
|
|
kmem_cache_free(ntfs_attr_ctx_cache, ctx);
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_find_in_attrdef - find an attribute in the $AttrDef system file
|
|
* @vol: ntfs volume to which the attribute belongs
|
|
* @type: attribute type which to find
|
|
*
|
|
* Search for the attribute definition record corresponding to the attribute
|
|
* @type in the $AttrDef system file.
|
|
*
|
|
* Return the attribute type definition record if found and NULL if not found.
|
|
*/
|
|
static struct attr_def *ntfs_attr_find_in_attrdef(const struct ntfs_volume *vol,
|
|
const __le32 type)
|
|
{
|
|
struct attr_def *ad;
|
|
|
|
WARN_ON(!type);
|
|
for (ad = vol->attrdef; (u8 *)ad - (u8 *)vol->attrdef <
|
|
vol->attrdef_size && ad->type; ++ad) {
|
|
/* We have not found it yet, carry on searching. */
|
|
if (likely(le32_to_cpu(ad->type) < le32_to_cpu(type)))
|
|
continue;
|
|
/* We found the attribute; return it. */
|
|
if (likely(ad->type == type))
|
|
return ad;
|
|
/* We have gone too far already. No point in continuing. */
|
|
break;
|
|
}
|
|
/* Attribute not found. */
|
|
ntfs_debug("Attribute type 0x%x not found in $AttrDef.",
|
|
le32_to_cpu(type));
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_size_bounds_check - check a size of an attribute type for validity
|
|
* @vol: ntfs volume to which the attribute belongs
|
|
* @type: attribute type which to check
|
|
* @size: size which to check
|
|
*
|
|
* Check whether the @size in bytes is valid for an attribute of @type on the
|
|
* ntfs volume @vol. This information is obtained from $AttrDef system file.
|
|
*/
|
|
int ntfs_attr_size_bounds_check(const struct ntfs_volume *vol, const __le32 type,
|
|
const s64 size)
|
|
{
|
|
struct attr_def *ad;
|
|
|
|
if (size < 0)
|
|
return -EINVAL;
|
|
|
|
/*
|
|
* $ATTRIBUTE_LIST has a maximum size of 256kiB, but this is not
|
|
* listed in $AttrDef.
|
|
*/
|
|
if (unlikely(type == AT_ATTRIBUTE_LIST && size > 256 * 1024))
|
|
return -ERANGE;
|
|
/* Get the $AttrDef entry for the attribute @type. */
|
|
ad = ntfs_attr_find_in_attrdef(vol, type);
|
|
if (unlikely(!ad))
|
|
return -ENOENT;
|
|
/* Do the bounds check. */
|
|
if (((le64_to_cpu(ad->min_size) > 0) &&
|
|
size < le64_to_cpu(ad->min_size)) ||
|
|
((le64_to_cpu(ad->max_size) > 0) && size >
|
|
le64_to_cpu(ad->max_size)))
|
|
return -ERANGE;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_can_be_non_resident - check if an attribute can be non-resident
|
|
* @vol: ntfs volume to which the attribute belongs
|
|
* @type: attribute type which to check
|
|
*
|
|
* Check whether the attribute of @type on the ntfs volume @vol is allowed to
|
|
* be non-resident. This information is obtained from $AttrDef system file.
|
|
*/
|
|
static int ntfs_attr_can_be_non_resident(const struct ntfs_volume *vol,
|
|
const __le32 type)
|
|
{
|
|
struct attr_def *ad;
|
|
|
|
/* Find the attribute definition record in $AttrDef. */
|
|
ad = ntfs_attr_find_in_attrdef(vol, type);
|
|
if (unlikely(!ad))
|
|
return -ENOENT;
|
|
/* Check the flags and return the result. */
|
|
if (ad->flags & ATTR_DEF_RESIDENT)
|
|
return -EPERM;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_can_be_resident - check if an attribute can be resident
|
|
* @vol: ntfs volume to which the attribute belongs
|
|
* @type: attribute type which to check
|
|
*
|
|
* Check whether the attribute of @type on the ntfs volume @vol is allowed to
|
|
* be resident. This information is derived from our ntfs knowledge and may
|
|
* not be completely accurate, especially when user defined attributes are
|
|
* present. Basically we allow everything to be resident except for index
|
|
* allocation and $EA attributes.
|
|
*
|
|
* Return 0 if the attribute is allowed to be non-resident and -EPERM if not.
|
|
*
|
|
* Warning: In the system file $MFT the attribute $Bitmap must be non-resident
|
|
* otherwise windows will not boot (blue screen of death)! We cannot
|
|
* check for this here as we do not know which inode's $Bitmap is
|
|
* being asked about so the caller needs to special case this.
|
|
*/
|
|
int ntfs_attr_can_be_resident(const struct ntfs_volume *vol, const __le32 type)
|
|
{
|
|
if (type == AT_INDEX_ALLOCATION)
|
|
return -EPERM;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_record_resize - resize an attribute record
|
|
* @m: mft record containing attribute record
|
|
* @a: attribute record to resize
|
|
* @new_size: new size in bytes to which to resize the attribute record @a
|
|
*
|
|
* Resize the attribute record @a, i.e. the resident part of the attribute, in
|
|
* the mft record @m to @new_size bytes.
|
|
*/
|
|
int ntfs_attr_record_resize(struct mft_record *m, struct attr_record *a, u32 new_size)
|
|
{
|
|
u32 old_size, alloc_size, attr_size;
|
|
|
|
old_size = le32_to_cpu(m->bytes_in_use);
|
|
alloc_size = le32_to_cpu(m->bytes_allocated);
|
|
attr_size = le32_to_cpu(a->length);
|
|
|
|
ntfs_debug("Sizes: old=%u alloc=%u attr=%u new=%u\n",
|
|
(unsigned int)old_size, (unsigned int)alloc_size,
|
|
(unsigned int)attr_size, (unsigned int)new_size);
|
|
|
|
/* Align to 8 bytes if it is not already done. */
|
|
if (new_size & 7)
|
|
new_size = (new_size + 7) & ~7;
|
|
/* If the actual attribute length has changed, move things around. */
|
|
if (new_size != attr_size) {
|
|
u32 new_muse = le32_to_cpu(m->bytes_in_use) -
|
|
attr_size + new_size;
|
|
/* Not enough space in this mft record. */
|
|
if (new_muse > le32_to_cpu(m->bytes_allocated))
|
|
return -ENOSPC;
|
|
|
|
if (a->type == AT_INDEX_ROOT && new_size > attr_size &&
|
|
new_muse + 120 > alloc_size && old_size + 120 <= alloc_size) {
|
|
ntfs_debug("Too big struct index_root (%u > %u)\n",
|
|
new_muse, alloc_size);
|
|
return -ENOSPC;
|
|
}
|
|
|
|
/* Move attributes following @a to their new location. */
|
|
memmove((u8 *)a + new_size, (u8 *)a + le32_to_cpu(a->length),
|
|
le32_to_cpu(m->bytes_in_use) - ((u8 *)a -
|
|
(u8 *)m) - attr_size);
|
|
/* Adjust @m to reflect the change in used space. */
|
|
m->bytes_in_use = cpu_to_le32(new_muse);
|
|
/* Adjust @a to reflect the new size. */
|
|
if (new_size >= offsetof(struct attr_record, length) + sizeof(a->length))
|
|
a->length = cpu_to_le32(new_size);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_resident_attr_value_resize - resize the value of a resident attribute
|
|
* @m: mft record containing attribute record
|
|
* @a: attribute record whose value to resize
|
|
* @new_size: new size in bytes to which to resize the attribute value of @a
|
|
*
|
|
* Resize the value of the attribute @a in the mft record @m to @new_size bytes.
|
|
* If the value is made bigger, the newly allocated space is cleared.
|
|
*/
|
|
int ntfs_resident_attr_value_resize(struct mft_record *m, struct attr_record *a,
|
|
const u32 new_size)
|
|
{
|
|
u32 old_size;
|
|
|
|
/* Resize the resident part of the attribute record. */
|
|
if (ntfs_attr_record_resize(m, a,
|
|
le16_to_cpu(a->data.resident.value_offset) + new_size))
|
|
return -ENOSPC;
|
|
/*
|
|
* The resize succeeded! If we made the attribute value bigger, clear
|
|
* the area between the old size and @new_size.
|
|
*/
|
|
old_size = le32_to_cpu(a->data.resident.value_length);
|
|
if (new_size > old_size)
|
|
memset((u8 *)a + le16_to_cpu(a->data.resident.value_offset) +
|
|
old_size, 0, new_size - old_size);
|
|
/* Finally update the length of the attribute value. */
|
|
a->data.resident.value_length = cpu_to_le32(new_size);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_make_non_resident - convert a resident to a non-resident attribute
|
|
* @ni: ntfs inode describing the attribute to convert
|
|
* @data_size: size of the resident data to copy to the non-resident attribute
|
|
*
|
|
* Convert the resident ntfs attribute described by the ntfs inode @ni to a
|
|
* non-resident one.
|
|
*
|
|
* @data_size must be equal to the attribute value size. This is needed since
|
|
* we need to know the size before we can map the mft record and our callers
|
|
* always know it. The reason we cannot simply read the size from the vfs
|
|
* inode i_size is that this is not necessarily uptodate. This happens when
|
|
* ntfs_attr_make_non_resident() is called in the ->truncate call path(s).
|
|
*/
|
|
int ntfs_attr_make_non_resident(struct ntfs_inode *ni, const u32 data_size)
|
|
{
|
|
s64 new_size;
|
|
struct inode *vi = VFS_I(ni);
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct ntfs_inode *base_ni;
|
|
struct mft_record *m;
|
|
struct attr_record *a;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
struct folio *folio;
|
|
struct runlist_element *rl;
|
|
unsigned long flags;
|
|
int mp_size, mp_ofs, name_ofs, arec_size, err, err2;
|
|
u32 attr_size;
|
|
u8 old_res_attr_flags;
|
|
|
|
if (NInoNonResident(ni)) {
|
|
ntfs_warning(vol->sb,
|
|
"Trying to make non-resident attribute non-resident. Aborting...\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Check that the attribute is allowed to be non-resident. */
|
|
err = ntfs_attr_can_be_non_resident(vol, ni->type);
|
|
if (unlikely(err)) {
|
|
if (err == -EPERM)
|
|
ntfs_debug("Attribute is not allowed to be non-resident.");
|
|
else
|
|
ntfs_debug("Attribute not defined on the NTFS volume!");
|
|
return err;
|
|
}
|
|
|
|
if (NInoEncrypted(ni))
|
|
return -EIO;
|
|
|
|
if (!NInoAttr(ni))
|
|
base_ni = ni;
|
|
else
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
m = map_mft_record(base_ni);
|
|
if (IS_ERR(m)) {
|
|
err = PTR_ERR(m);
|
|
m = NULL;
|
|
ctx = NULL;
|
|
goto err_out;
|
|
}
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, m);
|
|
if (unlikely(!ctx)) {
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx);
|
|
if (unlikely(err)) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
goto err_out;
|
|
}
|
|
m = ctx->mrec;
|
|
a = ctx->attr;
|
|
|
|
/*
|
|
* The size needs to be aligned to a cluster boundary for allocation
|
|
* purposes.
|
|
*/
|
|
new_size = (data_size + vol->cluster_size - 1) &
|
|
~(vol->cluster_size - 1);
|
|
if (new_size > 0) {
|
|
if ((a->flags & ATTR_COMPRESSION_MASK) == ATTR_IS_COMPRESSED) {
|
|
/* must allocate full compression blocks */
|
|
new_size =
|
|
((new_size - 1) |
|
|
((1L << (STANDARD_COMPRESSION_UNIT +
|
|
vol->cluster_size_bits)) - 1)) + 1;
|
|
}
|
|
|
|
/*
|
|
* Will need folio later and since folio lock nests
|
|
* outside all ntfs locks, we need to get the folio now.
|
|
*/
|
|
folio = __filemap_get_folio(vi->i_mapping, 0,
|
|
FGP_CREAT | FGP_LOCK,
|
|
mapping_gfp_mask(vi->i_mapping));
|
|
if (IS_ERR(folio)) {
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
|
|
/* Start by allocating clusters to hold the attribute value. */
|
|
rl = ntfs_cluster_alloc(vol, 0,
|
|
ntfs_bytes_to_cluster(vol, new_size),
|
|
-1, DATA_ZONE, true, false, false);
|
|
if (IS_ERR(rl)) {
|
|
err = PTR_ERR(rl);
|
|
ntfs_debug("Failed to allocate cluster%s, error code %i.",
|
|
ntfs_bytes_to_cluster(vol, new_size) > 1 ? "s" : "",
|
|
err);
|
|
goto folio_err_out;
|
|
}
|
|
} else {
|
|
rl = NULL;
|
|
folio = NULL;
|
|
}
|
|
|
|
down_write(&ni->runlist.lock);
|
|
/* Determine the size of the mapping pairs array. */
|
|
mp_size = ntfs_get_size_for_mapping_pairs(vol, rl, 0, -1, -1);
|
|
if (unlikely(mp_size < 0)) {
|
|
err = mp_size;
|
|
ntfs_debug("Failed to get size for mapping pairs array, error code %i.\n", err);
|
|
goto rl_err_out;
|
|
}
|
|
|
|
if (NInoNonResident(ni) || a->non_resident) {
|
|
err = -EIO;
|
|
goto rl_err_out;
|
|
}
|
|
|
|
/*
|
|
* Calculate new offsets for the name and the mapping pairs array.
|
|
*/
|
|
if (NInoSparse(ni) || NInoCompressed(ni))
|
|
name_ofs = (offsetof(struct attr_record,
|
|
data.non_resident.compressed_size) +
|
|
sizeof(a->data.non_resident.compressed_size) +
|
|
7) & ~7;
|
|
else
|
|
name_ofs = (offsetof(struct attr_record,
|
|
data.non_resident.compressed_size) + 7) & ~7;
|
|
mp_ofs = (name_ofs + a->name_length * sizeof(__le16) + 7) & ~7;
|
|
/*
|
|
* Determine the size of the resident part of the now non-resident
|
|
* attribute record.
|
|
*/
|
|
arec_size = (mp_ofs + mp_size + 7) & ~7;
|
|
/*
|
|
* If the folio is not uptodate bring it uptodate by copying from the
|
|
* attribute value.
|
|
*/
|
|
attr_size = le32_to_cpu(a->data.resident.value_length);
|
|
WARN_ON(attr_size != data_size);
|
|
if (folio && !folio_test_uptodate(folio)) {
|
|
folio_fill_tail(folio, 0, (u8 *)a +
|
|
le16_to_cpu(a->data.resident.value_offset),
|
|
attr_size);
|
|
folio_mark_uptodate(folio);
|
|
}
|
|
|
|
/* Backup the attribute flag. */
|
|
old_res_attr_flags = a->data.resident.flags;
|
|
/* Resize the resident part of the attribute record. */
|
|
err = ntfs_attr_record_resize(m, a, arec_size);
|
|
if (unlikely(err))
|
|
goto rl_err_out;
|
|
|
|
/*
|
|
* Convert the resident part of the attribute record to describe a
|
|
* non-resident attribute.
|
|
*/
|
|
a->non_resident = 1;
|
|
/* Move the attribute name if it exists and update the offset. */
|
|
if (a->name_length)
|
|
memmove((u8 *)a + name_ofs, (u8 *)a + le16_to_cpu(a->name_offset),
|
|
a->name_length * sizeof(__le16));
|
|
a->name_offset = cpu_to_le16(name_ofs);
|
|
/* Setup the fields specific to non-resident attributes. */
|
|
a->data.non_resident.lowest_vcn = 0;
|
|
a->data.non_resident.highest_vcn =
|
|
cpu_to_le64(ntfs_bytes_to_cluster(vol, new_size - 1));
|
|
a->data.non_resident.mapping_pairs_offset = cpu_to_le16(mp_ofs);
|
|
memset(&a->data.non_resident.reserved, 0,
|
|
sizeof(a->data.non_resident.reserved));
|
|
a->data.non_resident.allocated_size = cpu_to_le64(new_size);
|
|
a->data.non_resident.data_size =
|
|
a->data.non_resident.initialized_size =
|
|
cpu_to_le64(attr_size);
|
|
if (NInoSparse(ni) || NInoCompressed(ni)) {
|
|
a->data.non_resident.compression_unit = 0;
|
|
if (NInoCompressed(ni) || vol->major_ver < 3)
|
|
a->data.non_resident.compression_unit = 4;
|
|
a->data.non_resident.compressed_size =
|
|
a->data.non_resident.allocated_size;
|
|
} else
|
|
a->data.non_resident.compression_unit = 0;
|
|
/* Generate the mapping pairs array into the attribute record. */
|
|
err = ntfs_mapping_pairs_build(vol, (u8 *)a + mp_ofs,
|
|
arec_size - mp_ofs, rl, 0, -1, NULL, NULL, NULL);
|
|
if (unlikely(err)) {
|
|
ntfs_error(vol->sb, "Failed to build mapping pairs, error code %i.",
|
|
err);
|
|
goto undo_err_out;
|
|
}
|
|
|
|
/* Setup the in-memory attribute structure to be non-resident. */
|
|
ni->runlist.rl = rl;
|
|
if (rl) {
|
|
for (ni->runlist.count = 1; rl->length != 0; rl++)
|
|
ni->runlist.count++;
|
|
} else
|
|
ni->runlist.count = 0;
|
|
write_lock_irqsave(&ni->size_lock, flags);
|
|
ni->allocated_size = new_size;
|
|
if (NInoSparse(ni) || NInoCompressed(ni)) {
|
|
ni->itype.compressed.size = ni->allocated_size;
|
|
if (a->data.non_resident.compression_unit) {
|
|
ni->itype.compressed.block_size = 1U <<
|
|
(a->data.non_resident.compression_unit +
|
|
vol->cluster_size_bits);
|
|
ni->itype.compressed.block_size_bits =
|
|
ffs(ni->itype.compressed.block_size) -
|
|
1;
|
|
ni->itype.compressed.block_clusters = 1U <<
|
|
a->data.non_resident.compression_unit;
|
|
} else {
|
|
ni->itype.compressed.block_size = 0;
|
|
ni->itype.compressed.block_size_bits = 0;
|
|
ni->itype.compressed.block_clusters = 0;
|
|
}
|
|
vi->i_blocks = ni->itype.compressed.size >> 9;
|
|
} else
|
|
vi->i_blocks = ni->allocated_size >> 9;
|
|
write_unlock_irqrestore(&ni->size_lock, flags);
|
|
/*
|
|
* This needs to be last since the address space operations ->read_folio
|
|
* and ->writepage can run concurrently with us as they are not
|
|
* serialized on i_mutex. Note, we are not allowed to fail once we flip
|
|
* this switch, which is another reason to do this last.
|
|
*/
|
|
NInoSetNonResident(ni);
|
|
NInoSetFullyMapped(ni);
|
|
/* Mark the mft record dirty, so it gets written back. */
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
unmap_mft_record(base_ni);
|
|
up_write(&ni->runlist.lock);
|
|
if (folio) {
|
|
iomap_dirty_folio(vi->i_mapping, folio);
|
|
folio_unlock(folio);
|
|
folio_put(folio);
|
|
}
|
|
ntfs_debug("Done.");
|
|
return 0;
|
|
undo_err_out:
|
|
/* Convert the attribute back into a resident attribute. */
|
|
a->non_resident = 0;
|
|
/* Move the attribute name if it exists and update the offset. */
|
|
name_ofs = (offsetof(struct attr_record, data.resident.reserved) +
|
|
sizeof(a->data.resident.reserved) + 7) & ~7;
|
|
if (a->name_length)
|
|
memmove((u8 *)a + name_ofs, (u8 *)a + le16_to_cpu(a->name_offset),
|
|
a->name_length * sizeof(__le16));
|
|
mp_ofs = (name_ofs + a->name_length * sizeof(__le16) + 7) & ~7;
|
|
a->name_offset = cpu_to_le16(name_ofs);
|
|
arec_size = (mp_ofs + attr_size + 7) & ~7;
|
|
/* Resize the resident part of the attribute record. */
|
|
err2 = ntfs_attr_record_resize(m, a, arec_size);
|
|
if (unlikely(err2)) {
|
|
/*
|
|
* This cannot happen (well if memory corruption is at work it
|
|
* could happen in theory), but deal with it as well as we can.
|
|
* If the old size is too small, truncate the attribute,
|
|
* otherwise simply give it a larger allocated size.
|
|
*/
|
|
arec_size = le32_to_cpu(a->length);
|
|
if ((mp_ofs + attr_size) > arec_size) {
|
|
err2 = attr_size;
|
|
attr_size = arec_size - mp_ofs;
|
|
ntfs_error(vol->sb,
|
|
"Failed to undo partial resident to non-resident attribute conversion. Truncating inode 0x%llx, attribute type 0x%x from %i bytes to %i bytes to maintain metadata consistency. THIS MEANS YOU ARE LOSING %i BYTES DATA FROM THIS %s.",
|
|
ni->mft_no,
|
|
(unsigned int)le32_to_cpu(ni->type),
|
|
err2, attr_size, err2 - attr_size,
|
|
((ni->type == AT_DATA) &&
|
|
!ni->name_len) ? "FILE" : "ATTRIBUTE");
|
|
write_lock_irqsave(&ni->size_lock, flags);
|
|
ni->initialized_size = attr_size;
|
|
i_size_write(vi, attr_size);
|
|
write_unlock_irqrestore(&ni->size_lock, flags);
|
|
}
|
|
}
|
|
/* Setup the fields specific to resident attributes. */
|
|
a->data.resident.value_length = cpu_to_le32(attr_size);
|
|
a->data.resident.value_offset = cpu_to_le16(mp_ofs);
|
|
a->data.resident.flags = old_res_attr_flags;
|
|
memset(&a->data.resident.reserved, 0,
|
|
sizeof(a->data.resident.reserved));
|
|
/* Copy the data from folio back to the attribute value. */
|
|
if (folio)
|
|
memcpy_from_folio((u8 *)a + mp_ofs, folio, 0, attr_size);
|
|
/* Setup the allocated size in the ntfs inode in case it changed. */
|
|
write_lock_irqsave(&ni->size_lock, flags);
|
|
ni->allocated_size = arec_size - mp_ofs;
|
|
write_unlock_irqrestore(&ni->size_lock, flags);
|
|
/* Mark the mft record dirty, so it gets written back. */
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
rl_err_out:
|
|
up_write(&ni->runlist.lock);
|
|
if (rl) {
|
|
if (ntfs_cluster_free_from_rl(vol, rl) < 0) {
|
|
ntfs_error(vol->sb,
|
|
"Failed to release allocated cluster(s) in error code path. Run chkdsk to recover the lost cluster(s).");
|
|
NVolSetErrors(vol);
|
|
}
|
|
kvfree(rl);
|
|
folio_err_out:
|
|
folio_unlock(folio);
|
|
folio_put(folio);
|
|
}
|
|
err_out:
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
if (m)
|
|
unmap_mft_record(base_ni);
|
|
ni->runlist.rl = NULL;
|
|
|
|
if (err == -EINVAL)
|
|
err = -EIO;
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_set - fill (a part of) an attribute with a byte
|
|
* @ni: ntfs inode describing the attribute to fill
|
|
* @ofs: offset inside the attribute at which to start to fill
|
|
* @cnt: number of bytes to fill
|
|
* @val: the unsigned 8-bit value with which to fill the attribute
|
|
*
|
|
* Fill @cnt bytes of the attribute described by the ntfs inode @ni starting at
|
|
* byte offset @ofs inside the attribute with the constant byte @val.
|
|
*
|
|
* This function is effectively like memset() applied to an ntfs attribute.
|
|
* Note thie function actually only operates on the page cache pages belonging
|
|
* to the ntfs attribute and it marks them dirty after doing the memset().
|
|
* Thus it relies on the vm dirty page write code paths to cause the modified
|
|
* pages to be written to the mft record/disk.
|
|
*/
|
|
int ntfs_attr_set(struct ntfs_inode *ni, s64 ofs, s64 cnt, const u8 val)
|
|
{
|
|
struct address_space *mapping = VFS_I(ni)->i_mapping;
|
|
struct folio *folio;
|
|
pgoff_t index;
|
|
u8 *addr;
|
|
unsigned long offset;
|
|
size_t attr_len;
|
|
int ret = 0;
|
|
|
|
index = ofs >> PAGE_SHIFT;
|
|
while (cnt) {
|
|
folio = read_mapping_folio(mapping, index, NULL);
|
|
if (IS_ERR(folio)) {
|
|
ret = PTR_ERR(folio);
|
|
ntfs_error(VFS_I(ni)->i_sb, "Failed to read a page %lu for attr %#x: %ld",
|
|
index, ni->type, PTR_ERR(folio));
|
|
break;
|
|
}
|
|
|
|
offset = offset_in_folio(folio, ofs);
|
|
attr_len = min_t(size_t, (size_t)cnt, folio_size(folio) - offset);
|
|
|
|
folio_lock(folio);
|
|
addr = kmap_local_folio(folio, offset);
|
|
memset(addr, val, attr_len);
|
|
kunmap_local(addr);
|
|
|
|
folio_mark_dirty(folio);
|
|
folio_unlock(folio);
|
|
folio_put(folio);
|
|
|
|
ofs += attr_len;
|
|
cnt -= attr_len;
|
|
index++;
|
|
cond_resched();
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_attr_set_initialized_size(struct ntfs_inode *ni, loff_t new_size)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
int err = 0;
|
|
|
|
if (!NInoNonResident(ni))
|
|
return -EINVAL;
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx)
|
|
return -ENOMEM;
|
|
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx);
|
|
if (err)
|
|
goto out_ctx;
|
|
|
|
ctx->attr->data.non_resident.initialized_size = cpu_to_le64(new_size);
|
|
ni->initialized_size = new_size;
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
out_ctx:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_make_room_for_attr - make room for an attribute inside an mft record
|
|
* @m: mft record
|
|
* @pos: position at which to make space
|
|
* @size: byte size to make available at this position
|
|
*
|
|
* @pos points to the attribute in front of which we want to make space.
|
|
*/
|
|
static int ntfs_make_room_for_attr(struct mft_record *m, u8 *pos, u32 size)
|
|
{
|
|
u32 biu;
|
|
|
|
ntfs_debug("Entering for pos 0x%x, size %u.\n",
|
|
(int)(pos - (u8 *)m), (unsigned int) size);
|
|
|
|
/* Make size 8-byte alignment. */
|
|
size = (size + 7) & ~7;
|
|
|
|
/* Rigorous consistency checks. */
|
|
if (!m || !pos || pos < (u8 *)m) {
|
|
pr_err("%s: pos=%p m=%p\n", __func__, pos, m);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* The -8 is for the attribute terminator. */
|
|
if (pos - (u8 *)m > (int)le32_to_cpu(m->bytes_in_use) - 8)
|
|
return -EINVAL;
|
|
/* Nothing to do. */
|
|
if (!size)
|
|
return 0;
|
|
|
|
biu = le32_to_cpu(m->bytes_in_use);
|
|
/* Do we have enough space? */
|
|
if (biu + size > le32_to_cpu(m->bytes_allocated) ||
|
|
pos + size > (u8 *)m + le32_to_cpu(m->bytes_allocated)) {
|
|
ntfs_debug("No enough space in the MFT record\n");
|
|
return -ENOSPC;
|
|
}
|
|
/* Move everything after pos to pos + size. */
|
|
memmove(pos + size, pos, biu - (pos - (u8 *)m));
|
|
/* Update mft record. */
|
|
m->bytes_in_use = cpu_to_le32(biu + size);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_resident_attr_record_add - add resident attribute to inode
|
|
* @ni: opened ntfs inode to which MFT record add attribute
|
|
* @type: type of the new attribute
|
|
* @name: name of the new attribute
|
|
* @name_len: name length of the new attribute
|
|
* @val: value of the new attribute
|
|
* @size: size of new attribute (length of @val, if @val != NULL)
|
|
* @flags: flags of the new attribute
|
|
*/
|
|
int ntfs_resident_attr_record_add(struct ntfs_inode *ni, __le32 type,
|
|
__le16 *name, u8 name_len, u8 *val, u32 size,
|
|
__le16 flags)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
u32 length;
|
|
struct attr_record *a;
|
|
struct mft_record *m;
|
|
int err, offset;
|
|
struct ntfs_inode *base_ni;
|
|
|
|
if (!ni || (!name && name_len))
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x, flags 0x%x.\n",
|
|
(long long) ni->mft_no, (unsigned int) le32_to_cpu(type),
|
|
(unsigned int) le16_to_cpu(flags));
|
|
|
|
err = ntfs_attr_can_be_resident(ni->vol, type);
|
|
if (err) {
|
|
if (err == -EPERM)
|
|
ntfs_debug("Attribute can't be resident.\n");
|
|
else
|
|
ntfs_debug("ntfs_attr_can_be_resident failed.\n");
|
|
return err;
|
|
}
|
|
|
|
/* Locate place where record should be. */
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(ni->vol->sb, "%s: Failed to get search context",
|
|
__func__);
|
|
return -ENOMEM;
|
|
}
|
|
/*
|
|
* Use ntfs_attr_find instead of ntfs_attr_lookup to find place for
|
|
* attribute in @ni->mrec, not any extent inode in case if @ni is base
|
|
* file record.
|
|
*/
|
|
err = ntfs_attr_find(type, name, name_len, CASE_SENSITIVE, val, size, ctx);
|
|
if (!err) {
|
|
err = -EEXIST;
|
|
ntfs_debug("Attribute already present.\n");
|
|
goto put_err_out;
|
|
}
|
|
if (err != -ENOENT) {
|
|
err = -EIO;
|
|
goto put_err_out;
|
|
}
|
|
a = ctx->attr;
|
|
m = ctx->mrec;
|
|
|
|
/* Make room for attribute. */
|
|
length = offsetof(struct attr_record, data.resident.reserved) +
|
|
sizeof(a->data.resident.reserved) +
|
|
((name_len * sizeof(__le16) + 7) & ~7) +
|
|
((size + 7) & ~7);
|
|
err = ntfs_make_room_for_attr(ctx->mrec, (u8 *) ctx->attr, length);
|
|
if (err) {
|
|
ntfs_debug("Failed to make room for attribute.\n");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Setup record fields. */
|
|
offset = ((u8 *)a - (u8 *)m);
|
|
a->type = type;
|
|
a->length = cpu_to_le32(length);
|
|
a->non_resident = 0;
|
|
a->name_length = name_len;
|
|
a->name_offset =
|
|
name_len ? cpu_to_le16((offsetof(struct attr_record, data.resident.reserved) +
|
|
sizeof(a->data.resident.reserved))) : cpu_to_le16(0);
|
|
|
|
a->flags = flags;
|
|
a->instance = m->next_attr_instance;
|
|
a->data.resident.value_length = cpu_to_le32(size);
|
|
a->data.resident.value_offset = cpu_to_le16(length - ((size + 7) & ~7));
|
|
if (val)
|
|
memcpy((u8 *)a + le16_to_cpu(a->data.resident.value_offset), val, size);
|
|
else
|
|
memset((u8 *)a + le16_to_cpu(a->data.resident.value_offset), 0, size);
|
|
if (type == AT_FILE_NAME)
|
|
a->data.resident.flags = RESIDENT_ATTR_IS_INDEXED;
|
|
else
|
|
a->data.resident.flags = 0;
|
|
if (name_len)
|
|
memcpy((u8 *)a + le16_to_cpu(a->name_offset),
|
|
name, sizeof(__le16) * name_len);
|
|
m->next_attr_instance =
|
|
cpu_to_le16((le16_to_cpu(m->next_attr_instance) + 1) & 0xffff);
|
|
if (ni->nr_extents == -1)
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
if (type != AT_ATTRIBUTE_LIST && NInoAttrList(base_ni)) {
|
|
err = ntfs_attrlist_entry_add(ni, a);
|
|
if (err) {
|
|
ntfs_attr_record_resize(m, a, 0);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
ntfs_debug("Failed add attribute entry to ATTRIBUTE_LIST.\n");
|
|
goto put_err_out;
|
|
}
|
|
}
|
|
mark_mft_record_dirty(ni);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return offset;
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return -EIO;
|
|
}
|
|
|
|
/*
|
|
* ntfs_non_resident_attr_record_add - add extent of non-resident attribute
|
|
* @ni: opened ntfs inode to which MFT record add attribute
|
|
* @type: type of the new attribute extent
|
|
* @name: name of the new attribute extent
|
|
* @name_len: name length of the new attribute extent
|
|
* @lowest_vcn: lowest vcn of the new attribute extent
|
|
* @dataruns_size: dataruns size of the new attribute extent
|
|
* @flags: flags of the new attribute extent
|
|
*/
|
|
static int ntfs_non_resident_attr_record_add(struct ntfs_inode *ni, __le32 type,
|
|
__le16 *name, u8 name_len, s64 lowest_vcn, int dataruns_size,
|
|
__le16 flags)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
u32 length;
|
|
struct attr_record *a;
|
|
struct mft_record *m;
|
|
struct ntfs_inode *base_ni;
|
|
int err, offset;
|
|
|
|
if (!ni || dataruns_size <= 0 || (!name && name_len))
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x, lowest_vcn %lld, dataruns_size %d, flags 0x%x.\n",
|
|
(long long) ni->mft_no, (unsigned int) le32_to_cpu(type),
|
|
(long long) lowest_vcn, dataruns_size,
|
|
(unsigned int) le16_to_cpu(flags));
|
|
|
|
err = ntfs_attr_can_be_non_resident(ni->vol, type);
|
|
if (err) {
|
|
if (err == -EPERM)
|
|
pr_err("Attribute can't be non resident\n");
|
|
else
|
|
pr_err("ntfs_attr_can_be_non_resident failed\n");
|
|
return err;
|
|
}
|
|
|
|
/* Locate place where record should be. */
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
pr_err("%s: Failed to get search context\n", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
/*
|
|
* Use ntfs_attr_find instead of ntfs_attr_lookup to find place for
|
|
* attribute in @ni->mrec, not any extent inode in case if @ni is base
|
|
* file record.
|
|
*/
|
|
err = ntfs_attr_find(type, name, name_len, CASE_SENSITIVE, NULL, 0, ctx);
|
|
if (!err) {
|
|
err = -EEXIST;
|
|
pr_err("Attribute 0x%x already present\n", type);
|
|
goto put_err_out;
|
|
}
|
|
if (err != -ENOENT) {
|
|
pr_err("ntfs_attr_find failed\n");
|
|
err = -EIO;
|
|
goto put_err_out;
|
|
}
|
|
a = ctx->attr;
|
|
m = ctx->mrec;
|
|
|
|
/* Make room for attribute. */
|
|
dataruns_size = (dataruns_size + 7) & ~7;
|
|
length = offsetof(struct attr_record, data.non_resident.compressed_size) +
|
|
((sizeof(__le16) * name_len + 7) & ~7) + dataruns_size +
|
|
((flags & (ATTR_IS_COMPRESSED | ATTR_IS_SPARSE)) ?
|
|
sizeof(a->data.non_resident.compressed_size) : 0);
|
|
err = ntfs_make_room_for_attr(ctx->mrec, (u8 *) ctx->attr, length);
|
|
if (err) {
|
|
pr_err("Failed to make room for attribute\n");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Setup record fields. */
|
|
a->type = type;
|
|
a->length = cpu_to_le32(length);
|
|
a->non_resident = 1;
|
|
a->name_length = name_len;
|
|
a->name_offset = cpu_to_le16(offsetof(struct attr_record,
|
|
data.non_resident.compressed_size) +
|
|
((flags & (ATTR_IS_COMPRESSED | ATTR_IS_SPARSE)) ?
|
|
sizeof(a->data.non_resident.compressed_size) : 0));
|
|
a->flags = flags;
|
|
a->instance = m->next_attr_instance;
|
|
a->data.non_resident.lowest_vcn = cpu_to_le64(lowest_vcn);
|
|
a->data.non_resident.mapping_pairs_offset = cpu_to_le16(length - dataruns_size);
|
|
a->data.non_resident.compression_unit =
|
|
(flags & ATTR_IS_COMPRESSED) ? STANDARD_COMPRESSION_UNIT : 0;
|
|
/* If @lowest_vcn == 0, than setup empty attribute. */
|
|
if (!lowest_vcn) {
|
|
a->data.non_resident.highest_vcn = cpu_to_le64(-1);
|
|
a->data.non_resident.allocated_size = 0;
|
|
a->data.non_resident.data_size = 0;
|
|
a->data.non_resident.initialized_size = 0;
|
|
/* Set empty mapping pairs. */
|
|
*((u8 *)a + le16_to_cpu(a->data.non_resident.mapping_pairs_offset)) = 0;
|
|
}
|
|
if (name_len)
|
|
memcpy((u8 *)a + le16_to_cpu(a->name_offset),
|
|
name, sizeof(__le16) * name_len);
|
|
m->next_attr_instance =
|
|
cpu_to_le16((le16_to_cpu(m->next_attr_instance) + 1) & 0xffff);
|
|
if (ni->nr_extents == -1)
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
if (type != AT_ATTRIBUTE_LIST && NInoAttrList(base_ni)) {
|
|
err = ntfs_attrlist_entry_add(ni, a);
|
|
if (err) {
|
|
pr_err("Failed add attr entry to attrlist\n");
|
|
ntfs_attr_record_resize(m, a, 0);
|
|
goto put_err_out;
|
|
}
|
|
}
|
|
mark_mft_record_dirty(ni);
|
|
/*
|
|
* Locate offset from start of the MFT record where new attribute is
|
|
* placed. We need relookup it, because record maybe moved during
|
|
* update of attribute list.
|
|
*/
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
err = ntfs_attr_lookup(type, name, name_len, CASE_SENSITIVE,
|
|
lowest_vcn, NULL, 0, ctx);
|
|
if (err) {
|
|
pr_err("%s: attribute lookup failed\n", __func__);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
|
|
}
|
|
offset = (u8 *)ctx->attr - (u8 *)ctx->mrec;
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return offset;
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return -1;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_record_rm - remove attribute extent
|
|
* @ctx: search context describing the attribute which should be removed
|
|
*
|
|
* If this function succeed, user should reinit search context if he/she wants
|
|
* use it anymore.
|
|
*/
|
|
int ntfs_attr_record_rm(struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
struct ntfs_inode *base_ni, *ni;
|
|
__le32 type;
|
|
int err;
|
|
|
|
if (!ctx || !ctx->ntfs_ino || !ctx->mrec || !ctx->attr)
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x.\n",
|
|
(long long) ctx->ntfs_ino->mft_no,
|
|
(unsigned int) le32_to_cpu(ctx->attr->type));
|
|
type = ctx->attr->type;
|
|
ni = ctx->ntfs_ino;
|
|
if (ctx->base_ntfs_ino)
|
|
base_ni = ctx->base_ntfs_ino;
|
|
else
|
|
base_ni = ctx->ntfs_ino;
|
|
|
|
/* Remove attribute itself. */
|
|
if (ntfs_attr_record_resize(ctx->mrec, ctx->attr, 0)) {
|
|
ntfs_debug("Couldn't remove attribute record. Bug or damaged MFT record.\n");
|
|
return -EIO;
|
|
}
|
|
mark_mft_record_dirty(ni);
|
|
|
|
/*
|
|
* Remove record from $ATTRIBUTE_LIST if present and we don't want
|
|
* delete $ATTRIBUTE_LIST itself.
|
|
*/
|
|
if (NInoAttrList(base_ni) && type != AT_ATTRIBUTE_LIST) {
|
|
err = ntfs_attrlist_entry_rm(ctx);
|
|
if (err) {
|
|
ntfs_debug("Couldn't delete record from $ATTRIBUTE_LIST.\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* Post $ATTRIBUTE_LIST delete setup. */
|
|
if (type == AT_ATTRIBUTE_LIST) {
|
|
if (NInoAttrList(base_ni) && base_ni->attr_list)
|
|
kvfree(base_ni->attr_list);
|
|
base_ni->attr_list = NULL;
|
|
NInoClearAttrList(base_ni);
|
|
}
|
|
|
|
/* Free MFT record, if it doesn't contain attributes. */
|
|
if (le32_to_cpu(ctx->mrec->bytes_in_use) -
|
|
le16_to_cpu(ctx->mrec->attrs_offset) == 8) {
|
|
if (ntfs_mft_record_free(ni->vol, ni)) {
|
|
ntfs_debug("Couldn't free MFT record.\n");
|
|
return -EIO;
|
|
}
|
|
/* Remove done if we freed base inode. */
|
|
if (ni == base_ni)
|
|
return 0;
|
|
ntfs_inode_close(ni);
|
|
ctx->ntfs_ino = ni = NULL;
|
|
}
|
|
|
|
if (type == AT_ATTRIBUTE_LIST || !NInoAttrList(base_ni))
|
|
return 0;
|
|
|
|
/* Remove attribute list if we don't need it any more. */
|
|
if (!ntfs_attrlist_need(base_ni)) {
|
|
struct ntfs_attr na;
|
|
struct inode *attr_vi;
|
|
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
if (ntfs_attr_lookup(AT_ATTRIBUTE_LIST, NULL, 0, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx)) {
|
|
ntfs_debug("Couldn't find attribute list. Succeed anyway.\n");
|
|
return 0;
|
|
}
|
|
/* Deallocate clusters. */
|
|
if (ctx->attr->non_resident) {
|
|
struct runlist_element *al_rl;
|
|
size_t new_rl_count;
|
|
|
|
al_rl = ntfs_mapping_pairs_decompress(base_ni->vol,
|
|
ctx->attr, NULL, &new_rl_count);
|
|
if (IS_ERR(al_rl)) {
|
|
ntfs_debug("Couldn't decompress attribute list runlist. Succeed anyway.\n");
|
|
return 0;
|
|
}
|
|
if (ntfs_cluster_free_from_rl(base_ni->vol, al_rl))
|
|
ntfs_debug("Leaking clusters! Run chkdsk. Couldn't free clusters from attribute list runlist.\n");
|
|
kvfree(al_rl);
|
|
}
|
|
/* Remove attribute record itself. */
|
|
if (ntfs_attr_record_rm(ctx)) {
|
|
ntfs_debug("Couldn't remove attribute list. Succeed anyway.\n");
|
|
return 0;
|
|
}
|
|
|
|
na.mft_no = VFS_I(base_ni)->i_ino;
|
|
na.type = AT_ATTRIBUTE_LIST;
|
|
na.name = NULL;
|
|
na.name_len = 0;
|
|
|
|
attr_vi = ilookup5(VFS_I(base_ni)->i_sb, VFS_I(base_ni)->i_ino,
|
|
ntfs_test_inode, &na);
|
|
if (attr_vi) {
|
|
clear_nlink(attr_vi);
|
|
iput(attr_vi);
|
|
}
|
|
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_add - add attribute to inode
|
|
* @ni: opened ntfs inode to which add attribute
|
|
* @type: type of the new attribute
|
|
* @name: name in unicode of the new attribute
|
|
* @name_len: name length in unicode characters of the new attribute
|
|
* @val: value of new attribute
|
|
* @size: size of the new attribute / length of @val (if specified)
|
|
*
|
|
* @val should always be specified for always resident attributes (eg. FILE_NAME
|
|
* attribute), for attributes that can become non-resident @val can be NULL
|
|
* (eg. DATA attribute). @size can be specified even if @val is NULL, in this
|
|
* case data size will be equal to @size and initialized size will be equal
|
|
* to 0.
|
|
*
|
|
* If inode haven't got enough space to add attribute, add attribute to one of
|
|
* it extents, if no extents present or no one of them have enough space, than
|
|
* allocate new extent and add attribute to it.
|
|
*
|
|
* If on one of this steps attribute list is needed but not present, than it is
|
|
* added transparently to caller. So, this function should not be called with
|
|
* @type == AT_ATTRIBUTE_LIST, if you really need to add attribute list call
|
|
* ntfs_inode_add_attrlist instead.
|
|
*
|
|
* On success return 0. On error return -1 with errno set to the error code.
|
|
*/
|
|
int ntfs_attr_add(struct ntfs_inode *ni, __le32 type,
|
|
__le16 *name, u8 name_len, u8 *val, s64 size)
|
|
{
|
|
struct super_block *sb;
|
|
u32 attr_rec_size;
|
|
int err, i, offset;
|
|
bool is_resident;
|
|
bool can_be_non_resident = false;
|
|
struct ntfs_inode *attr_ni;
|
|
struct inode *attr_vi;
|
|
struct mft_record *ni_mrec;
|
|
|
|
if (!ni || size < 0 || type == AT_ATTRIBUTE_LIST)
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr %x, size %lld.\n",
|
|
(long long) ni->mft_no, type, size);
|
|
|
|
if (ni->nr_extents == -1)
|
|
ni = ni->ext.base_ntfs_ino;
|
|
|
|
/* Check the attribute type and the size. */
|
|
err = ntfs_attr_size_bounds_check(ni->vol, type, size);
|
|
if (err) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
return err;
|
|
}
|
|
|
|
sb = ni->vol->sb;
|
|
/* Sanity checks for always resident attributes. */
|
|
err = ntfs_attr_can_be_non_resident(ni->vol, type);
|
|
if (err) {
|
|
if (err != -EPERM) {
|
|
ntfs_error(sb, "ntfs_attr_can_be_non_resident failed");
|
|
goto err_out;
|
|
}
|
|
/* @val is mandatory. */
|
|
if (!val) {
|
|
ntfs_error(sb,
|
|
"val is mandatory for always resident attributes");
|
|
return -EINVAL;
|
|
}
|
|
if (size > ni->vol->mft_record_size) {
|
|
ntfs_error(sb, "Attribute is too big");
|
|
return -ERANGE;
|
|
}
|
|
} else
|
|
can_be_non_resident = true;
|
|
|
|
/*
|
|
* Determine resident or not will be new attribute. We add 8 to size in
|
|
* non resident case for mapping pairs.
|
|
*/
|
|
err = ntfs_attr_can_be_resident(ni->vol, type);
|
|
if (!err) {
|
|
is_resident = true;
|
|
} else {
|
|
if (err != -EPERM) {
|
|
ntfs_error(sb, "ntfs_attr_can_be_resident failed");
|
|
goto err_out;
|
|
}
|
|
is_resident = false;
|
|
}
|
|
|
|
/* Calculate attribute record size. */
|
|
if (is_resident)
|
|
attr_rec_size = offsetof(struct attr_record, data.resident.reserved) +
|
|
1 +
|
|
((name_len * sizeof(__le16) + 7) & ~7) +
|
|
((size + 7) & ~7);
|
|
else
|
|
attr_rec_size = offsetof(struct attr_record, data.non_resident.compressed_size) +
|
|
((name_len * sizeof(__le16) + 7) & ~7) + 8;
|
|
|
|
/*
|
|
* If we have enough free space for the new attribute in the base MFT
|
|
* record, then add attribute to it.
|
|
*/
|
|
retry:
|
|
ni_mrec = map_mft_record(ni);
|
|
if (IS_ERR(ni_mrec)) {
|
|
err = -EIO;
|
|
goto err_out;
|
|
}
|
|
|
|
if (le32_to_cpu(ni_mrec->bytes_allocated) -
|
|
le32_to_cpu(ni_mrec->bytes_in_use) >= attr_rec_size) {
|
|
attr_ni = ni;
|
|
unmap_mft_record(ni);
|
|
goto add_attr_record;
|
|
}
|
|
unmap_mft_record(ni);
|
|
|
|
/* Try to add to extent inodes. */
|
|
err = ntfs_inode_attach_all_extents(ni);
|
|
if (err) {
|
|
ntfs_error(sb, "Failed to attach all extents to inode");
|
|
goto err_out;
|
|
}
|
|
|
|
for (i = 0; i < ni->nr_extents; i++) {
|
|
attr_ni = ni->ext.extent_ntfs_inos[i];
|
|
ni_mrec = map_mft_record(attr_ni);
|
|
if (IS_ERR(ni_mrec)) {
|
|
err = -EIO;
|
|
goto err_out;
|
|
}
|
|
|
|
if (le32_to_cpu(ni_mrec->bytes_allocated) -
|
|
le32_to_cpu(ni_mrec->bytes_in_use) >=
|
|
attr_rec_size) {
|
|
unmap_mft_record(attr_ni);
|
|
goto add_attr_record;
|
|
}
|
|
unmap_mft_record(attr_ni);
|
|
}
|
|
|
|
/* There is no extent that contain enough space for new attribute. */
|
|
if (!NInoAttrList(ni)) {
|
|
/* Add attribute list not present, add it and retry. */
|
|
err = ntfs_inode_add_attrlist(ni);
|
|
if (err) {
|
|
ntfs_error(sb, "Failed to add attribute list");
|
|
goto err_out;
|
|
}
|
|
goto retry;
|
|
}
|
|
|
|
attr_ni = NULL;
|
|
/* Allocate new extent. */
|
|
err = ntfs_mft_record_alloc(ni->vol, 0, &attr_ni, ni, NULL);
|
|
if (err) {
|
|
ntfs_error(sb, "Failed to allocate extent record");
|
|
goto err_out;
|
|
}
|
|
unmap_mft_record(attr_ni);
|
|
|
|
add_attr_record:
|
|
if (is_resident) {
|
|
/* Add resident attribute. */
|
|
offset = ntfs_resident_attr_record_add(attr_ni, type, name,
|
|
name_len, val, size, 0);
|
|
if (offset < 0) {
|
|
if (offset == -ENOSPC && can_be_non_resident)
|
|
goto add_non_resident;
|
|
err = offset;
|
|
ntfs_error(sb, "Failed to add resident attribute");
|
|
goto free_err_out;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
add_non_resident:
|
|
/* Add non resident attribute. */
|
|
offset = ntfs_non_resident_attr_record_add(attr_ni, type, name,
|
|
name_len, 0, 8, 0);
|
|
if (offset < 0) {
|
|
err = offset;
|
|
ntfs_error(sb, "Failed to add non resident attribute");
|
|
goto free_err_out;
|
|
}
|
|
|
|
/* If @size == 0, we are done. */
|
|
if (!size)
|
|
return 0;
|
|
|
|
/* Open new attribute and resize it. */
|
|
attr_vi = ntfs_attr_iget(VFS_I(ni), type, name, name_len);
|
|
if (IS_ERR(attr_vi)) {
|
|
err = PTR_ERR(attr_vi);
|
|
ntfs_error(sb, "Failed to open just added attribute");
|
|
goto rm_attr_err_out;
|
|
}
|
|
attr_ni = NTFS_I(attr_vi);
|
|
|
|
/* Resize and set attribute value. */
|
|
if (ntfs_attr_truncate(attr_ni, size) ||
|
|
(val && (ntfs_inode_attr_pwrite(attr_vi, 0, size, val, false) != size))) {
|
|
err = -EIO;
|
|
ntfs_error(sb, "Failed to initialize just added attribute");
|
|
if (ntfs_attr_rm(attr_ni))
|
|
ntfs_error(sb, "Failed to remove just added attribute");
|
|
iput(attr_vi);
|
|
goto err_out;
|
|
}
|
|
iput(attr_vi);
|
|
return 0;
|
|
|
|
rm_attr_err_out:
|
|
/* Remove just added attribute. */
|
|
ni_mrec = map_mft_record(attr_ni);
|
|
if (!IS_ERR(ni_mrec)) {
|
|
if (ntfs_attr_record_resize(ni_mrec,
|
|
(struct attr_record *)((u8 *)ni_mrec + offset), 0))
|
|
ntfs_error(sb, "Failed to remove just added attribute #2");
|
|
unmap_mft_record(attr_ni);
|
|
} else
|
|
pr_err("EIO when try to remove new added attr\n");
|
|
|
|
free_err_out:
|
|
/* Free MFT record, if it doesn't contain attributes. */
|
|
ni_mrec = map_mft_record(attr_ni);
|
|
if (!IS_ERR(ni_mrec)) {
|
|
int attr_size;
|
|
|
|
attr_size = le32_to_cpu(ni_mrec->bytes_in_use) -
|
|
le16_to_cpu(ni_mrec->attrs_offset);
|
|
unmap_mft_record(attr_ni);
|
|
if (attr_size == 8) {
|
|
if (ntfs_mft_record_free(attr_ni->vol, attr_ni))
|
|
ntfs_error(sb, "Failed to free MFT record");
|
|
if (attr_ni->nr_extents < 0)
|
|
ntfs_inode_close(attr_ni);
|
|
}
|
|
} else
|
|
pr_err("EIO when testing mft record is free-able\n");
|
|
|
|
err_out:
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* __ntfs_attr_init - primary initialization of an ntfs attribute structure
|
|
* @ni: ntfs attribute inode to initialize
|
|
* @ni: ntfs inode with which to initialize the ntfs attribute
|
|
* @type: attribute type
|
|
* @name: attribute name in little endian Unicode or NULL
|
|
* @name_len: length of attribute @name in Unicode characters (if @name given)
|
|
*
|
|
* Initialize the ntfs attribute @na with @ni, @type, @name, and @name_len.
|
|
*/
|
|
static void __ntfs_attr_init(struct ntfs_inode *ni,
|
|
const __le32 type, __le16 *name, const u32 name_len)
|
|
{
|
|
ni->runlist.rl = NULL;
|
|
ni->type = type;
|
|
ni->name = name;
|
|
if (name)
|
|
ni->name_len = name_len;
|
|
else
|
|
ni->name_len = 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_init - initialize an ntfs_attr with data sizes and status
|
|
* @ni: ntfs inode to initialize
|
|
* @non_resident: true if attribute is non-resident
|
|
* @compressed: true if attribute is compressed
|
|
* @encrypted: true if attribute is encrypted
|
|
* @sparse: true if attribute is sparse
|
|
* @allocated_size: allocated size of the attribute
|
|
* @data_size: actual data size of the attribute
|
|
* @initialized_size: initialized size of the attribute
|
|
* @compressed_size: compressed size (if compressed or sparse)
|
|
* @compression_unit: compression unit size (log2 of clusters)
|
|
*
|
|
* Final initialization for an ntfs attribute.
|
|
*/
|
|
static void ntfs_attr_init(struct ntfs_inode *ni, const bool non_resident,
|
|
const bool compressed, const bool encrypted, const bool sparse,
|
|
const s64 allocated_size, const s64 data_size,
|
|
const s64 initialized_size, const s64 compressed_size,
|
|
const u8 compression_unit)
|
|
{
|
|
if (non_resident)
|
|
NInoSetNonResident(ni);
|
|
if (compressed) {
|
|
NInoSetCompressed(ni);
|
|
ni->flags |= FILE_ATTR_COMPRESSED;
|
|
}
|
|
if (encrypted) {
|
|
NInoSetEncrypted(ni);
|
|
ni->flags |= FILE_ATTR_ENCRYPTED;
|
|
}
|
|
if (sparse) {
|
|
NInoSetSparse(ni);
|
|
ni->flags |= FILE_ATTR_SPARSE_FILE;
|
|
}
|
|
ni->allocated_size = allocated_size;
|
|
ni->data_size = data_size;
|
|
ni->initialized_size = initialized_size;
|
|
if (compressed || sparse) {
|
|
struct ntfs_volume *vol = ni->vol;
|
|
|
|
ni->itype.compressed.size = compressed_size;
|
|
ni->itype.compressed.block_clusters = 1 << compression_unit;
|
|
ni->itype.compressed.block_size = 1 << (compression_unit +
|
|
vol->cluster_size_bits);
|
|
ni->itype.compressed.block_size_bits = ffs(
|
|
ni->itype.compressed.block_size) - 1;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_open - open an ntfs attribute for access
|
|
* @ni: open ntfs inode in which the ntfs attribute resides
|
|
* @type: attribute type
|
|
* @name: attribute name in little endian Unicode or AT_UNNAMED or NULL
|
|
* @name_len: length of attribute @name in Unicode characters (if @name given)
|
|
*/
|
|
int ntfs_attr_open(struct ntfs_inode *ni, const __le32 type,
|
|
__le16 *name, u32 name_len)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
__le16 *newname = NULL;
|
|
struct attr_record *a;
|
|
bool cs;
|
|
struct ntfs_inode *base_ni;
|
|
int err;
|
|
|
|
ntfs_debug("Entering for inode %lld, attr 0x%x.\n",
|
|
(unsigned long long)ni->mft_no, type);
|
|
|
|
if (!ni || !ni->vol)
|
|
return -EINVAL;
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
if (name && name != AT_UNNAMED && name != I30) {
|
|
name = ntfs_ucsndup(name, name_len);
|
|
if (!name) {
|
|
err = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
newname = name;
|
|
}
|
|
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
err = -ENOMEM;
|
|
pr_err("%s: Failed to get search context\n", __func__);
|
|
goto err_out;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(type, name, name_len, 0, 0, NULL, 0, ctx);
|
|
if (err)
|
|
goto put_err_out;
|
|
|
|
a = ctx->attr;
|
|
|
|
if (!name) {
|
|
if (a->name_length) {
|
|
name = ntfs_ucsndup((__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length);
|
|
if (!name)
|
|
goto put_err_out;
|
|
newname = name;
|
|
name_len = a->name_length;
|
|
} else {
|
|
name = AT_UNNAMED;
|
|
name_len = 0;
|
|
}
|
|
}
|
|
|
|
__ntfs_attr_init(ni, type, name, name_len);
|
|
|
|
/*
|
|
* Wipe the flags in case they are not zero for an attribute list
|
|
* attribute. Windows does not complain about invalid flags and chkdsk
|
|
* does not detect or fix them so we need to cope with it, too.
|
|
*/
|
|
if (type == AT_ATTRIBUTE_LIST)
|
|
a->flags = 0;
|
|
|
|
if ((type == AT_DATA) &&
|
|
(a->non_resident ? !a->data.non_resident.initialized_size :
|
|
!a->data.resident.value_length)) {
|
|
/*
|
|
* Define/redefine the compression state if stream is
|
|
* empty, based on the compression mark on parent
|
|
* directory (for unnamed data streams) or on current
|
|
* inode (for named data streams). The compression mark
|
|
* may change any time, the compression state can only
|
|
* change when stream is wiped out.
|
|
*
|
|
* Also prevent compression on NTFS version < 3.0
|
|
* or cluster size > 4K or compression is disabled
|
|
*/
|
|
a->flags &= ~ATTR_COMPRESSION_MASK;
|
|
if (NInoCompressed(ni)
|
|
&& (ni->vol->major_ver >= 3)
|
|
&& NVolCompression(ni->vol)
|
|
&& (ni->vol->cluster_size <= MAX_COMPRESSION_CLUSTER_SIZE))
|
|
a->flags |= ATTR_IS_COMPRESSED;
|
|
}
|
|
|
|
cs = a->flags & (ATTR_IS_COMPRESSED | ATTR_IS_SPARSE);
|
|
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED &&
|
|
((!(a->flags & ATTR_IS_COMPRESSED) != !NInoCompressed(ni)) ||
|
|
(!(a->flags & ATTR_IS_SPARSE) != !NInoSparse(ni)) ||
|
|
(!(a->flags & ATTR_IS_ENCRYPTED) != !NInoEncrypted(ni)))) {
|
|
err = -EIO;
|
|
pr_err("Inode %lld has corrupt attribute flags (0x%x <> 0x%x)\n",
|
|
(unsigned long long)ni->mft_no,
|
|
a->flags, ni->flags);
|
|
goto put_err_out;
|
|
}
|
|
|
|
if (a->non_resident) {
|
|
if (((a->flags & ATTR_COMPRESSION_MASK) || a->data.non_resident.compression_unit) &&
|
|
(ni->vol->major_ver < 3)) {
|
|
err = -EIO;
|
|
pr_err("Compressed inode %lld not allowed on NTFS %d.%d\n",
|
|
(unsigned long long)ni->mft_no,
|
|
ni->vol->major_ver,
|
|
ni->vol->major_ver);
|
|
goto put_err_out;
|
|
}
|
|
|
|
if ((a->flags & ATTR_IS_COMPRESSED) && !a->data.non_resident.compression_unit) {
|
|
err = -EIO;
|
|
pr_err("Compressed inode %lld attr 0x%x has no compression unit\n",
|
|
(unsigned long long)ni->mft_no, type);
|
|
goto put_err_out;
|
|
}
|
|
if ((a->flags & ATTR_COMPRESSION_MASK) &&
|
|
(a->data.non_resident.compression_unit != STANDARD_COMPRESSION_UNIT)) {
|
|
err = -EIO;
|
|
pr_err("Compressed inode %lld attr 0x%lx has an unsupported compression unit %d\n",
|
|
(unsigned long long)ni->mft_no,
|
|
(long)le32_to_cpu(type),
|
|
(int)a->data.non_resident.compression_unit);
|
|
goto put_err_out;
|
|
}
|
|
ntfs_attr_init(ni, true, a->flags & ATTR_IS_COMPRESSED,
|
|
a->flags & ATTR_IS_ENCRYPTED,
|
|
a->flags & ATTR_IS_SPARSE,
|
|
le64_to_cpu(a->data.non_resident.allocated_size),
|
|
le64_to_cpu(a->data.non_resident.data_size),
|
|
le64_to_cpu(a->data.non_resident.initialized_size),
|
|
cs ? le64_to_cpu(a->data.non_resident.compressed_size) : 0,
|
|
cs ? a->data.non_resident.compression_unit : 0);
|
|
} else {
|
|
s64 l = le32_to_cpu(a->data.resident.value_length);
|
|
|
|
ntfs_attr_init(ni, false, a->flags & ATTR_IS_COMPRESSED,
|
|
a->flags & ATTR_IS_ENCRYPTED,
|
|
a->flags & ATTR_IS_SPARSE, (l + 7) & ~7, l, l,
|
|
cs ? (l + 7) & ~7 : 0, 0);
|
|
}
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
out:
|
|
ntfs_debug("\n");
|
|
return err;
|
|
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
err_out:
|
|
kfree(newname);
|
|
goto out;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_close - free an ntfs attribute structure
|
|
* @ni: ntfs inode to free
|
|
*
|
|
* Release all memory associated with the ntfs attribute @na and then release
|
|
* @na itself.
|
|
*/
|
|
void ntfs_attr_close(struct ntfs_inode *ni)
|
|
{
|
|
if (NInoNonResident(ni) && ni->runlist.rl)
|
|
kvfree(ni->runlist.rl);
|
|
/* Don't release if using an internal constant. */
|
|
if (ni->name != AT_UNNAMED && ni->name != I30)
|
|
kfree(ni->name);
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_map_whole_runlist - map the whole runlist of an ntfs attribute
|
|
* @ni: ntfs inode for which to map the runlist
|
|
*
|
|
* Map the whole runlist of the ntfs attribute @na. For an attribute made up
|
|
* of only one attribute extent this is the same as calling
|
|
* ntfs_map_runlist(ni, 0) but for an attribute with multiple extents this
|
|
* will map the runlist fragments from each of the extents thus giving access
|
|
* to the entirety of the disk allocation of an attribute.
|
|
*/
|
|
int ntfs_attr_map_whole_runlist(struct ntfs_inode *ni)
|
|
{
|
|
s64 next_vcn, last_vcn, highest_vcn;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct super_block *sb = vol->sb;
|
|
struct attr_record *a;
|
|
int err;
|
|
struct ntfs_inode *base_ni;
|
|
int not_mapped;
|
|
size_t new_rl_count;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x.\n",
|
|
(unsigned long long)ni->mft_no, ni->type);
|
|
|
|
if (NInoFullyMapped(ni) && ni->runlist.rl)
|
|
return 0;
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Map all attribute extents one by one. */
|
|
next_vcn = last_vcn = highest_vcn = 0;
|
|
a = NULL;
|
|
while (1) {
|
|
struct runlist_element *rl;
|
|
|
|
not_mapped = 0;
|
|
if (ntfs_rl_vcn_to_lcn(ni->runlist.rl, next_vcn) == LCN_RL_NOT_MAPPED)
|
|
not_mapped = 1;
|
|
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, next_vcn, NULL, 0, ctx);
|
|
if (err)
|
|
break;
|
|
|
|
a = ctx->attr;
|
|
|
|
if (not_mapped) {
|
|
/* Decode the runlist. */
|
|
rl = ntfs_mapping_pairs_decompress(ni->vol, a, &ni->runlist,
|
|
&new_rl_count);
|
|
if (IS_ERR(rl)) {
|
|
err = PTR_ERR(rl);
|
|
goto err_out;
|
|
}
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_count;
|
|
}
|
|
|
|
/* Are we in the first extent? */
|
|
if (!next_vcn) {
|
|
if (a->data.non_resident.lowest_vcn) {
|
|
err = -EIO;
|
|
ntfs_error(sb,
|
|
"First extent of inode %llu attribute has non-zero lowest_vcn",
|
|
(unsigned long long)ni->mft_no);
|
|
goto err_out;
|
|
}
|
|
/* Get the last vcn in the attribute. */
|
|
last_vcn = ntfs_bytes_to_cluster(vol,
|
|
le64_to_cpu(a->data.non_resident.allocated_size));
|
|
}
|
|
|
|
/* Get the lowest vcn for the next extent. */
|
|
highest_vcn = le64_to_cpu(a->data.non_resident.highest_vcn);
|
|
next_vcn = highest_vcn + 1;
|
|
|
|
/* Only one extent or error, which we catch below. */
|
|
if (next_vcn <= 0) {
|
|
err = -ENOENT;
|
|
break;
|
|
}
|
|
|
|
/* Avoid endless loops due to corruption. */
|
|
if (next_vcn < le64_to_cpu(a->data.non_resident.lowest_vcn)) {
|
|
err = -EIO;
|
|
ntfs_error(sb, "Inode %llu has corrupt attribute list",
|
|
(unsigned long long)ni->mft_no);
|
|
goto err_out;
|
|
}
|
|
}
|
|
if (!a) {
|
|
ntfs_error(sb, "Couldn't find attribute for runlist mapping");
|
|
goto err_out;
|
|
}
|
|
if (not_mapped && highest_vcn && highest_vcn != last_vcn - 1) {
|
|
err = -EIO;
|
|
ntfs_error(sb,
|
|
"Failed to load full runlist: inode: %llu highest_vcn: 0x%llx last_vcn: 0x%llx",
|
|
(unsigned long long)ni->mft_no,
|
|
(long long)highest_vcn, (long long)last_vcn);
|
|
goto err_out;
|
|
}
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
if (err == -ENOENT) {
|
|
NInoSetFullyMapped(ni);
|
|
return 0;
|
|
}
|
|
|
|
return err;
|
|
|
|
err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_record_move_to - move attribute record to target inode
|
|
* @ctx: attribute search context describing the attribute record
|
|
* @ni: opened ntfs inode to which move attribute record
|
|
*/
|
|
int ntfs_attr_record_move_to(struct ntfs_attr_search_ctx *ctx, struct ntfs_inode *ni)
|
|
{
|
|
struct ntfs_attr_search_ctx *nctx;
|
|
struct attr_record *a;
|
|
int err;
|
|
struct mft_record *ni_mrec;
|
|
struct super_block *sb;
|
|
|
|
if (!ctx || !ctx->attr || !ctx->ntfs_ino || !ni) {
|
|
ntfs_debug("Invalid arguments passed.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
sb = ni->vol->sb;
|
|
ntfs_debug("Entering for ctx->attr->type 0x%x, ctx->ntfs_ino->mft_no 0x%llx, ni->mft_no 0x%llx.\n",
|
|
(unsigned int) le32_to_cpu(ctx->attr->type),
|
|
(long long) ctx->ntfs_ino->mft_no,
|
|
(long long) ni->mft_no);
|
|
|
|
if (ctx->ntfs_ino == ni)
|
|
return 0;
|
|
|
|
if (!ctx->al_entry) {
|
|
ntfs_debug("Inode should contain attribute list to use this function.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Find place in MFT record where attribute will be moved. */
|
|
a = ctx->attr;
|
|
nctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!nctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/*
|
|
* Use ntfs_attr_find instead of ntfs_attr_lookup to find place for
|
|
* attribute in @ni->mrec, not any extent inode in case if @ni is base
|
|
* file record.
|
|
*/
|
|
err = ntfs_attr_find(a->type, (__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length, CASE_SENSITIVE, NULL,
|
|
0, nctx);
|
|
if (!err) {
|
|
ntfs_debug("Attribute of such type, with same name already present in this MFT record.\n");
|
|
err = -EEXIST;
|
|
goto put_err_out;
|
|
}
|
|
if (err != -ENOENT) {
|
|
ntfs_debug("Attribute lookup failed.\n");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Make space and move attribute. */
|
|
ni_mrec = map_mft_record(ni);
|
|
if (IS_ERR(ni_mrec)) {
|
|
err = -EIO;
|
|
goto put_err_out;
|
|
}
|
|
|
|
err = ntfs_make_room_for_attr(ni_mrec, (u8 *) nctx->attr,
|
|
le32_to_cpu(a->length));
|
|
if (err) {
|
|
ntfs_debug("Couldn't make space for attribute.\n");
|
|
unmap_mft_record(ni);
|
|
goto put_err_out;
|
|
}
|
|
memcpy(nctx->attr, a, le32_to_cpu(a->length));
|
|
nctx->attr->instance = nctx->mrec->next_attr_instance;
|
|
nctx->mrec->next_attr_instance =
|
|
cpu_to_le16((le16_to_cpu(nctx->mrec->next_attr_instance) + 1) & 0xffff);
|
|
ntfs_attr_record_resize(ctx->mrec, a, 0);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
mark_mft_record_dirty(ni);
|
|
|
|
/* Update attribute list. */
|
|
ctx->al_entry->mft_reference =
|
|
MK_LE_MREF(ni->mft_no, le16_to_cpu(ni_mrec->sequence_number));
|
|
ctx->al_entry->instance = nctx->attr->instance;
|
|
unmap_mft_record(ni);
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(nctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_record_move_away - move away attribute record from it's mft record
|
|
* @ctx: attribute search context describing the attribute record
|
|
* @extra: minimum amount of free space in the new holder of record
|
|
*/
|
|
int ntfs_attr_record_move_away(struct ntfs_attr_search_ctx *ctx, int extra)
|
|
{
|
|
struct ntfs_inode *base_ni, *ni = NULL;
|
|
struct mft_record *m;
|
|
int i, err;
|
|
struct super_block *sb;
|
|
|
|
if (!ctx || !ctx->attr || !ctx->ntfs_ino || extra < 0)
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for attr 0x%x, inode %llu\n",
|
|
(unsigned int) le32_to_cpu(ctx->attr->type),
|
|
(unsigned long long)ctx->ntfs_ino->mft_no);
|
|
|
|
if (ctx->ntfs_ino->nr_extents == -1)
|
|
base_ni = ctx->base_ntfs_ino;
|
|
else
|
|
base_ni = ctx->ntfs_ino;
|
|
|
|
sb = ctx->ntfs_ino->vol->sb;
|
|
if (!NInoAttrList(base_ni)) {
|
|
ntfs_error(sb, "Inode %llu has no attrlist",
|
|
(unsigned long long)base_ni->mft_no);
|
|
return -EINVAL;
|
|
}
|
|
|
|
err = ntfs_inode_attach_all_extents(ctx->ntfs_ino);
|
|
if (err) {
|
|
ntfs_error(sb, "Couldn't attach extents, inode=%llu",
|
|
(unsigned long long)base_ni->mft_no);
|
|
return err;
|
|
}
|
|
|
|
mutex_lock(&base_ni->extent_lock);
|
|
/* Walk through all extents and try to move attribute to them. */
|
|
for (i = 0; i < base_ni->nr_extents; i++) {
|
|
ni = base_ni->ext.extent_ntfs_inos[i];
|
|
|
|
if (ctx->ntfs_ino->mft_no == ni->mft_no)
|
|
continue;
|
|
m = map_mft_record(ni);
|
|
if (IS_ERR(m)) {
|
|
ntfs_error(sb, "Can not map mft record for mft_no %lld",
|
|
(unsigned long long)ni->mft_no);
|
|
mutex_unlock(&base_ni->extent_lock);
|
|
return -EIO;
|
|
}
|
|
if (le32_to_cpu(m->bytes_allocated) -
|
|
le32_to_cpu(m->bytes_in_use) < le32_to_cpu(ctx->attr->length) + extra) {
|
|
unmap_mft_record(ni);
|
|
continue;
|
|
}
|
|
unmap_mft_record(ni);
|
|
|
|
/*
|
|
* ntfs_attr_record_move_to can fail if extent with other lowest
|
|
* s64 already present in inode we trying move record to. So,
|
|
* do not return error.
|
|
*/
|
|
if (!ntfs_attr_record_move_to(ctx, ni)) {
|
|
mutex_unlock(&base_ni->extent_lock);
|
|
return 0;
|
|
}
|
|
}
|
|
mutex_unlock(&base_ni->extent_lock);
|
|
|
|
/*
|
|
* Failed to move attribute to one of the current extents, so allocate
|
|
* new extent and move attribute to it.
|
|
*/
|
|
ni = NULL;
|
|
err = ntfs_mft_record_alloc(base_ni->vol, 0, &ni, base_ni, NULL);
|
|
if (err) {
|
|
ntfs_error(sb, "Couldn't allocate MFT record, err : %d", err);
|
|
return err;
|
|
}
|
|
unmap_mft_record(ni);
|
|
|
|
err = ntfs_attr_record_move_to(ctx, ni);
|
|
if (err)
|
|
ntfs_error(sb, "Couldn't move attribute to MFT record");
|
|
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* If we are in the first extent, then set/clean sparse bit,
|
|
* update allocated and compressed size.
|
|
*/
|
|
static int ntfs_attr_update_meta(struct attr_record *a, struct ntfs_inode *ni,
|
|
struct mft_record *m, struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
int sparse, err = 0;
|
|
struct ntfs_inode *base_ni;
|
|
struct super_block *sb = ni->vol->sb;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x\n",
|
|
(unsigned long long)ni->mft_no, ni->type);
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
if (a->data.non_resident.lowest_vcn)
|
|
goto out;
|
|
|
|
a->data.non_resident.allocated_size = cpu_to_le64(ni->allocated_size);
|
|
|
|
sparse = ntfs_rl_sparse(ni->runlist.rl);
|
|
if (sparse < 0) {
|
|
err = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
/* Attribute become sparse. */
|
|
if (sparse && !(a->flags & (ATTR_IS_SPARSE | ATTR_IS_COMPRESSED))) {
|
|
/*
|
|
* Move attribute to another mft record, if attribute is too
|
|
* small to add compressed_size field to it and we have no
|
|
* free space in the current mft record.
|
|
*/
|
|
if ((le32_to_cpu(a->length) -
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset) == 8) &&
|
|
!(le32_to_cpu(m->bytes_allocated) - le32_to_cpu(m->bytes_in_use))) {
|
|
|
|
if (!NInoAttrList(base_ni)) {
|
|
err = ntfs_inode_add_attrlist(base_ni);
|
|
if (err)
|
|
goto out;
|
|
err = -EAGAIN;
|
|
goto out;
|
|
}
|
|
err = ntfs_attr_record_move_away(ctx, 8);
|
|
if (err) {
|
|
ntfs_error(sb, "Failed to move attribute");
|
|
goto out;
|
|
}
|
|
|
|
err = ntfs_attrlist_update(base_ni);
|
|
if (err)
|
|
goto out;
|
|
err = -EAGAIN;
|
|
goto out;
|
|
}
|
|
if (!(le32_to_cpu(a->length) -
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset))) {
|
|
err = -EIO;
|
|
ntfs_error(sb, "Mapping pairs space is 0");
|
|
goto out;
|
|
}
|
|
|
|
NInoSetSparse(ni);
|
|
ni->flags |= FILE_ATTR_SPARSE_FILE;
|
|
a->flags |= ATTR_IS_SPARSE;
|
|
a->data.non_resident.compression_unit = 0;
|
|
|
|
memmove((u8 *)a + le16_to_cpu(a->name_offset) + 8,
|
|
(u8 *)a + le16_to_cpu(a->name_offset),
|
|
a->name_length * sizeof(__le16));
|
|
|
|
a->name_offset = cpu_to_le16(le16_to_cpu(a->name_offset) + 8);
|
|
|
|
a->data.non_resident.mapping_pairs_offset =
|
|
cpu_to_le16(le16_to_cpu(a->data.non_resident.mapping_pairs_offset) + 8);
|
|
}
|
|
|
|
/* Attribute no longer sparse. */
|
|
if (!sparse && (a->flags & ATTR_IS_SPARSE) &&
|
|
!(a->flags & ATTR_IS_COMPRESSED)) {
|
|
NInoClearSparse(ni);
|
|
ni->flags &= ~FILE_ATTR_SPARSE_FILE;
|
|
a->flags &= ~ATTR_IS_SPARSE;
|
|
a->data.non_resident.compression_unit = 0;
|
|
|
|
memmove((u8 *)a + le16_to_cpu(a->name_offset) - 8,
|
|
(u8 *)a + le16_to_cpu(a->name_offset),
|
|
a->name_length * sizeof(__le16));
|
|
|
|
if (le16_to_cpu(a->name_offset) >= 8)
|
|
a->name_offset = cpu_to_le16(le16_to_cpu(a->name_offset) - 8);
|
|
|
|
a->data.non_resident.mapping_pairs_offset =
|
|
cpu_to_le16(le16_to_cpu(a->data.non_resident.mapping_pairs_offset) - 8);
|
|
}
|
|
|
|
/* Update compressed size if required. */
|
|
if (NInoFullyMapped(ni) && (sparse || NInoCompressed(ni))) {
|
|
s64 new_compr_size;
|
|
|
|
new_compr_size = ntfs_rl_get_compressed_size(ni->vol, ni->runlist.rl);
|
|
if (new_compr_size < 0) {
|
|
err = new_compr_size;
|
|
goto out;
|
|
}
|
|
|
|
ni->itype.compressed.size = new_compr_size;
|
|
a->data.non_resident.compressed_size = cpu_to_le64(new_compr_size);
|
|
}
|
|
|
|
if (NInoSparse(ni) || NInoCompressed(ni))
|
|
VFS_I(base_ni)->i_blocks = ni->itype.compressed.size >> 9;
|
|
else
|
|
VFS_I(base_ni)->i_blocks = ni->allocated_size >> 9;
|
|
/*
|
|
* Set FILE_NAME dirty flag, to update sparse bit and
|
|
* allocated size in the index.
|
|
*/
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(ni);
|
|
out:
|
|
return err;
|
|
}
|
|
|
|
#define NTFS_VCN_DELETE_MARK -2
|
|
/*
|
|
* ntfs_attr_update_mapping_pairs - update mapping pairs for ntfs attribute
|
|
* @ni: non-resident ntfs inode for which we need update
|
|
* @from_vcn: update runlist starting this VCN
|
|
*
|
|
* Build mapping pairs from @na->rl and write them to the disk. Also, this
|
|
* function updates sparse bit, allocated and compressed size (allocates/frees
|
|
* space for this field if required).
|
|
*
|
|
* @na->allocated_size should be set to correct value for the new runlist before
|
|
* call to this function. Vice-versa @na->compressed_size will be calculated and
|
|
* set to correct value during this function.
|
|
*/
|
|
int ntfs_attr_update_mapping_pairs(struct ntfs_inode *ni, s64 from_vcn)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
struct ntfs_inode *base_ni;
|
|
struct mft_record *m;
|
|
struct attr_record *a;
|
|
s64 stop_vcn;
|
|
int err = 0, mp_size, cur_max_mp_size, exp_max_mp_size;
|
|
bool finished_build;
|
|
bool first_updated = false;
|
|
struct super_block *sb;
|
|
struct runlist_element *start_rl;
|
|
unsigned int de_cluster_count = 0;
|
|
|
|
retry:
|
|
if (!ni || !ni->runlist.rl)
|
|
return -EINVAL;
|
|
|
|
ntfs_debug("Entering for inode %llu, attr 0x%x\n",
|
|
(unsigned long long)ni->mft_no, ni->type);
|
|
|
|
sb = ni->vol->sb;
|
|
if (!NInoNonResident(ni)) {
|
|
ntfs_error(sb, "%s: resident attribute", __func__);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (ni->nr_extents == -1)
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Fill attribute records with new mapping pairs. */
|
|
stop_vcn = 0;
|
|
finished_build = false;
|
|
start_rl = ni->runlist.rl;
|
|
while (!(err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, from_vcn, NULL, 0, ctx))) {
|
|
unsigned int de_cnt = 0;
|
|
|
|
a = ctx->attr;
|
|
m = ctx->mrec;
|
|
if (!a->data.non_resident.lowest_vcn)
|
|
first_updated = true;
|
|
|
|
/*
|
|
* If runlist is updating not from the beginning, then set
|
|
* @stop_vcn properly, i.e. to the lowest vcn of record that
|
|
* contain @from_vcn. Also we do not need @from_vcn anymore,
|
|
* set it to 0 to make ntfs_attr_lookup enumerate attributes.
|
|
*/
|
|
if (from_vcn) {
|
|
s64 first_lcn;
|
|
|
|
stop_vcn = le64_to_cpu(a->data.non_resident.lowest_vcn);
|
|
from_vcn = 0;
|
|
/*
|
|
* Check whether the first run we need to update is
|
|
* the last run in runlist, if so, then deallocate
|
|
* all attrubute extents starting this one.
|
|
*/
|
|
first_lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, stop_vcn);
|
|
if (first_lcn == LCN_EINVAL) {
|
|
err = -EIO;
|
|
ntfs_error(sb, "Bad runlist");
|
|
goto put_err_out;
|
|
}
|
|
if (first_lcn == LCN_ENOENT ||
|
|
first_lcn == LCN_RL_NOT_MAPPED)
|
|
finished_build = true;
|
|
}
|
|
|
|
/*
|
|
* Check whether we finished mapping pairs build, if so mark
|
|
* extent as need to delete (by setting highest vcn to
|
|
* NTFS_VCN_DELETE_MARK (-2), we shall check it later and
|
|
* delete extent) and continue search.
|
|
*/
|
|
if (finished_build) {
|
|
ntfs_debug("Mark attr 0x%x for delete in inode 0x%llx.\n",
|
|
(unsigned int)le32_to_cpu(a->type), ctx->ntfs_ino->mft_no);
|
|
a->data.non_resident.highest_vcn = cpu_to_le64(NTFS_VCN_DELETE_MARK);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
continue;
|
|
}
|
|
|
|
err = ntfs_attr_update_meta(a, ni, m, ctx);
|
|
if (err < 0) {
|
|
if (err == -EAGAIN) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
goto retry;
|
|
}
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* Determine maximum possible length of mapping pairs,
|
|
* if we shall *not* expand space for mapping pairs.
|
|
*/
|
|
cur_max_mp_size = le32_to_cpu(a->length) -
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset);
|
|
/*
|
|
* Determine maximum possible length of mapping pairs in the
|
|
* current mft record, if we shall expand space for mapping
|
|
* pairs.
|
|
*/
|
|
exp_max_mp_size = le32_to_cpu(m->bytes_allocated) -
|
|
le32_to_cpu(m->bytes_in_use) + cur_max_mp_size;
|
|
|
|
/* Get the size for the rest of mapping pairs array. */
|
|
mp_size = ntfs_get_size_for_mapping_pairs(ni->vol, start_rl,
|
|
stop_vcn, -1, exp_max_mp_size);
|
|
if (mp_size <= 0) {
|
|
err = mp_size;
|
|
ntfs_error(sb, "%s: get MP size failed", __func__);
|
|
goto put_err_out;
|
|
}
|
|
/* Test mapping pairs for fitting in the current mft record. */
|
|
if (mp_size > exp_max_mp_size) {
|
|
/*
|
|
* Mapping pairs of $ATTRIBUTE_LIST attribute must fit
|
|
* in the base mft record. Try to move out other
|
|
* attributes and try again.
|
|
*/
|
|
if (ni->type == AT_ATTRIBUTE_LIST) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
if (ntfs_inode_free_space(base_ni, mp_size -
|
|
cur_max_mp_size)) {
|
|
ntfs_debug("Attribute list is too big. Defragment the volume\n");
|
|
return -ENOSPC;
|
|
}
|
|
if (ntfs_attrlist_update(base_ni))
|
|
return -EIO;
|
|
goto retry;
|
|
}
|
|
|
|
/* Add attribute list if it isn't present, and retry. */
|
|
if (!NInoAttrList(base_ni)) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
if (ntfs_inode_add_attrlist(base_ni)) {
|
|
ntfs_error(sb, "Can not add attrlist");
|
|
return -EIO;
|
|
}
|
|
goto retry;
|
|
}
|
|
|
|
/*
|
|
* Set mapping pairs size to maximum possible for this
|
|
* mft record. We shall write the rest of mapping pairs
|
|
* to another MFT records.
|
|
*/
|
|
mp_size = exp_max_mp_size;
|
|
}
|
|
|
|
/* Change space for mapping pairs if we need it. */
|
|
if (((mp_size + 7) & ~7) != cur_max_mp_size) {
|
|
if (ntfs_attr_record_resize(m, a,
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset) +
|
|
mp_size)) {
|
|
err = -EIO;
|
|
ntfs_error(sb, "Failed to resize attribute");
|
|
goto put_err_out;
|
|
}
|
|
}
|
|
|
|
/* Update lowest vcn. */
|
|
a->data.non_resident.lowest_vcn = cpu_to_le64(stop_vcn);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
if ((ctx->ntfs_ino->nr_extents == -1 || NInoAttrList(ctx->ntfs_ino)) &&
|
|
ctx->attr->type != AT_ATTRIBUTE_LIST) {
|
|
ctx->al_entry->lowest_vcn = cpu_to_le64(stop_vcn);
|
|
err = ntfs_attrlist_update(base_ni);
|
|
if (err)
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* Generate the new mapping pairs array directly into the
|
|
* correct destination, i.e. the attribute record itself.
|
|
*/
|
|
err = ntfs_mapping_pairs_build(ni->vol,
|
|
(u8 *)a + le16_to_cpu(a->data.non_resident.mapping_pairs_offset),
|
|
mp_size, start_rl, stop_vcn, -1, &stop_vcn, &start_rl, &de_cnt);
|
|
if (!err)
|
|
finished_build = true;
|
|
if (!finished_build && err != -ENOSPC) {
|
|
ntfs_error(sb, "Failed to build mapping pairs");
|
|
goto put_err_out;
|
|
}
|
|
a->data.non_resident.highest_vcn = cpu_to_le64(stop_vcn - 1);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
de_cluster_count += de_cnt;
|
|
}
|
|
|
|
/* Check whether error occurred. */
|
|
if (err && err != -ENOENT) {
|
|
ntfs_error(sb, "%s: Attribute lookup failed", __func__);
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* If the base extent was skipped in the above process,
|
|
* we still may have to update the sizes.
|
|
*/
|
|
if (!first_updated) {
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx);
|
|
if (!err) {
|
|
a = ctx->attr;
|
|
a->data.non_resident.allocated_size = cpu_to_le64(ni->allocated_size);
|
|
if (NInoCompressed(ni) || NInoSparse(ni))
|
|
a->data.non_resident.compressed_size =
|
|
cpu_to_le64(ni->itype.compressed.size);
|
|
/* Updating sizes taints the extent holding the attr */
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(ni);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
} else {
|
|
ntfs_error(sb, "Failed to update sizes in base extent\n");
|
|
goto put_err_out;
|
|
}
|
|
}
|
|
|
|
/* Deallocate not used attribute extents and return with success. */
|
|
if (finished_build) {
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
ntfs_debug("Deallocate marked extents.\n");
|
|
while (!(err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx))) {
|
|
if (le64_to_cpu(ctx->attr->data.non_resident.highest_vcn) !=
|
|
NTFS_VCN_DELETE_MARK)
|
|
continue;
|
|
/* Remove unused attribute record. */
|
|
err = ntfs_attr_record_rm(ctx);
|
|
if (err) {
|
|
ntfs_error(sb, "Could not remove unused attr");
|
|
goto put_err_out;
|
|
}
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
}
|
|
if (err && err != -ENOENT) {
|
|
ntfs_error(sb, "%s: Attr lookup failed", __func__);
|
|
goto put_err_out;
|
|
}
|
|
ntfs_debug("Deallocate done.\n");
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
goto out;
|
|
}
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
ctx = NULL;
|
|
|
|
/* Allocate new MFT records for the rest of mapping pairs. */
|
|
while (1) {
|
|
struct ntfs_inode *ext_ni = NULL;
|
|
unsigned int de_cnt = 0;
|
|
|
|
/* Allocate new mft record. */
|
|
err = ntfs_mft_record_alloc(ni->vol, 0, &ext_ni, base_ni, NULL);
|
|
if (err) {
|
|
ntfs_error(sb, "Failed to allocate extent record");
|
|
goto put_err_out;
|
|
}
|
|
unmap_mft_record(ext_ni);
|
|
|
|
m = map_mft_record(ext_ni);
|
|
if (IS_ERR(m)) {
|
|
ntfs_error(sb, "Could not map new MFT record");
|
|
if (ntfs_mft_record_free(ni->vol, ext_ni))
|
|
ntfs_error(sb, "Could not free MFT record");
|
|
ntfs_inode_close(ext_ni);
|
|
err = -ENOMEM;
|
|
ext_ni = NULL;
|
|
goto put_err_out;
|
|
}
|
|
/*
|
|
* If mapping size exceed available space, set them to
|
|
* possible maximum.
|
|
*/
|
|
cur_max_mp_size = le32_to_cpu(m->bytes_allocated) -
|
|
le32_to_cpu(m->bytes_in_use) -
|
|
(sizeof(struct attr_record) +
|
|
((NInoCompressed(ni) || NInoSparse(ni)) ?
|
|
sizeof(a->data.non_resident.compressed_size) : 0)) -
|
|
((sizeof(__le16) * ni->name_len + 7) & ~7);
|
|
|
|
/* Calculate size of rest mapping pairs. */
|
|
mp_size = ntfs_get_size_for_mapping_pairs(ni->vol,
|
|
start_rl, stop_vcn, -1, cur_max_mp_size);
|
|
if (mp_size <= 0) {
|
|
unmap_mft_record(ext_ni);
|
|
ntfs_inode_close(ext_ni);
|
|
err = mp_size;
|
|
ntfs_error(sb, "%s: get mp size failed", __func__);
|
|
goto put_err_out;
|
|
}
|
|
|
|
if (mp_size > cur_max_mp_size)
|
|
mp_size = cur_max_mp_size;
|
|
/* Add attribute extent to new record. */
|
|
err = ntfs_non_resident_attr_record_add(ext_ni, ni->type,
|
|
ni->name, ni->name_len, stop_vcn, mp_size, 0);
|
|
if (err < 0) {
|
|
ntfs_error(sb, "Could not add attribute extent");
|
|
unmap_mft_record(ext_ni);
|
|
if (ntfs_mft_record_free(ni->vol, ext_ni))
|
|
ntfs_error(sb, "Could not free MFT record");
|
|
ntfs_inode_close(ext_ni);
|
|
goto put_err_out;
|
|
}
|
|
a = (struct attr_record *)((u8 *)m + err);
|
|
|
|
err = ntfs_mapping_pairs_build(ni->vol, (u8 *)a +
|
|
le16_to_cpu(a->data.non_resident.mapping_pairs_offset),
|
|
mp_size, start_rl, stop_vcn, -1, &stop_vcn, &start_rl,
|
|
&de_cnt);
|
|
if (err < 0 && err != -ENOSPC) {
|
|
ntfs_error(sb, "Failed to build MP");
|
|
unmap_mft_record(ext_ni);
|
|
if (ntfs_mft_record_free(ni->vol, ext_ni))
|
|
ntfs_error(sb, "Couldn't free MFT record");
|
|
goto put_err_out;
|
|
}
|
|
a->data.non_resident.highest_vcn = cpu_to_le64(stop_vcn - 1);
|
|
mark_mft_record_dirty(ext_ni);
|
|
unmap_mft_record(ext_ni);
|
|
|
|
de_cluster_count += de_cnt;
|
|
/* All mapping pairs has been written. */
|
|
if (!err)
|
|
break;
|
|
}
|
|
out:
|
|
if (from_vcn == 0)
|
|
ni->i_dealloc_clusters = de_cluster_count;
|
|
return 0;
|
|
|
|
put_err_out:
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_make_resident - convert a non-resident to a resident attribute
|
|
* @ni: open ntfs attribute to make resident
|
|
* @ctx: ntfs search context describing the attribute
|
|
*
|
|
* Convert a non-resident ntfs attribute to a resident one.
|
|
*/
|
|
static int ntfs_attr_make_resident(struct ntfs_inode *ni, struct ntfs_attr_search_ctx *ctx)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct super_block *sb = vol->sb;
|
|
struct attr_record *a = ctx->attr;
|
|
int name_ofs, val_ofs, err;
|
|
s64 arec_size;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x.\n",
|
|
(unsigned long long)ni->mft_no, ni->type);
|
|
|
|
/* Should be called for the first extent of the attribute. */
|
|
if (le64_to_cpu(a->data.non_resident.lowest_vcn)) {
|
|
ntfs_debug("Eeek! Should be called for the first extent of the attribute. Aborting...\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Some preliminary sanity checking. */
|
|
if (!NInoNonResident(ni)) {
|
|
ntfs_debug("Eeek! Trying to make resident attribute resident. Aborting...\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Make sure this is not $MFT/$BITMAP or Windows will not boot! */
|
|
if (ni->type == AT_BITMAP && ni->mft_no == FILE_MFT)
|
|
return -EPERM;
|
|
|
|
/* Check that the attribute is allowed to be resident. */
|
|
err = ntfs_attr_can_be_resident(vol, ni->type);
|
|
if (err)
|
|
return err;
|
|
|
|
if (NInoCompressed(ni) || NInoEncrypted(ni)) {
|
|
ntfs_debug("Making compressed or encrypted files resident is not implemented yet.\n");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
/* Work out offsets into and size of the resident attribute. */
|
|
name_ofs = 24; /* = sizeof(resident_struct attr_record); */
|
|
val_ofs = (name_ofs + a->name_length * sizeof(__le16) + 7) & ~7;
|
|
arec_size = (val_ofs + ni->data_size + 7) & ~7;
|
|
|
|
/* Sanity check the size before we start modifying the attribute. */
|
|
if (le32_to_cpu(ctx->mrec->bytes_in_use) - le32_to_cpu(a->length) +
|
|
arec_size > le32_to_cpu(ctx->mrec->bytes_allocated)) {
|
|
ntfs_debug("Not enough space to make attribute resident\n");
|
|
return -ENOSPC;
|
|
}
|
|
|
|
/* Read and cache the whole runlist if not already done. */
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err)
|
|
return err;
|
|
|
|
/* Move the attribute name if it exists and update the offset. */
|
|
if (a->name_length) {
|
|
memmove((u8 *)a + name_ofs, (u8 *)a + le16_to_cpu(a->name_offset),
|
|
a->name_length * sizeof(__le16));
|
|
}
|
|
a->name_offset = cpu_to_le16(name_ofs);
|
|
|
|
/* Resize the resident part of the attribute record. */
|
|
if (ntfs_attr_record_resize(ctx->mrec, a, arec_size) < 0) {
|
|
/*
|
|
* Bug, because ntfs_attr_record_resize should not fail (we
|
|
* already checked that attribute fits MFT record).
|
|
*/
|
|
ntfs_error(ctx->ntfs_ino->vol->sb, "BUG! Failed to resize attribute record. ");
|
|
return -EIO;
|
|
}
|
|
|
|
/* Convert the attribute record to describe a resident attribute. */
|
|
a->non_resident = 0;
|
|
a->flags = 0;
|
|
a->data.resident.value_length = cpu_to_le32(ni->data_size);
|
|
a->data.resident.value_offset = cpu_to_le16(val_ofs);
|
|
/*
|
|
* File names cannot be non-resident so we would never see this here
|
|
* but at least it serves as a reminder that there may be attributes
|
|
* for which we do need to set this flag. (AIA)
|
|
*/
|
|
if (a->type == AT_FILE_NAME)
|
|
a->data.resident.flags = RESIDENT_ATTR_IS_INDEXED;
|
|
else
|
|
a->data.resident.flags = 0;
|
|
a->data.resident.reserved = 0;
|
|
|
|
/*
|
|
* Deallocate clusters from the runlist.
|
|
*
|
|
* NOTE: We can use ntfs_cluster_free() because we have already mapped
|
|
* the whole run list and thus it doesn't matter that the attribute
|
|
* record is in a transiently corrupted state at this moment in time.
|
|
*/
|
|
err = ntfs_cluster_free(ni, 0, -1, ctx);
|
|
if (err) {
|
|
ntfs_error(sb, "Eeek! Failed to release allocated clusters");
|
|
ntfs_debug("Ignoring error and leaving behind wasted clusters.\n");
|
|
}
|
|
|
|
/* Throw away the now unused runlist. */
|
|
kvfree(ni->runlist.rl);
|
|
ni->runlist.rl = NULL;
|
|
ni->runlist.count = 0;
|
|
/* Update in-memory struct ntfs_attr. */
|
|
NInoClearNonResident(ni);
|
|
NInoClearCompressed(ni);
|
|
ni->flags &= ~FILE_ATTR_COMPRESSED;
|
|
NInoClearSparse(ni);
|
|
ni->flags &= ~FILE_ATTR_SPARSE_FILE;
|
|
NInoClearEncrypted(ni);
|
|
ni->flags &= ~FILE_ATTR_ENCRYPTED;
|
|
ni->initialized_size = ni->data_size;
|
|
ni->allocated_size = ni->itype.compressed.size = (ni->data_size + 7) & ~7;
|
|
ni->itype.compressed.block_size = 0;
|
|
ni->itype.compressed.block_size_bits = ni->itype.compressed.block_clusters = 0;
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* ntfs_non_resident_attr_shrink - shrink a non-resident, open ntfs attribute
|
|
* @ni: non-resident ntfs attribute to shrink
|
|
* @newsize: new size (in bytes) to which to shrink the attribute
|
|
*
|
|
* Reduce the size of a non-resident, open ntfs attribute @na to @newsize bytes.
|
|
*/
|
|
static int ntfs_non_resident_attr_shrink(struct ntfs_inode *ni, const s64 newsize)
|
|
{
|
|
struct ntfs_volume *vol;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
s64 first_free_vcn;
|
|
s64 nr_freed_clusters;
|
|
int err;
|
|
struct ntfs_inode *base_ni;
|
|
|
|
ntfs_debug("Inode 0x%llx attr 0x%x new size %lld\n",
|
|
(unsigned long long)ni->mft_no, ni->type, (long long)newsize);
|
|
|
|
vol = ni->vol;
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
/*
|
|
* Check the attribute type and the corresponding minimum size
|
|
* against @newsize and fail if @newsize is too small.
|
|
*/
|
|
err = ntfs_attr_size_bounds_check(vol, ni->type, newsize);
|
|
if (err) {
|
|
if (err == -ERANGE)
|
|
ntfs_debug("Eeek! Size bounds check failed. Aborting...\n");
|
|
else if (err == -ENOENT)
|
|
err = -EIO;
|
|
return err;
|
|
}
|
|
|
|
/* The first cluster outside the new allocation. */
|
|
if (NInoCompressed(ni))
|
|
/*
|
|
* For compressed files we must keep full compressions blocks,
|
|
* but currently we do not decompress/recompress the last
|
|
* block to truncate the data, so we may leave more allocated
|
|
* clusters than really needed.
|
|
*/
|
|
first_free_vcn = ntfs_bytes_to_cluster(vol,
|
|
((newsize - 1) | (ni->itype.compressed.block_size - 1)) + 1);
|
|
else
|
|
first_free_vcn =
|
|
ntfs_bytes_to_cluster(vol, newsize + vol->cluster_size - 1);
|
|
|
|
if (first_free_vcn < 0)
|
|
return -EINVAL;
|
|
/*
|
|
* Compare the new allocation with the old one and only deallocate
|
|
* clusters if there is a change.
|
|
*/
|
|
if (ntfs_bytes_to_cluster(vol, ni->allocated_size) != first_free_vcn) {
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err) {
|
|
ntfs_debug("Eeek! ntfs_attr_map_whole_runlist failed.\n");
|
|
return err;
|
|
}
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(vol->sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Deallocate all clusters starting with the first free one. */
|
|
nr_freed_clusters = ntfs_cluster_free(ni, first_free_vcn, -1, ctx);
|
|
if (nr_freed_clusters < 0) {
|
|
ntfs_debug("Eeek! Freeing of clusters failed. Aborting...\n");
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return (int)nr_freed_clusters;
|
|
}
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
|
|
/* Truncate the runlist itself. */
|
|
if (ntfs_rl_truncate_nolock(vol, &ni->runlist, first_free_vcn)) {
|
|
/*
|
|
* Failed to truncate the runlist, so just throw it
|
|
* away, it will be mapped afresh on next use.
|
|
*/
|
|
kvfree(ni->runlist.rl);
|
|
ni->runlist.rl = NULL;
|
|
ntfs_error(vol->sb, "Eeek! Run list truncation failed.\n");
|
|
return -EIO;
|
|
}
|
|
|
|
/* Prepare to mapping pairs update. */
|
|
ni->allocated_size = ntfs_cluster_to_bytes(vol, first_free_vcn);
|
|
|
|
if (NInoSparse(ni) || NInoCompressed(ni)) {
|
|
if (nr_freed_clusters) {
|
|
ni->itype.compressed.size -=
|
|
ntfs_cluster_to_bytes(vol, nr_freed_clusters);
|
|
VFS_I(base_ni)->i_blocks = ni->itype.compressed.size >> 9;
|
|
}
|
|
} else
|
|
VFS_I(base_ni)->i_blocks = ni->allocated_size >> 9;
|
|
|
|
/* Write mapping pairs for new runlist. */
|
|
err = ntfs_attr_update_mapping_pairs(ni, 0 /*first_free_vcn*/);
|
|
if (err) {
|
|
ntfs_debug("Eeek! Mapping pairs update failed. Leaving inconstant metadata. Run chkdsk.\n");
|
|
return err;
|
|
}
|
|
}
|
|
|
|
/* Get the first attribute record. */
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(vol->sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx);
|
|
if (err) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
ntfs_debug("Eeek! Lookup of first attribute extent failed. Leaving inconstant metadata.\n");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Update data and initialized size. */
|
|
ni->data_size = newsize;
|
|
ctx->attr->data.non_resident.data_size = cpu_to_le64(newsize);
|
|
if (newsize < ni->initialized_size) {
|
|
ni->initialized_size = newsize;
|
|
ctx->attr->data.non_resident.initialized_size = cpu_to_le64(newsize);
|
|
}
|
|
/* Update data size in the index. */
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(ni);
|
|
|
|
/* If the attribute now has zero size, make it resident. */
|
|
if (!newsize && !NInoEncrypted(ni) && !NInoCompressed(ni)) {
|
|
err = ntfs_attr_make_resident(ni, ctx);
|
|
if (err) {
|
|
/* If couldn't make resident, just continue. */
|
|
if (err != -EPERM)
|
|
ntfs_error(ni->vol->sb,
|
|
"Failed to make attribute resident. Leaving as is...\n");
|
|
}
|
|
}
|
|
|
|
/* Set the inode dirty so it is written out later. */
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
/* Done! */
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return 0;
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_non_resident_attr_expand - expand a non-resident, open ntfs attribute
|
|
* @ni: non-resident ntfs attribute to expand
|
|
* @prealloc_size: preallocation size (in bytes) to which to expand the attribute
|
|
* @newsize: new size (in bytes) to which to expand the attribute
|
|
* @holes: how to create a hole if expanding
|
|
* @need_lock: whether mrec lock is needed or not
|
|
*
|
|
* Expand the size of a non-resident, open ntfs attribute @na to @newsize bytes,
|
|
* by allocating new clusters.
|
|
*/
|
|
static int ntfs_non_resident_attr_expand(struct ntfs_inode *ni, const s64 newsize,
|
|
const s64 prealloc_size, unsigned int holes, bool need_lock)
|
|
{
|
|
s64 lcn_seek_from;
|
|
s64 first_free_vcn;
|
|
struct ntfs_volume *vol;
|
|
struct ntfs_attr_search_ctx *ctx = NULL;
|
|
struct runlist_element *rl, *rln;
|
|
s64 org_alloc_size, org_compressed_size;
|
|
int err, err2;
|
|
struct ntfs_inode *base_ni;
|
|
struct super_block *sb = ni->vol->sb;
|
|
size_t new_rl_count;
|
|
|
|
ntfs_debug("Inode 0x%llx, attr 0x%x, new size %lld old size %lld\n",
|
|
(unsigned long long)ni->mft_no, ni->type,
|
|
(long long)newsize, (long long)ni->data_size);
|
|
|
|
vol = ni->vol;
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
/*
|
|
* Check the attribute type and the corresponding maximum size
|
|
* against @newsize and fail if @newsize is too big.
|
|
*/
|
|
err = ntfs_attr_size_bounds_check(vol, ni->type, newsize);
|
|
if (err < 0) {
|
|
ntfs_error(sb, "%s: bounds check failed", __func__);
|
|
return err;
|
|
}
|
|
|
|
/* Save for future use. */
|
|
org_alloc_size = ni->allocated_size;
|
|
org_compressed_size = ni->itype.compressed.size;
|
|
|
|
/* The first cluster outside the new allocation. */
|
|
if (prealloc_size)
|
|
first_free_vcn =
|
|
ntfs_bytes_to_cluster(vol, prealloc_size + vol->cluster_size - 1);
|
|
else
|
|
first_free_vcn =
|
|
ntfs_bytes_to_cluster(vol, newsize + vol->cluster_size - 1);
|
|
if (first_free_vcn < 0)
|
|
return -EFBIG;
|
|
|
|
/*
|
|
* Compare the new allocation with the old one and only allocate
|
|
* clusters if there is a change.
|
|
*/
|
|
if (ntfs_bytes_to_cluster(vol, ni->allocated_size) < first_free_vcn) {
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err) {
|
|
ntfs_error(sb, "ntfs_attr_map_whole_runlist failed");
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* If we extend $DATA attribute on NTFS 3+ volume, we can add
|
|
* sparse runs instead of real allocation of clusters.
|
|
*/
|
|
if ((ni->type == AT_DATA && (vol->major_ver >= 3 || !NInoSparseDisabled(ni))) &&
|
|
(holes != HOLES_NO)) {
|
|
if (NInoCompressed(ni)) {
|
|
int last = 0, i = 0;
|
|
s64 alloc_size;
|
|
u64 more_entries = round_up(first_free_vcn -
|
|
ntfs_bytes_to_cluster(vol, ni->allocated_size),
|
|
ni->itype.compressed.block_clusters);
|
|
|
|
do_div(more_entries, ni->itype.compressed.block_clusters);
|
|
|
|
while (ni->runlist.rl[last].length)
|
|
last++;
|
|
|
|
rl = ntfs_rl_realloc(ni->runlist.rl, last + 1,
|
|
last + more_entries + 1);
|
|
if (IS_ERR(rl)) {
|
|
err = -ENOMEM;
|
|
goto put_err_out;
|
|
}
|
|
|
|
alloc_size = ni->allocated_size;
|
|
while (i++ < more_entries) {
|
|
rl[last].vcn = ntfs_bytes_to_cluster(vol,
|
|
round_up(alloc_size, vol->cluster_size));
|
|
rl[last].length = ni->itype.compressed.block_clusters -
|
|
(rl[last].vcn &
|
|
(ni->itype.compressed.block_clusters - 1));
|
|
rl[last].lcn = LCN_HOLE;
|
|
last++;
|
|
alloc_size += ni->itype.compressed.block_size;
|
|
}
|
|
|
|
rl[last].vcn = first_free_vcn;
|
|
rl[last].lcn = LCN_ENOENT;
|
|
rl[last].length = 0;
|
|
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count += more_entries;
|
|
} else {
|
|
rl = kmalloc(sizeof(struct runlist_element) * 2, GFP_NOFS);
|
|
if (!rl) {
|
|
err = -ENOMEM;
|
|
goto put_err_out;
|
|
}
|
|
|
|
rl[0].vcn = ntfs_bytes_to_cluster(vol, ni->allocated_size);
|
|
rl[0].lcn = LCN_HOLE;
|
|
rl[0].length = first_free_vcn -
|
|
ntfs_bytes_to_cluster(vol, ni->allocated_size);
|
|
rl[1].vcn = first_free_vcn;
|
|
rl[1].lcn = LCN_ENOENT;
|
|
rl[1].length = 0;
|
|
}
|
|
} else {
|
|
/*
|
|
* Determine first after last LCN of attribute.
|
|
* We will start seek clusters from this LCN to avoid
|
|
* fragmentation. If there are no valid LCNs in the
|
|
* attribute let the cluster allocator choose the
|
|
* starting LCN.
|
|
*/
|
|
lcn_seek_from = -1;
|
|
if (ni->runlist.rl->length) {
|
|
/* Seek to the last run list element. */
|
|
for (rl = ni->runlist.rl; (rl + 1)->length; rl++)
|
|
;
|
|
/*
|
|
* If the last LCN is a hole or similar seek
|
|
* back to last valid LCN.
|
|
*/
|
|
while (rl->lcn < 0 && rl != ni->runlist.rl)
|
|
rl--;
|
|
/*
|
|
* Only set lcn_seek_from it the LCN is valid.
|
|
*/
|
|
if (rl->lcn >= 0)
|
|
lcn_seek_from = rl->lcn + rl->length;
|
|
}
|
|
|
|
rl = ntfs_cluster_alloc(vol,
|
|
ntfs_bytes_to_cluster(vol, ni->allocated_size),
|
|
first_free_vcn -
|
|
ntfs_bytes_to_cluster(vol, ni->allocated_size),
|
|
lcn_seek_from, DATA_ZONE, false, false, false);
|
|
if (IS_ERR(rl)) {
|
|
ntfs_debug("Cluster allocation failed (%lld)",
|
|
(long long)first_free_vcn -
|
|
ntfs_bytes_to_cluster(vol, ni->allocated_size));
|
|
return PTR_ERR(rl);
|
|
}
|
|
}
|
|
|
|
if (!NInoCompressed(ni)) {
|
|
/* Append new clusters to attribute runlist. */
|
|
rln = ntfs_runlists_merge(&ni->runlist, rl, 0, &new_rl_count);
|
|
if (IS_ERR(rln)) {
|
|
/* Failed, free just allocated clusters. */
|
|
ntfs_error(sb, "Run list merge failed");
|
|
ntfs_cluster_free_from_rl(vol, rl);
|
|
kvfree(rl);
|
|
return -EIO;
|
|
}
|
|
ni->runlist.rl = rln;
|
|
ni->runlist.count = new_rl_count;
|
|
}
|
|
|
|
/* Prepare to mapping pairs update. */
|
|
ni->allocated_size = ntfs_cluster_to_bytes(vol, first_free_vcn);
|
|
err = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
if (err) {
|
|
ntfs_debug("Mapping pairs update failed");
|
|
goto rollback;
|
|
}
|
|
}
|
|
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
err = -ENOMEM;
|
|
if (ni->allocated_size == org_alloc_size)
|
|
return err;
|
|
goto rollback;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx);
|
|
if (err) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
if (ni->allocated_size != org_alloc_size)
|
|
goto rollback;
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Update data size. */
|
|
ni->data_size = newsize;
|
|
ctx->attr->data.non_resident.data_size = cpu_to_le64(newsize);
|
|
/* Update data size in the index. */
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(ni);
|
|
/* Set the inode dirty so it is written out later. */
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
/* Done! */
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return 0;
|
|
rollback:
|
|
/* Free allocated clusters. */
|
|
err2 = ntfs_cluster_free(ni, ntfs_bytes_to_cluster(vol, org_alloc_size),
|
|
-1, ctx);
|
|
if (err2)
|
|
ntfs_debug("Leaking clusters");
|
|
|
|
/* Now, truncate the runlist itself. */
|
|
if (need_lock)
|
|
down_write(&ni->runlist.lock);
|
|
err2 = ntfs_rl_truncate_nolock(vol, &ni->runlist,
|
|
ntfs_bytes_to_cluster(vol, org_alloc_size));
|
|
if (need_lock)
|
|
up_write(&ni->runlist.lock);
|
|
if (err2) {
|
|
/*
|
|
* Failed to truncate the runlist, so just throw it away, it
|
|
* will be mapped afresh on next use.
|
|
*/
|
|
kvfree(ni->runlist.rl);
|
|
ni->runlist.rl = NULL;
|
|
ntfs_error(sb, "Couldn't truncate runlist. Rollback failed");
|
|
} else {
|
|
/* Prepare to mapping pairs update. */
|
|
ni->allocated_size = org_alloc_size;
|
|
/* Restore mapping pairs. */
|
|
if (need_lock)
|
|
down_read(&ni->runlist.lock);
|
|
if (ntfs_attr_update_mapping_pairs(ni, 0))
|
|
ntfs_error(sb, "Failed to restore old mapping pairs");
|
|
if (need_lock)
|
|
up_read(&ni->runlist.lock);
|
|
|
|
if (NInoSparse(ni) || NInoCompressed(ni)) {
|
|
ni->itype.compressed.size = org_compressed_size;
|
|
VFS_I(base_ni)->i_blocks = ni->itype.compressed.size >> 9;
|
|
} else
|
|
VFS_I(base_ni)->i_blocks = ni->allocated_size >> 9;
|
|
}
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
put_err_out:
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_resident_attr_resize - resize a resident, open ntfs attribute
|
|
* @attr_ni: resident ntfs inode to resize
|
|
* @newsize: new size (in bytes) to which to resize the attribute
|
|
* @prealloc_size: preallocation size (in bytes) to which to resize the attribute
|
|
* @holes: flags indicating how to handle holes
|
|
*
|
|
* Change the size of a resident, open ntfs attribute @na to @newsize bytes.
|
|
*/
|
|
static int ntfs_resident_attr_resize(struct ntfs_inode *attr_ni, const s64 newsize,
|
|
const s64 prealloc_size, unsigned int holes)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
struct ntfs_volume *vol = attr_ni->vol;
|
|
struct super_block *sb = vol->sb;
|
|
int err = -EIO;
|
|
struct ntfs_inode *base_ni, *ext_ni = NULL;
|
|
|
|
attr_resize_again:
|
|
ntfs_debug("Inode 0x%llx attr 0x%x new size %lld\n",
|
|
(unsigned long long)attr_ni->mft_no, attr_ni->type,
|
|
(long long)newsize);
|
|
|
|
if (NInoAttr(attr_ni))
|
|
base_ni = attr_ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = attr_ni;
|
|
|
|
/* Get the attribute record that needs modification. */
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(attr_ni->type, attr_ni->name, attr_ni->name_len,
|
|
0, 0, NULL, 0, ctx);
|
|
if (err) {
|
|
ntfs_error(sb, "ntfs_attr_lookup failed");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* Check the attribute type and the corresponding minimum and maximum
|
|
* sizes against @newsize and fail if @newsize is out of bounds.
|
|
*/
|
|
err = ntfs_attr_size_bounds_check(vol, attr_ni->type, newsize);
|
|
if (err) {
|
|
if (err == -ENOENT)
|
|
err = -EIO;
|
|
ntfs_debug("%s: bounds check failed", __func__);
|
|
goto put_err_out;
|
|
}
|
|
/*
|
|
* If @newsize is bigger than the mft record we need to make the
|
|
* attribute non-resident if the attribute type supports it. If it is
|
|
* smaller we can go ahead and attempt the resize.
|
|
*/
|
|
if (newsize < vol->mft_record_size) {
|
|
/* Perform the resize of the attribute record. */
|
|
err = ntfs_resident_attr_value_resize(ctx->mrec, ctx->attr,
|
|
newsize);
|
|
if (!err) {
|
|
/* Update attribute size everywhere. */
|
|
attr_ni->data_size = attr_ni->initialized_size = newsize;
|
|
attr_ni->allocated_size = (newsize + 7) & ~7;
|
|
if (NInoCompressed(attr_ni) || NInoSparse(attr_ni))
|
|
attr_ni->itype.compressed.size = attr_ni->allocated_size;
|
|
if (attr_ni->type == AT_DATA && attr_ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(attr_ni);
|
|
goto resize_done;
|
|
}
|
|
|
|
/* Prefer AT_INDEX_ALLOCATION instead of AT_ATTRIBUTE_LIST */
|
|
if (err == -ENOSPC && ctx->attr->type == AT_INDEX_ROOT)
|
|
goto put_err_out;
|
|
|
|
}
|
|
/* There is not enough space in the mft record to perform the resize. */
|
|
|
|
/* Make the attribute non-resident if possible. */
|
|
err = ntfs_attr_make_non_resident(attr_ni,
|
|
le32_to_cpu(ctx->attr->data.resident.value_length));
|
|
if (!err) {
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
/* Resize non-resident attribute */
|
|
return ntfs_non_resident_attr_expand(attr_ni, newsize, prealloc_size, holes, true);
|
|
} else if (err != -ENOSPC && err != -EPERM) {
|
|
ntfs_error(sb, "Failed to make attribute non-resident");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Try to make other attributes non-resident and retry each time. */
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
while (!(err = ntfs_attr_lookup(AT_UNUSED, NULL, 0, 0, 0, NULL, 0, ctx))) {
|
|
struct inode *tvi;
|
|
struct attr_record *a;
|
|
|
|
a = ctx->attr;
|
|
if (a->non_resident || a->type == AT_ATTRIBUTE_LIST)
|
|
continue;
|
|
|
|
if (ntfs_attr_can_be_non_resident(vol, a->type))
|
|
continue;
|
|
|
|
/*
|
|
* Check out whether convert is reasonable. Assume that mapping
|
|
* pairs will take 8 bytes.
|
|
*/
|
|
if (le32_to_cpu(a->length) <= (sizeof(struct attr_record) - sizeof(s64)) +
|
|
((a->name_length * sizeof(__le16) + 7) & ~7) + 8)
|
|
continue;
|
|
|
|
if (a->type == AT_DATA)
|
|
tvi = ntfs_iget(sb, base_ni->mft_no);
|
|
else
|
|
tvi = ntfs_attr_iget(VFS_I(base_ni), a->type,
|
|
(__le16 *)((u8 *)a + le16_to_cpu(a->name_offset)),
|
|
a->name_length);
|
|
if (IS_ERR(tvi)) {
|
|
ntfs_error(sb, "Couldn't open attribute");
|
|
continue;
|
|
}
|
|
|
|
if (ntfs_attr_make_non_resident(NTFS_I(tvi),
|
|
le32_to_cpu(ctx->attr->data.resident.value_length))) {
|
|
iput(tvi);
|
|
continue;
|
|
}
|
|
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
iput(tvi);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
goto attr_resize_again;
|
|
}
|
|
|
|
/* Check whether error occurred. */
|
|
if (err != -ENOENT) {
|
|
ntfs_error(sb, "%s: Attribute lookup failed 1", __func__);
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* The standard information and attribute list attributes can't be
|
|
* moved out from the base MFT record, so try to move out others.
|
|
*/
|
|
if (attr_ni->type == AT_STANDARD_INFORMATION ||
|
|
attr_ni->type == AT_ATTRIBUTE_LIST) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
|
|
if (!NInoAttrList(base_ni)) {
|
|
err = ntfs_inode_add_attrlist(base_ni);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
err = ntfs_inode_free_space(base_ni, sizeof(struct attr_record));
|
|
if (err) {
|
|
err = -ENOSPC;
|
|
ntfs_error(sb,
|
|
"Couldn't free space in the MFT record to make attribute list non resident");
|
|
return err;
|
|
}
|
|
err = ntfs_attrlist_update(base_ni);
|
|
if (err)
|
|
return err;
|
|
goto attr_resize_again;
|
|
}
|
|
|
|
/*
|
|
* Move the attribute to a new mft record, creating an attribute list
|
|
* attribute or modifying it if it is already present.
|
|
*/
|
|
|
|
/* Point search context back to attribute which we need resize. */
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
err = ntfs_attr_lookup(attr_ni->type, attr_ni->name, attr_ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx);
|
|
if (err) {
|
|
ntfs_error(sb, "%s: Attribute lookup failed 2", __func__);
|
|
goto put_err_out;
|
|
}
|
|
|
|
/*
|
|
* Check whether attribute is already single in this MFT record.
|
|
* 8 added for the attribute terminator.
|
|
*/
|
|
if (le32_to_cpu(ctx->mrec->bytes_in_use) ==
|
|
le16_to_cpu(ctx->mrec->attrs_offset) + le32_to_cpu(ctx->attr->length) + 8) {
|
|
err = -ENOSPC;
|
|
ntfs_debug("MFT record is filled with one attribute\n");
|
|
goto put_err_out;
|
|
}
|
|
|
|
/* Add attribute list if not present. */
|
|
if (!NInoAttrList(base_ni)) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
err = ntfs_inode_add_attrlist(base_ni);
|
|
if (err)
|
|
return err;
|
|
goto attr_resize_again;
|
|
}
|
|
|
|
/* Allocate new mft record. */
|
|
err = ntfs_mft_record_alloc(base_ni->vol, 0, &ext_ni, base_ni, NULL);
|
|
if (err) {
|
|
ntfs_error(sb, "Couldn't allocate MFT record");
|
|
goto put_err_out;
|
|
}
|
|
unmap_mft_record(ext_ni);
|
|
|
|
/* Move attribute to it. */
|
|
err = ntfs_attr_record_move_to(ctx, ext_ni);
|
|
if (err) {
|
|
ntfs_error(sb, "Couldn't move attribute to new MFT record");
|
|
err = -ENOMEM;
|
|
goto put_err_out;
|
|
}
|
|
|
|
err = ntfs_attrlist_update(base_ni);
|
|
if (err < 0)
|
|
goto put_err_out;
|
|
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
/* Try to perform resize once again. */
|
|
goto attr_resize_again;
|
|
|
|
resize_done:
|
|
/*
|
|
* Set the inode (and its base inode if it exists) dirty so it is
|
|
* written out later.
|
|
*/
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return 0;
|
|
|
|
put_err_out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
int __ntfs_attr_truncate_vfs(struct ntfs_inode *ni, const s64 newsize,
|
|
const s64 i_size)
|
|
{
|
|
int err = 0;
|
|
|
|
if (newsize < 0 ||
|
|
(ni->mft_no == FILE_MFT && ni->type == AT_DATA)) {
|
|
ntfs_debug("Invalid arguments passed.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x, size %lld\n",
|
|
(unsigned long long)ni->mft_no, ni->type, newsize);
|
|
|
|
if (NInoNonResident(ni)) {
|
|
if (newsize > i_size) {
|
|
down_write(&ni->runlist.lock);
|
|
err = ntfs_non_resident_attr_expand(ni, newsize, 0,
|
|
NVolDisableSparse(ni->vol) ?
|
|
HOLES_NO : HOLES_OK,
|
|
false);
|
|
up_write(&ni->runlist.lock);
|
|
} else
|
|
err = ntfs_non_resident_attr_shrink(ni, newsize);
|
|
} else
|
|
err = ntfs_resident_attr_resize(ni, newsize, 0,
|
|
NVolDisableSparse(ni->vol) ?
|
|
HOLES_NO : HOLES_OK);
|
|
ntfs_debug("Return status %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
int ntfs_attr_expand(struct ntfs_inode *ni, const s64 newsize, const s64 prealloc_size)
|
|
{
|
|
int err = 0;
|
|
|
|
if (newsize < 0 ||
|
|
(ni->mft_no == FILE_MFT && ni->type == AT_DATA)) {
|
|
ntfs_debug("Invalid arguments passed.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x, size %lld\n",
|
|
(unsigned long long)ni->mft_no, ni->type, newsize);
|
|
|
|
if (ni->data_size == newsize) {
|
|
ntfs_debug("Size is already ok\n");
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Encrypted attributes are not supported. We return access denied,
|
|
* which is what Windows NT4 does, too.
|
|
*/
|
|
if (NInoEncrypted(ni)) {
|
|
pr_err("Failed to truncate encrypted attribute\n");
|
|
return -EACCES;
|
|
}
|
|
|
|
if (NInoNonResident(ni)) {
|
|
if (newsize > ni->data_size)
|
|
err = ntfs_non_resident_attr_expand(ni, newsize, prealloc_size,
|
|
NVolDisableSparse(ni->vol) ?
|
|
HOLES_NO : HOLES_OK, true);
|
|
} else
|
|
err = ntfs_resident_attr_resize(ni, newsize, prealloc_size,
|
|
NVolDisableSparse(ni->vol) ?
|
|
HOLES_NO : HOLES_OK);
|
|
if (!err)
|
|
i_size_write(VFS_I(ni), newsize);
|
|
ntfs_debug("Return status %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_truncate_i - resize an ntfs attribute
|
|
* @ni: open ntfs inode to resize
|
|
* @newsize: new size (in bytes) to which to resize the attribute
|
|
* @holes: how to create a hole if expanding
|
|
*
|
|
* Change the size of an open ntfs attribute @na to @newsize bytes. If the
|
|
* attribute is made bigger and the attribute is resident the newly
|
|
* "allocated" space is cleared and if the attribute is non-resident the
|
|
* newly allocated space is marked as not initialised and no real allocation
|
|
* on disk is performed.
|
|
*/
|
|
int ntfs_attr_truncate_i(struct ntfs_inode *ni, const s64 newsize, unsigned int holes)
|
|
{
|
|
int err;
|
|
|
|
if (newsize < 0 ||
|
|
(ni->mft_no == FILE_MFT && ni->type == AT_DATA)) {
|
|
ntfs_debug("Invalid arguments passed.\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x, size %lld\n",
|
|
(unsigned long long)ni->mft_no, ni->type, newsize);
|
|
|
|
if (ni->data_size == newsize) {
|
|
ntfs_debug("Size is already ok\n");
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Encrypted attributes are not supported. We return access denied,
|
|
* which is what Windows NT4 does, too.
|
|
*/
|
|
if (NInoEncrypted(ni)) {
|
|
pr_err("Failed to truncate encrypted attribute\n");
|
|
return -EACCES;
|
|
}
|
|
|
|
if (NInoCompressed(ni)) {
|
|
pr_err("Failed to truncate compressed attribute\n");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
if (NInoNonResident(ni)) {
|
|
if (newsize > ni->data_size)
|
|
err = ntfs_non_resident_attr_expand(ni, newsize, 0, holes, true);
|
|
else
|
|
err = ntfs_non_resident_attr_shrink(ni, newsize);
|
|
} else
|
|
err = ntfs_resident_attr_resize(ni, newsize, 0, holes);
|
|
ntfs_debug("Return status %d\n", err);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* Resize an attribute, creating a hole if relevant
|
|
*/
|
|
int ntfs_attr_truncate(struct ntfs_inode *ni, const s64 newsize)
|
|
{
|
|
return ntfs_attr_truncate_i(ni, newsize,
|
|
NVolDisableSparse(ni->vol) ?
|
|
HOLES_NO : HOLES_OK);
|
|
}
|
|
|
|
int ntfs_attr_map_cluster(struct ntfs_inode *ni, s64 vcn_start, s64 *lcn_start,
|
|
s64 *lcn_count, s64 max_clu_count, bool *balloc, bool update_mp,
|
|
bool skip_holes)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
struct runlist_element *rl, *rlc;
|
|
s64 vcn = vcn_start, lcn, clu_count;
|
|
s64 lcn_seek_from = -1;
|
|
int err = 0;
|
|
size_t new_rl_count;
|
|
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err)
|
|
return err;
|
|
|
|
if (NInoAttr(ni))
|
|
ctx = ntfs_attr_get_search_ctx(ni->ext.base_ntfs_ino, NULL);
|
|
else
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(vol->sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, vcn, NULL, 0, ctx);
|
|
if (err) {
|
|
ntfs_error(vol->sb,
|
|
"ntfs_attr_lookup failed, ntfs inode(mft_no : %llu) type : 0x%x, err : %d",
|
|
ni->mft_no, ni->type, err);
|
|
goto out;
|
|
}
|
|
|
|
rl = ntfs_attr_find_vcn_nolock(ni, vcn, ctx);
|
|
if (IS_ERR(rl)) {
|
|
ntfs_error(vol->sb, "Failed to find run after mapping runlist.");
|
|
err = PTR_ERR(rl);
|
|
goto out;
|
|
}
|
|
|
|
lcn = ntfs_rl_vcn_to_lcn(rl, vcn);
|
|
clu_count = min(max_clu_count, rl->length - (vcn - rl->vcn));
|
|
if (lcn >= LCN_HOLE) {
|
|
if (lcn > LCN_DELALLOC ||
|
|
(lcn == LCN_HOLE && skip_holes)) {
|
|
*lcn_start = lcn;
|
|
*lcn_count = clu_count;
|
|
*balloc = false;
|
|
goto out;
|
|
}
|
|
} else {
|
|
WARN_ON(lcn == LCN_RL_NOT_MAPPED);
|
|
if (lcn == LCN_ENOENT)
|
|
err = -ENOENT;
|
|
else
|
|
err = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
/* Search backwards to find the best lcn to start seek from. */
|
|
rlc = rl;
|
|
while (rlc->vcn) {
|
|
rlc--;
|
|
if (rlc->lcn >= 0) {
|
|
/*
|
|
* avoid fragmenting a compressed file
|
|
* Windows does not do that, and that may
|
|
* not be desirable for files which can
|
|
* be updated
|
|
*/
|
|
if (NInoCompressed(ni))
|
|
lcn_seek_from = rlc->lcn + rlc->length;
|
|
else
|
|
lcn_seek_from = rlc->lcn + (vcn - rlc->vcn);
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (lcn_seek_from == -1) {
|
|
/* Backwards search failed, search forwards. */
|
|
rlc = rl;
|
|
while (rlc->length) {
|
|
rlc++;
|
|
if (rlc->lcn >= 0) {
|
|
lcn_seek_from = rlc->lcn - (rlc->vcn - vcn);
|
|
if (lcn_seek_from < -1)
|
|
lcn_seek_from = -1;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
rlc = ntfs_cluster_alloc(vol, vcn, clu_count, lcn_seek_from, DATA_ZONE,
|
|
false, true, true);
|
|
if (IS_ERR(rlc)) {
|
|
err = PTR_ERR(rlc);
|
|
goto out;
|
|
}
|
|
|
|
WARN_ON(rlc->vcn != vcn);
|
|
lcn = rlc->lcn;
|
|
clu_count = rlc->length;
|
|
|
|
rl = ntfs_runlists_merge(&ni->runlist, rlc, 0, &new_rl_count);
|
|
if (IS_ERR(rl)) {
|
|
ntfs_error(vol->sb, "Failed to merge runlists");
|
|
err = PTR_ERR(rl);
|
|
if (ntfs_cluster_free_from_rl(vol, rlc))
|
|
ntfs_error(vol->sb, "Failed to free hot clusters.");
|
|
kvfree(rlc);
|
|
goto out;
|
|
}
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_count;
|
|
|
|
if (!update_mp) {
|
|
u64 free = atomic64_read(&vol->free_clusters) * 100;
|
|
|
|
do_div(free, vol->nr_clusters);
|
|
if (free <= 5)
|
|
update_mp = true;
|
|
}
|
|
|
|
if (update_mp) {
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
err = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
if (err) {
|
|
int err2;
|
|
|
|
err2 = ntfs_cluster_free(ni, vcn, clu_count, ctx);
|
|
if (err2 < 0)
|
|
ntfs_error(vol->sb,
|
|
"Failed to free cluster allocation. Leaving inconstant metadata.\n");
|
|
goto out;
|
|
}
|
|
} else {
|
|
VFS_I(ni)->i_blocks += clu_count << (vol->cluster_size_bits - 9);
|
|
NInoSetRunlistDirty(ni);
|
|
mark_mft_record_dirty(ni);
|
|
}
|
|
|
|
*lcn_start = lcn;
|
|
*lcn_count = clu_count;
|
|
*balloc = true;
|
|
out:
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_rm - remove attribute from ntfs inode
|
|
* @ni: opened ntfs attribute to delete
|
|
*
|
|
* Remove attribute and all it's extents from ntfs inode. If attribute was non
|
|
* resident also free all clusters allocated by attribute.
|
|
*/
|
|
int ntfs_attr_rm(struct ntfs_inode *ni)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
int err = 0, ret = 0;
|
|
struct ntfs_inode *base_ni;
|
|
struct super_block *sb = ni->vol->sb;
|
|
|
|
if (NInoAttr(ni))
|
|
base_ni = ni->ext.base_ntfs_ino;
|
|
else
|
|
base_ni = ni;
|
|
|
|
ntfs_debug("Entering for inode 0x%llx, attr 0x%x.\n",
|
|
(long long) ni->mft_no, ni->type);
|
|
|
|
/* Free cluster allocation. */
|
|
if (NInoNonResident(ni)) {
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
if (err)
|
|
return err;
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
ret = ntfs_cluster_free(ni, 0, -1, ctx);
|
|
if (ret < 0)
|
|
ntfs_error(sb,
|
|
"Failed to free cluster allocation. Leaving inconstant metadata.\n");
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
}
|
|
|
|
/* Search for attribute extents and remove them all. */
|
|
ctx = ntfs_attr_get_search_ctx(base_ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(sb, "%s: Failed to get search context", __func__);
|
|
return -ENOMEM;
|
|
}
|
|
while (!(err = ntfs_attr_lookup(ni->type, ni->name, ni->name_len,
|
|
CASE_SENSITIVE, 0, NULL, 0, ctx))) {
|
|
err = ntfs_attr_record_rm(ctx);
|
|
if (err) {
|
|
ntfs_error(sb,
|
|
"Failed to remove attribute extent. Leaving inconstant metadata.\n");
|
|
ret = err;
|
|
}
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
}
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
if (err != -ENOENT) {
|
|
ntfs_error(sb, "Attribute lookup failed. Probably leaving inconstant metadata.\n");
|
|
ret = err;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_attr_exist(struct ntfs_inode *ni, const __le32 type, __le16 *name,
|
|
u32 name_len)
|
|
{
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
int ret;
|
|
|
|
ntfs_debug("Entering\n");
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ntfs_error(ni->vol->sb, "%s: Failed to get search context",
|
|
__func__);
|
|
return 0;
|
|
}
|
|
|
|
ret = ntfs_attr_lookup(type, name, name_len, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
|
|
return !ret;
|
|
}
|
|
|
|
int ntfs_attr_remove(struct ntfs_inode *ni, const __le32 type, __le16 *name,
|
|
u32 name_len)
|
|
{
|
|
int err;
|
|
struct inode *attr_vi;
|
|
struct ntfs_inode *attr_ni;
|
|
|
|
ntfs_debug("Entering\n");
|
|
|
|
if (!ni)
|
|
return -EINVAL;
|
|
|
|
attr_vi = ntfs_attr_iget(VFS_I(ni), type, name, name_len);
|
|
if (IS_ERR(attr_vi)) {
|
|
err = PTR_ERR(attr_vi);
|
|
ntfs_error(ni->vol->sb, "Failed to open attribute 0x%02x of inode 0x%llx",
|
|
type, (unsigned long long)ni->mft_no);
|
|
return err;
|
|
}
|
|
attr_ni = NTFS_I(attr_vi);
|
|
|
|
err = ntfs_attr_rm(attr_ni);
|
|
if (err)
|
|
ntfs_error(ni->vol->sb, "Failed to remove attribute 0x%02x of inode 0x%llx",
|
|
type, (unsigned long long)ni->mft_no);
|
|
iput(attr_vi);
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* ntfs_attr_readall - read the entire data from an ntfs attribute
|
|
* @ni: open ntfs inode in which the ntfs attribute resides
|
|
* @type: attribute type
|
|
* @name: attribute name in little endian Unicode or AT_UNNAMED or NULL
|
|
* @name_len: length of attribute @name in Unicode characters (if @name given)
|
|
* @data_size: if non-NULL then store here the data size
|
|
*
|
|
* This function will read the entire content of an ntfs attribute.
|
|
* If @name is AT_UNNAMED then look specifically for an unnamed attribute.
|
|
* If @name is NULL then the attribute could be either named or not.
|
|
* In both those cases @name_len is not used at all.
|
|
*
|
|
* On success a buffer is allocated with the content of the attribute
|
|
* and which needs to be freed when it's not needed anymore. If the
|
|
* @data_size parameter is non-NULL then the data size is set there.
|
|
*/
|
|
void *ntfs_attr_readall(struct ntfs_inode *ni, const __le32 type,
|
|
__le16 *name, u32 name_len, s64 *data_size)
|
|
{
|
|
struct ntfs_inode *bmp_ni;
|
|
struct inode *bmp_vi;
|
|
void *data, *ret = NULL;
|
|
s64 size;
|
|
struct super_block *sb = ni->vol->sb;
|
|
|
|
ntfs_debug("Entering\n");
|
|
|
|
bmp_vi = ntfs_attr_iget(VFS_I(ni), type, name, name_len);
|
|
if (IS_ERR(bmp_vi)) {
|
|
ntfs_debug("ntfs_attr_iget failed");
|
|
goto err_exit;
|
|
}
|
|
bmp_ni = NTFS_I(bmp_vi);
|
|
|
|
data = kvmalloc(bmp_ni->data_size, GFP_NOFS);
|
|
if (!data)
|
|
goto out;
|
|
|
|
size = ntfs_inode_attr_pread(VFS_I(bmp_ni), 0, bmp_ni->data_size,
|
|
(u8 *)data);
|
|
if (size != bmp_ni->data_size) {
|
|
ntfs_error(sb, "ntfs_attr_pread failed");
|
|
kvfree(data);
|
|
goto out;
|
|
}
|
|
ret = data;
|
|
if (data_size)
|
|
*data_size = size;
|
|
out:
|
|
iput(bmp_vi);
|
|
err_exit:
|
|
ntfs_debug("\n");
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_non_resident_attr_insert_range(struct ntfs_inode *ni, s64 start_vcn, s64 len)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct runlist_element *hole_rl, *rl;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
int ret;
|
|
size_t new_rl_count;
|
|
|
|
if (NInoAttr(ni) || ni->type != AT_DATA)
|
|
return -EOPNOTSUPP;
|
|
if (start_vcn > ntfs_bytes_to_cluster(vol, ni->allocated_size))
|
|
return -EINVAL;
|
|
|
|
hole_rl = kmalloc(sizeof(*hole_rl) * 2, GFP_NOFS);
|
|
if (!hole_rl)
|
|
return -ENOMEM;
|
|
hole_rl[0].vcn = start_vcn;
|
|
hole_rl[0].lcn = LCN_HOLE;
|
|
hole_rl[0].length = len;
|
|
hole_rl[1].vcn = start_vcn + len;
|
|
hole_rl[1].lcn = LCN_ENOENT;
|
|
hole_rl[1].length = 0;
|
|
|
|
down_write(&ni->runlist.lock);
|
|
ret = ntfs_attr_map_whole_runlist(ni);
|
|
if (ret) {
|
|
up_write(&ni->runlist.lock);
|
|
return ret;
|
|
}
|
|
|
|
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
|
|
if (!rl) {
|
|
up_write(&ni->runlist.lock);
|
|
kfree(hole_rl);
|
|
return -EIO;
|
|
}
|
|
|
|
rl = ntfs_rl_insert_range(ni->runlist.rl, (int)ni->runlist.count,
|
|
hole_rl, 1, &new_rl_count);
|
|
if (IS_ERR(rl)) {
|
|
up_write(&ni->runlist.lock);
|
|
kfree(hole_rl);
|
|
return PTR_ERR(rl);
|
|
}
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_count;
|
|
|
|
ni->allocated_size += ntfs_cluster_to_bytes(vol, len);
|
|
ni->data_size += ntfs_cluster_to_bytes(vol, len);
|
|
if (ntfs_cluster_to_bytes(vol, start_vcn) < ni->initialized_size)
|
|
ni->initialized_size += ntfs_cluster_to_bytes(vol, len);
|
|
ret = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
up_write(&ni->runlist.lock);
|
|
if (ret)
|
|
return ret;
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ret = -ENOMEM;
|
|
return ret;
|
|
}
|
|
|
|
ret = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx);
|
|
if (ret) {
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return ret;
|
|
}
|
|
|
|
ctx->attr->data.non_resident.data_size = cpu_to_le64(ni->data_size);
|
|
ctx->attr->data.non_resident.initialized_size = cpu_to_le64(ni->initialized_size);
|
|
if (ni->type == AT_DATA && ni->name == AT_UNNAMED)
|
|
NInoSetFileNameDirty(ni);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_non_resident_attr_collapse_range(struct ntfs_inode *ni, s64 start_vcn, s64 len)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct runlist_element *punch_rl, *rl;
|
|
struct ntfs_attr_search_ctx *ctx = NULL;
|
|
s64 end_vcn;
|
|
int dst_cnt;
|
|
int ret;
|
|
size_t new_rl_cnt;
|
|
|
|
if (NInoAttr(ni) || ni->type != AT_DATA)
|
|
return -EOPNOTSUPP;
|
|
|
|
end_vcn = ntfs_bytes_to_cluster(vol, ni->allocated_size);
|
|
if (start_vcn >= end_vcn)
|
|
return -EINVAL;
|
|
|
|
down_write(&ni->runlist.lock);
|
|
ret = ntfs_attr_map_whole_runlist(ni);
|
|
if (ret) {
|
|
up_write(&ni->runlist.lock);
|
|
return ret;
|
|
}
|
|
|
|
len = min(len, end_vcn - start_vcn);
|
|
for (rl = ni->runlist.rl, dst_cnt = 0; rl && rl->length; rl++)
|
|
dst_cnt++;
|
|
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
|
|
if (!rl) {
|
|
up_write(&ni->runlist.lock);
|
|
return -EIO;
|
|
}
|
|
|
|
rl = ntfs_rl_collapse_range(ni->runlist.rl, dst_cnt + 1,
|
|
start_vcn, len, &punch_rl, &new_rl_cnt);
|
|
if (IS_ERR(rl)) {
|
|
up_write(&ni->runlist.lock);
|
|
return PTR_ERR(rl);
|
|
}
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_cnt;
|
|
|
|
ni->allocated_size -= ntfs_cluster_to_bytes(vol, len);
|
|
if (ni->data_size > ntfs_cluster_to_bytes(vol, start_vcn)) {
|
|
if (ni->data_size > ntfs_cluster_to_bytes(vol, (start_vcn + len)))
|
|
ni->data_size -= ntfs_cluster_to_bytes(vol, len);
|
|
else
|
|
ni->data_size = ntfs_cluster_to_bytes(vol, start_vcn);
|
|
}
|
|
if (ni->initialized_size > ntfs_cluster_to_bytes(vol, start_vcn)) {
|
|
if (ni->initialized_size >
|
|
ntfs_cluster_to_bytes(vol, start_vcn + len))
|
|
ni->initialized_size -= ntfs_cluster_to_bytes(vol, len);
|
|
else
|
|
ni->initialized_size = ntfs_cluster_to_bytes(vol, start_vcn);
|
|
}
|
|
|
|
if (ni->allocated_size > 0) {
|
|
ret = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
if (ret) {
|
|
up_write(&ni->runlist.lock);
|
|
goto out_rl;
|
|
}
|
|
}
|
|
up_write(&ni->runlist.lock);
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, NULL);
|
|
if (!ctx) {
|
|
ret = -ENOMEM;
|
|
goto out_rl;
|
|
}
|
|
|
|
ret = ntfs_attr_lookup(ni->type, ni->name, ni->name_len, CASE_SENSITIVE,
|
|
0, NULL, 0, ctx);
|
|
if (ret)
|
|
goto out_ctx;
|
|
|
|
ctx->attr->data.non_resident.data_size = cpu_to_le64(ni->data_size);
|
|
ctx->attr->data.non_resident.initialized_size = cpu_to_le64(ni->initialized_size);
|
|
if (ni->allocated_size == 0)
|
|
ntfs_attr_make_resident(ni, ctx);
|
|
mark_mft_record_dirty(ctx->ntfs_ino);
|
|
|
|
ret = ntfs_cluster_free_from_rl(vol, punch_rl);
|
|
if (ret)
|
|
ntfs_error(vol->sb, "Freeing of clusters failed");
|
|
out_ctx:
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
out_rl:
|
|
kvfree(punch_rl);
|
|
mark_mft_record_dirty(ni);
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_non_resident_attr_punch_hole(struct ntfs_inode *ni, s64 start_vcn, s64 len)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct runlist_element *punch_rl, *rl;
|
|
s64 end_vcn;
|
|
int dst_cnt;
|
|
int ret;
|
|
size_t new_rl_count;
|
|
|
|
if (NInoAttr(ni) || ni->type != AT_DATA)
|
|
return -EOPNOTSUPP;
|
|
|
|
end_vcn = ntfs_bytes_to_cluster(vol, ni->allocated_size);
|
|
if (start_vcn >= end_vcn)
|
|
return -EINVAL;
|
|
|
|
down_write(&ni->runlist.lock);
|
|
ret = ntfs_attr_map_whole_runlist(ni);
|
|
if (ret) {
|
|
up_write(&ni->runlist.lock);
|
|
return ret;
|
|
}
|
|
|
|
len = min(len, end_vcn - start_vcn + 1);
|
|
for (rl = ni->runlist.rl, dst_cnt = 0; rl && rl->length; rl++)
|
|
dst_cnt++;
|
|
rl = ntfs_rl_find_vcn_nolock(ni->runlist.rl, start_vcn);
|
|
if (!rl) {
|
|
up_write(&ni->runlist.lock);
|
|
return -EIO;
|
|
}
|
|
|
|
rl = ntfs_rl_punch_hole(ni->runlist.rl, dst_cnt + 1,
|
|
start_vcn, len, &punch_rl, &new_rl_count);
|
|
if (IS_ERR(rl)) {
|
|
up_write(&ni->runlist.lock);
|
|
return PTR_ERR(rl);
|
|
}
|
|
ni->runlist.rl = rl;
|
|
ni->runlist.count = new_rl_count;
|
|
|
|
ret = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
up_write(&ni->runlist.lock);
|
|
if (ret) {
|
|
kvfree(punch_rl);
|
|
return ret;
|
|
}
|
|
|
|
ret = ntfs_cluster_free_from_rl(vol, punch_rl);
|
|
if (ret)
|
|
ntfs_error(vol->sb, "Freeing of clusters failed");
|
|
|
|
kvfree(punch_rl);
|
|
mark_mft_record_dirty(ni);
|
|
return ret;
|
|
}
|
|
|
|
int ntfs_attr_fallocate(struct ntfs_inode *ni, loff_t start, loff_t byte_len, bool keep_size)
|
|
{
|
|
struct ntfs_volume *vol = ni->vol;
|
|
struct mft_record *mrec;
|
|
struct ntfs_attr_search_ctx *ctx;
|
|
s64 old_data_size;
|
|
s64 vcn_start, vcn_end, vcn_uninit, vcn, try_alloc_cnt;
|
|
s64 lcn, alloc_cnt;
|
|
int err = 0;
|
|
struct runlist_element *rl;
|
|
bool balloc;
|
|
|
|
if (NInoAttr(ni) || ni->type != AT_DATA)
|
|
return -EINVAL;
|
|
|
|
if (NInoNonResident(ni) && !NInoFullyMapped(ni)) {
|
|
down_write(&ni->runlist.lock);
|
|
err = ntfs_attr_map_whole_runlist(ni);
|
|
up_write(&ni->runlist.lock);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
mutex_lock_nested(&ni->mrec_lock, NTFS_INODE_MUTEX_NORMAL);
|
|
mrec = map_mft_record(ni);
|
|
if (IS_ERR(mrec)) {
|
|
mutex_unlock(&ni->mrec_lock);
|
|
return PTR_ERR(mrec);
|
|
}
|
|
|
|
ctx = ntfs_attr_get_search_ctx(ni, mrec);
|
|
if (!ctx) {
|
|
err = -ENOMEM;
|
|
goto out_unmap;
|
|
}
|
|
|
|
err = ntfs_attr_lookup(AT_DATA, AT_UNNAMED, 0, 0, 0, NULL, 0, ctx);
|
|
if (err) {
|
|
err = -EIO;
|
|
goto out_unmap;
|
|
}
|
|
|
|
old_data_size = ni->data_size;
|
|
if (start + byte_len > ni->data_size) {
|
|
err = ntfs_attr_truncate(ni, start + byte_len);
|
|
if (err)
|
|
goto out_unmap;
|
|
if (keep_size) {
|
|
ntfs_attr_reinit_search_ctx(ctx);
|
|
err = ntfs_attr_lookup(AT_DATA, AT_UNNAMED, 0, 0, 0, NULL, 0, ctx);
|
|
if (err) {
|
|
err = -EIO;
|
|
goto out_unmap;
|
|
}
|
|
ni->data_size = old_data_size;
|
|
if (NInoNonResident(ni))
|
|
ctx->attr->data.non_resident.data_size =
|
|
cpu_to_le64(old_data_size);
|
|
else
|
|
ctx->attr->data.resident.value_length =
|
|
cpu_to_le32((u32)old_data_size);
|
|
mark_mft_record_dirty(ni);
|
|
}
|
|
}
|
|
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
unmap_mft_record(ni);
|
|
mutex_unlock(&ni->mrec_lock);
|
|
|
|
if (!NInoNonResident(ni))
|
|
goto out;
|
|
|
|
vcn_start = (s64)ntfs_bytes_to_cluster(vol, start);
|
|
vcn_end = (s64)ntfs_bytes_to_cluster(vol,
|
|
round_up(start + byte_len, vol->cluster_size));
|
|
vcn_uninit = (s64)ntfs_bytes_to_cluster(vol,
|
|
round_up(ni->initialized_size, vol->cluster_size));
|
|
vcn_uninit = min_t(s64, vcn_uninit, vcn_end);
|
|
|
|
/*
|
|
* we have to allocate clusters for holes and delayed within initialized_size,
|
|
* and zero out the clusters only for the holes.
|
|
*/
|
|
vcn = vcn_start;
|
|
while (vcn < vcn_uninit) {
|
|
down_read(&ni->runlist.lock);
|
|
rl = ntfs_attr_find_vcn_nolock(ni, vcn, NULL);
|
|
up_read(&ni->runlist.lock);
|
|
if (IS_ERR(rl)) {
|
|
err = PTR_ERR(rl);
|
|
goto out;
|
|
}
|
|
|
|
if (rl->lcn > 0) {
|
|
vcn += rl->length - (vcn - rl->vcn);
|
|
} else if (rl->lcn == LCN_DELALLOC || rl->lcn == LCN_HOLE) {
|
|
try_alloc_cnt = min(rl->length - (vcn - rl->vcn),
|
|
vcn_uninit - vcn);
|
|
|
|
if (rl->lcn == LCN_DELALLOC) {
|
|
vcn += try_alloc_cnt;
|
|
continue;
|
|
}
|
|
|
|
while (try_alloc_cnt > 0) {
|
|
mutex_lock_nested(&ni->mrec_lock, NTFS_INODE_MUTEX_NORMAL);
|
|
down_write(&ni->runlist.lock);
|
|
err = ntfs_attr_map_cluster(ni, vcn, &lcn, &alloc_cnt,
|
|
try_alloc_cnt, &balloc, false, false);
|
|
up_write(&ni->runlist.lock);
|
|
mutex_unlock(&ni->mrec_lock);
|
|
if (err)
|
|
goto out;
|
|
|
|
err = ntfs_dio_zero_range(VFS_I(ni),
|
|
lcn << vol->cluster_size_bits,
|
|
alloc_cnt << vol->cluster_size_bits);
|
|
if (err > 0)
|
|
goto out;
|
|
|
|
if (signal_pending(current))
|
|
goto out;
|
|
|
|
vcn += alloc_cnt;
|
|
try_alloc_cnt -= alloc_cnt;
|
|
}
|
|
} else {
|
|
err = -EIO;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
/* allocate clusters outside of initialized_size */
|
|
try_alloc_cnt = vcn_end - vcn;
|
|
while (try_alloc_cnt > 0) {
|
|
mutex_lock_nested(&ni->mrec_lock, NTFS_INODE_MUTEX_NORMAL);
|
|
down_write(&ni->runlist.lock);
|
|
err = ntfs_attr_map_cluster(ni, vcn, &lcn, &alloc_cnt,
|
|
try_alloc_cnt, &balloc, false, false);
|
|
up_write(&ni->runlist.lock);
|
|
mutex_unlock(&ni->mrec_lock);
|
|
if (err || signal_pending(current))
|
|
goto out;
|
|
|
|
vcn += alloc_cnt;
|
|
try_alloc_cnt -= alloc_cnt;
|
|
cond_resched();
|
|
}
|
|
|
|
if (NInoRunlistDirty(ni)) {
|
|
mutex_lock_nested(&ni->mrec_lock, NTFS_INODE_MUTEX_NORMAL);
|
|
down_write(&ni->runlist.lock);
|
|
err = ntfs_attr_update_mapping_pairs(ni, 0);
|
|
if (err)
|
|
ntfs_error(ni->vol->sb, "Updating mapping pairs failed");
|
|
else
|
|
NInoClearRunlistDirty(ni);
|
|
up_write(&ni->runlist.lock);
|
|
mutex_unlock(&ni->mrec_lock);
|
|
}
|
|
return err;
|
|
out_unmap:
|
|
if (ctx)
|
|
ntfs_attr_put_search_ctx(ctx);
|
|
unmap_mft_record(ni);
|
|
mutex_unlock(&ni->mrec_lock);
|
|
out:
|
|
return err >= 0 ? 0 : err;
|
|
}
|